url
stringlengths
14
1.76k
text
stringlengths
100
1.02M
metadata
stringlengths
1.06k
1.1k
https://www.physicsforums.com/threads/interpetation-of-wavefunction-atom.521940/
# Interpetation of wavefunction - atom 1. Aug 16, 2011 ### liquidFuzz It's now years since I studied this, if anyone could help me remember. If I look at a hydrogen atom and it's shells. In the ground state there's 1s, the wave function is then: $\Psi_{nlm}(r,\theta,\phi) = Y_{lm}(\theta,\phi)R_{nl}(r) = Y_{00}R_{10}$ $Y_{00} = \frac{1}{\sqrt{4\pi}}$ $R_{10} = 2 \left(\frac{Z}{a_0}\right)^{3/2} e^{-\frac{\rho}{2}}$ With hydrogen's single proton and n = 1 etc. $\Psi_{100} = \frac{1}{\sqrt{\pi}} \left(\frac{1}{a_0}\right)^{3/2} e^{-\frac{r}{a_0}}$ - I just looked up this in Physics Handbook. Now I can't remember how to interpret this. My stumbling interpretations I the Bohr model there's fixed energy levels, but this calculation shows a high likeliness of the atom being in the centre. Is this right? Further, is the model predicting a likely radial position for a electron sitting in the 1s energy state? 2. Aug 16, 2011 ### Staff: Mentor The square of this expression gives the probability distribution for the electron being at location $(r, \theta, \phi)$ with respect to the nucleus. Obviously in this case (n = 1, l = 0, m = 0) it doesn't actually depend on either of the angles; but for other combinations of n, l, m it does. For this case (the ground state) the most likely location is in fact at the center of the atom, i.e. inside the nucleus, because this is the location that gives you the maximum value for the probability distribution. 3. Aug 16, 2011 ### vanhees71 The Bohr model is wrong in many respects. Particularly it's wrong to still learn about it in the 21st century. Better forget everything about the Bohr model first and then remember the modern version of quantum mechanics which has been discovered in 1925/26 by Heisenberg, Born, Jordan, Dirac, and Schrödinger. Quantum mechanics can be described in the position representation as representing the states of the electron in the hydrogen atom by a square-integrable wave function. In the context of atomic physics most important states are the energy-eigen states, which (according to the dynamics of quantum theory) represent stationary states of the electron. To determine the energy states completely, in addition to the energy eigenvalue you need also to specify the angular-momentum state. This is given by some quantum numbers, $l$ and $m$. The eigenvalues of the modulus of the orbital angular momentum are given by $l(l+1)\hbar^2$, where $l \in \{0,1,2,\ldots \}$, and then for each $l$ the "magnet-quantum number" can run in $m \in \{0,\pm 1,\pm 2,\ldots,\pm l \}$, and the $z$ component of orbital angular momentum is given by $m \hbar$. The lowest energy state has necessarily $l=m=0$, and thus is non-degenerate (I leave out the discussion of spin here!). Now, according to the funcamental rules of quantum theory, the squared modulus of the wave function is the position-probability distribution for the electron (Born's probability interpretation), if it is in the state discribed by this wave function. In your case, you are looking at the ground state, i.e., the energy eigenstate with the lowest possible energy value, and this is the most stable stationary state of the atom. Now, if you plot the position-probability distribution according to your ground-state wave function, you'll find it to be spherically symmetric (that's what $l=0$ is telling you) and having its maximum in the center (i.e., at the nuclues). The Bohr radius of the first orbit is only the average distance of the electron from the nucleus, and the atom is spherical and not a little disk as suggested by the Bohr model. The probablity-density distributions, you can fine nicely depicted in many books on introductory quantum mechanics. Nice colorful pictures can be found in Wikipedia, where you also find a nice description of the quantum theory of the hydrogen atom: http://en.wikipedia.org/wiki/Hydrogen_atom 4. Aug 16, 2011 ### liquidFuzz jtbell - Thanks for shedding a light on this. vanhees71 - Thanks for the I'll redeem this fella approach in your vigorous post. You're right about the Bohr model and the quantum model. But regarding the tutorial aspect of learning obsolete models are completely wrong. The atom as we see it underwent big changes, one step lead to an other. Te Bohr model was a big leap, for man kind. Ops, I got carried away in the last sentence. ;-) 5. Aug 16, 2011 ### vanhees71 Well, the question, whether one should learn/teach historically important but outdated views on nature in the physics curriculum. On the one hand, I'm strongly objecting the historical approach to teach physics. It's misleading to teach a beginner in quantum theory Bohr's and Sommerfeld's model of atoms. As much misleading is it to teach students the "relativistic-mass" concept or that temperature depends on the motion of the observer wrt. the body, whose temperature one likes to describe etc. although these were well founded ideas in the early times of quantum mechanics and relativity. The modern theories are better to comprehend and have at the same time a broader range of applicability since they represent the progress in our understanding of the fundamental laws of nature, and it does not make much sense to teach first some outdated points of view that lead to more confusion than they help to comprehend the subject. On the other hand, I love history of science since it is a very interesting subject. To figure out, how physics came to the knowledge about the fundamental laws of nature is not only very fascinating but can also help to understand better the more modern points of view, but as I said above, I think, one should present physics not in a historical but in a logical order. Of course, a good physics course also includes some historical context, but this should be presented after the students have learnt about the most logical and comprehensible models first. Then it is the more fascinating to learn how much effort of the greatest minds in history went into the world view we have today. I think the best example of this approach to teach (theoretical) physics are the textbooks by Steven Weinberg (e.g., his 3 volumes on quantum field theory). 6. Aug 17, 2011 ### liquidFuzz vanhees71 - yes, something along those lines. On topic Lets say my precious hydrogen atom finds a friend and becomes: $H_2^+$ To find out how they behave, or bond to each other I use the wave model. $\Psi = \frac{1}{\sqrt{2}}(1s_1 + 1s_2)$ and $\Psi = \frac{1}{\sqrt{2}}(1s_1 - 1s_2)$ I actually feel somewhat unsure of what the minus tells me. Up wave down wave..? Further, how do I determine wish states are possible? Similar Discussions: Interpetation of wavefunction - atom
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8249311447143555, "perplexity": 467.8355574062737}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-39/segments/1505818687484.46/warc/CC-MAIN-20170920213425-20170920233425-00386.warc.gz"}
http://openstudy.com/updates/55b6570ee4b0ce1056622942
Here's the question you clicked on: 55 members online • 0 viewing ## ksaimouli one year ago simplify Delete Cancel Submit • This Question is Closed 1. ksaimouli • one year ago Best Response You've already chosen the best response. 0 $\frac{ 2x^2-xy-3y^2 }{ 2x^2-11xy+12y^2 }$ 2. freckles • one year ago Best Response You've already chosen the best response. 0 You can factor the numerator and the denominator. 3. freckles • one year ago Best Response You've already chosen the best response. 0 $\text{ for example } \\ 2x^2-xy-3y^2 \\ 2x^2+2xy-3xy-3y^2 \\ 2x(x+y)-3y(x+y) \\ (x+y)(2x-3y)$ Try factoring the denominator. Hint: You can do what I did above. Just have to think what two integer numbers when multiplied gives 24 and add ups to be -11. 4. Not the answer you are looking for? Search for more explanations. • Attachments: Find more explanations on OpenStudy ##### spraguer (Moderator) 5→ View Detailed Profile 23 • Teamwork 19 Teammate • Problem Solving 19 Hero • You have blocked this person. • ✔ You're a fan Checking fan status... Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.998577356338501, "perplexity": 9308.977346458958}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988719566.74/warc/CC-MAIN-20161020183839-00071-ip-10-171-6-4.ec2.internal.warc.gz"}
https://codereview.stackexchange.com/questions/166136/datatable-runtime-grouping-on-user-defined-criteria
# DataTable runtime grouping on user defined criteria I one of my tools I use the user can define grouping for a DataTable. The criteria are known only at run-time. To achieve this I use a Dictionary<string, object> with a custom dictionary comparer that the GroupBy extension cosumes so that I can aggregate the groups. This works pretty well but I was wondering if there is still room for improvement. The number and names of columns can change. There is nothing known about the DataTable until a JSON configuration is loaded. This is an example of such a configuration from one of my unit tests. "Columns": [ "_nvarchar | key", "_datetime2", "_int | sum", "_float | avg", "_bit | count", "_money", "_numeric" ], (The configuration is already being taken care of and is not a part of the review, it's just for reference to show how the columns can be configured.) To make it simpler here's some example data: var dataTable = new DataTable(); The keyColumns are those marked with key - they are not the same as database keys and they are for aggregation later. var keyColumns = new[] { "FirstName", "LastName" }; The implementation is very simple. Just enumerate the rows and put the name of the key column in the dictionary's key and the value of that column as the value of that dictionary item. This is exactly the implementation I currently use: var rowGroups = dataTable.AsEnumerable().GroupBy(x => keyColumns.ToDictionary( column => column, column => x[column], StringComparer.OrdinalIgnoreCase ), new DictionaryComparer<string, object>() ).ToList(); To actually be able to group the dictionaries there is also this comparer: public class DictionaryComparer<TKey, TValue> : IEqualityComparer<IDictionary<TKey, TValue>> { public bool Equals(IDictionary<TKey, TValue> x, IDictionary<TKey, TValue> y) { return x.All(item => y.TryGetValue(item.Key, out TValue value) && item.Value.Equals(value)); } // Ignore dictionary hash-code and just compare the keys and values. public int GetHashCode(IDictionary<TKey, TValue> obj) => 0; } And the result is: Honestly, I am not sure if a dictionary is the right data structure for realizing composed keys. At least for me it is not directly obviously how the data are grouped. I think the main issue that confuses me is, that the grouping key contains not just the values to group by, but also the column headers which is absolutely irrelevant for grouping. Additionally I have to understand the comparer. For your example, you could easily use a tuple which provides build-in comparison of each item: var rowGroups = dataTable.AsEnumerable() .GroupBy(x => Tuple.Create(x["FirstName"], x["LastName"])) .ToList(); Another more flexible approach could be to create a string value to be used as grouping key: var rowGroups = dataTable.AsEnumerable() .GroupBy(x => string.Join(":", keyColumns.Select(key => x[key]))) .ToList(); • The column name for the group seems really to be a stupid idea... I'm wondering why did I think I need it. I'll throw it away but I'll have to keep the comparer and instead it'll use the SequenceEqual for two collections to determine if the keys are same because the columns can by of any type so it needs to use Equals internally. – t3chb0t Jun 19 '17 at 18:13 • Do you mean something like .GroupBy(x => keyColumns.Select(key => x[key]).ToArray()) + custom comparer using SequenceEqual? – JanDotNet Jun 19 '17 at 18:29 • Exactly, I've just tested it and it works great. Now I'm also able to add a proper hash-code for such a collection. – t3chb0t Jun 19 '17 at 18:31 • Sounds good. Another option could be to create a class like ComposableKey, getting a list of objects with custom Equals / GetHashCode implemenation. That is probably more descriptive in usage but less flexible compared to a comparer. – JanDotNet Jun 19 '17 at 18:41 This is what I ended up with. I removed the meaningless dictionary and replaced it with a CompositeKey. I need additional processing for the values thus the additional let to make the lines more compact. var rowGroups = from dataRow in dataRows let values = keyColumns.Select(column => column.Filter.Apply(dataRow[column.Name])) group dataRow by new CompositeKey<object>(values) into g select g; I like the suggestion with the CompositeKey so I move the loging from the comparer into this one. It now uses SequenceEqual to compare the keys and has a proper GetHashCode implementation (I hope so). public class CompositeKey<T> : IEnumerable<T>, IEquatable<CompositeKey<T>> { public CompositeKey(IEnumerable<T> keys) => _keys = new List<T>(keys); public IEnumerator<T> GetEnumerator() => _keys.GetEnumerator(); IEnumerator IEnumerable.GetEnumerator() => GetEnumerator(); public bool Equals(CompositeKey<T> other) { if (ReferenceEquals(other, null)) return false; return ReferenceEquals(this, other) || this.SequenceEqual(other); } public override bool Equals(object obj) { if (ReferenceEquals(null, obj)) return false; if (ReferenceEquals(this, obj)) return true; if (obj.GetType() != this.GetType()) return false; return Equals((CompositeKey<T>)obj); } public override int GetHashCode() { unchecked { return this.Aggregate(0, (current, next) => (current * 397) ^ next?.GetHashCode() ?? 0); } } }
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.19538705050945282, "perplexity": 4576.331262809271}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153491.18/warc/CC-MAIN-20210727202227-20210727232227-00595.warc.gz"}
https://kb.osu.edu/dspace/handle/1811/18475
THE $3\nu_{2}$ OVERTONE BAND OF $H_{3}^{+}$ Please use this identifier to cite or link to this item: http://hdl.handle.net/1811/18475 Files Size Format View 1993-MG-02.jpg 76.60Kb JPEG image Title: THE $3\nu_{2}$ OVERTONE BAND OF $H_{3}^{+}$ Creators: Ventrudo, Brian F.; Guo, Zhaoyu; Cassidy, Daniel T.; Joo, Sangwoo; Oka, T. Issue Date: 1993 Publisher: Ohio State University Abstract: Because of the small mass of hydrogen and the relatively shallow vibrational potential, the transition probabilities of overtone bands of $H_{3}^{+}$ do not decrease as drastically as those of other molecules. According to the first principle calculation by Dinelli, Miller, and $Tennyson^{1}$, the spontaneous emission probability of the second overtone band $3\nu_{2} (\ell = 1) \rightarrow 0$ is smaller than that of the fundamental band by a factor of less than 10. This corresponds to the reduction of absorption coefficient of about 180 and the spectrum is observable at the $H_{3}^{+}$ ion concentration of $3 \times 10^{10} cm^{-3}$. We have observed 15 vibration rotation transitions from 7265.9 to $6806 cm^{-1}$. InGaAsP diode lasers were used in a short-external cavity configuration to achieve laser operation in a single longitudinal mode. The observed frequencies agree with earlier prediction by Miller and Tennyson within $3 cm^{-1}$ but agree better with their new $prediction^{2}$. Description: $^{1}$. B.M. Dinelli, S. Miller, and J. Tennyson, J. Mol. Spectrose. 153, 718-725(1992) $^{2}$. B.M. Dinelli, private communication. Author Institution: Department of Engineering Physics, McMaster University; Department of Chemistry and Department of Astronomy and Astrophysics., The University of Chicago URI: http://hdl.handle.net/1811/18475 Other Identifiers: 1993-MG-2
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8180495500564575, "perplexity": 3888.9593627425274}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988720380.80/warc/CC-MAIN-20161020183840-00543-ip-10-171-6-4.ec2.internal.warc.gz"}
http://www.ni.com/documentation/en/labview/latest/analysis-node-ref/pattern-generator-sine/
# Pattern Generator (Sine) (G Dataflow) Version: Generates a signal containing a sine pattern. ## amplitude Amplitude of the pattern. Default: 1 ## cycles Number of complete periods of the sine pattern. Note Because this input is a floating-point number, fractional cycles are possible for the sine pattern. Furthermore, setting this input to a negative number does not generate an error condition because it is mathematically correct and useful to consider negative frequencies in Fourier and spectral analysis. Default: 1 ## phase Phase, in degrees, of the sine wave without reset. This input must be in degrees rather than radians. If this input is in radians, convert the value to degrees before using this node. Default: 0 ## error in Error conditions that occur before this node runs. The node responds to this input according to standard error behavior. Standard Error Behavior Many nodes provide an error in input and an error out output so that the node can respond to and communicate errors that occur while code is running. The value of error in specifies whether an error occurred before the node runs. Most nodes respond to values of error in in a standard, predictable way. error in does not contain an error error in contains an error If no error occurred before the node runs, the node begins execution normally. If no error occurs while the node runs, it returns no error. If an error does occur while the node runs, it returns that error information as error out. If an error occurred before the node runs, the node does not execute. Instead, it returns the error in value as error out. Default: No error ## dt Sampling rate in samples per second. This input is available only if you configure this node to return a waveform. Default: 0.1 ## samples Number of samples in the pattern. samples must be greater than 0. Otherwise, this node returns an error. Default: 128 ## t0 Timestamp of the output signal. If this input is unwired, this node uses the current time as the timestamp of the output signal. This input is available only if you configure this node to return a waveform. ## sine pattern Output sine pattern. The largest output pattern this node can generate depends on the amount of memory in your system and is theoretically limited to 2,147,483,647 (231 - 1) elements. This output can return a waveform or an array of double-precision, floating-point numbers. ## error out Error information. The node produces this output according to standard error behavior. Standard Error Behavior Many nodes provide an error in input and an error out output so that the node can respond to and communicate errors that occur while code is running. The value of error in specifies whether an error occurred before the node runs. Most nodes respond to values of error in in a standard, predictable way. error in does not contain an error error in contains an error If no error occurred before the node runs, the node begins execution normally. If no error occurs while the node runs, it returns no error. If an error does occur while the node runs, it returns that error information as error out. If an error occurred before the node runs, the node does not execute. Instead, it returns the error in value as error out. ## Algorithm for Generating the Sine Pattern If the sequence Y represents sine pattern, this node generates the pattern according to the following equations: ${y}_{i}=a\mathrm{sin}\left({x}_{i}\right)$ ${x}_{i}=\frac{2\pi k}{n}i+\frac{\pi {\varphi }_{0}}{180}$ for i = 0, 1, 2, ..., n - 1 where • a is amplitude • k is cycles in the pattern • ${\varphi }_{0}$ is the initial phase • n is samples Where This Node Can Run: Desktop OS: Windows FPGA: Not supported Web Server: Not supported in VIs that run in a web application
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 3, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7690582871437073, "perplexity": 2220.472791177087}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-43/segments/1539583519859.77/warc/CC-MAIN-20181024042701-20181024064201-00066.warc.gz"}
https://infoscience.epfl.ch/record/184373
Infoscience Journal article Experimental investigation and thermodynamic modeling of the Au-Ge-Ni system Phase equilibria in the Au-Ge-Ni ternary system were studied by means of scanning electron microscopy, electron probe microanalysis, X-ray diffraction, and differential scanning calorimetry. The phase relations in the solid state at 600 A degrees C as well as a vertical section at Au72Ge28-Ni were established. No ternary compound was found at 600 A degrees C. On the basis of the experimental phase equilibria data, a thermodynamic model of the Au-Ge-Ni ternary system was developed using the CALPHAD method. Thermodynamically calculated phase diagrams are shown at 600 A degrees C, in two vertical sections and the liquidus projection. Reasonable agreement between the calculations and the experimental results was achieved.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.854954183101654, "perplexity": 2434.386451076426}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988719754.86/warc/CC-MAIN-20161020183839-00064-ip-10-171-6-4.ec2.internal.warc.gz"}
https://scicomp.stackexchange.com/tags/c%2B%2B/hot
# Tag Info 46 It's a bad idea because vector needs to allocate as many objects in space as there are rows in your matrix. Allocation is expensive, but primarily it is a bad idea because the data of your matrix now exists in a number of arrays scattered around memory, rather than all in one place where the processor cache can easily access it. It's also a wasteful storage ... 32 I'm also throwing my two cents in kind of late, but I've only just seen this thread and I feel that, for posterity, there are a few points that desperately need to be made. Note in the following that I will talk about C and not C++. Why? Well, otherwise it's apples and oranges to compare a full-fledged dynamically typed object-oriented language with ... 32 First, if your undergraduates are like ours and had no prior introduction to computers, expect to spend some time teaching them how to use basic stuff like using a proper editor (i.e., not MS Word), the command line, etc. I think the answer somewhat depends on where you set the focus of your course (or what you are required to teach). For example: How ... 22 In 2014, I would've said Python. In 2017, I wholeheartedly believe that the language to teach undergraduates is Julia. Teaching is always about a tradeoff. On one hand, you want to choose something that is simple enough that it is easy to grasp. But secondly, you want to teach something that has staying power, i.e. something that can grow with you. The ... 20 In addition to the reasons Wolfgang mentioned, if you use a vector<vector<double> >, you'll have to dereference it twice every time you want to retrieve an element, which is more computationally costly than a single dereferencing operation. One typical approach is to allocate a single array (a vector<double> or a double *) instead. I've ... 19 Let me try and break down your requirements: Maintainability Reading/writing text data Strong interfaces/capability for LU factorizations Sparse linear solvers Performance and scalability to large data From this list, I would consider the following languages: C, C++, Fortran, Python, MATLAB, Java Julia is a promising new language, but the community is ... 19 Possibly one could start with the function $\mathtt{expm1}$ which is part of the C99 standard, and calculates $e^x-1$ accurately near $x=0$. 17 If you're looking for a good bound on your rounding error, you don't necessarily need an aribtrary-precision library. You can use running error analysis instead. I wasn't able to find a good online reference, but it's all described in Section 3.3 of Nick Higham's book "Accuracy and Stability of Numerical Algorithms". The idea is quite simple: Re-factor ... 17 My problem with expression templates is that they are a very leaky abstraction. You spend a lot of work writing very complicated code to do a simple task with nicer syntax. But if you want to change the algorithm, you have to mess with the dirty code and if you slip up with types or syntax, you get completely unintelligible error messages. If your ... 17 Let me first answer why I think C++ interfaces to MPI have generally not been overly successful, having thought about the issue for a good long time when trying to decide whether we should just use the standard C bindings of MPI or building on something at higher level: When you look at real-world MPI codes (say, PETSc, or in my case deal.II), one finds ... 17 This is an instance of cancellation error. The C standard library (as of C99) includes a function called expm1 that avoids this problem. If you use expm1(x) / x instead of (exp(x) - 1.0) / x, you won't experience this issue (see graph below). The details and solution of this particular problem are discussed at length in Section 1.14.1 of Accuracy and ... 17 There seems to be quite a bit of confusion about how to apply multi-step (e.g. Runge-Kutta) methods to 2nd or higher order ODEs or systems of ODEs. The process is very simple once you understand it, but perhaps not obvious without a good explanation. The following method is the one I find simplest. In your case, the differential equation you would like to ... 16 I believe this discussion has come up a number of times on the PETSc list. My main reasons are: The C++ standard states that std::complex is only defined for the float, double, and long double datatypes. Thus it cannot be used for other datatypes, such as quad-precision. The standard makes no guarantees about the stability of the complex arithmetic. The ... 15 Let me give an example based on experience. Most libraries I use from a day to day basis use OOP in some way. OOP is able to hide the complexity required for many domains, it is not a mechanism that really helps with performance. What can happen is that a library is able to use specific optimizations based upon the object hierarchy, but for the most part ... 15 A difficulty with any of these types of questions is that the answer is highly community-dependent. To answer some of your questions in haphazard order: MATLAB is used a lot both in academia and in industry. One of the reasons it's used quite a bit in industry is because it is taught in academia. I know for a fact that MATLAB is used at Lincoln Laboratory ... 15 If you're doing celestial mechanics over long time scales, using a classical Runge-Kutta integrator will not preserve energy. In that case, using a symplectic integrator would probably be better. Boost.odeint also implements a 4th-order symplectic Runge-Kutta scheme that would work better for long time intervals. GSL does not implement any symplectic methods,... 14 I think by and large, template metaprogramming has been found to be unusable in practice -- it compiles too slow, and the error messages we get are just impossible to decipher. The barrier to entry for newcomers is also just too high when using metaprogramming. Of course, generic programming is an entirely different issue, as witnessed by Trilinos, deal.II (... 13 I have always thought that we should use it in our own project, deal.II, because it is higher level than pure MPI and can save a few lines of code here and there. That said, what I learned over the years is that most high-level code doesn't actually have that much MPI code to begin with -- the 600,000 lines of code in deal.II have only ~50 calls to MPI. That'... 13 I will address only the comparison of C to C++. While it is true that anything written in C can be ported to C++ with a few syntactic touch-ups, the communities have different values. The C library community, more than almost any other, values binary stability. Binary stability is critical for low-level libraries to avoid inflicting constant pain on those ... 13 In fundamental C++, I find the problem here is that C++ will allocate a new object of cx_mat to store evolutionMatrix*stateMatrix, and then copy the new object to stateMatrix with operator=(). I think you're right that it's creating temporaries, which is too slow, but I think the reason for why it's doing that is wrong. Armadillo, like any good C++ linear ... 13 If you have constants that will not change before runs, declare them in a header file: //constants.hpp #ifndef PROJECT_NAME_constants_hpp #define PROJECT_NAME_constants_hpp namespace constants { constexpr double G = 6.67408e-11; constexpr double M_EARTH = 5.972e24; constexpr double GM_EARTH = G*M_EARTH; } #endif //main.cpp using namespace ... 13 Feature test macros: HPC is generally stuck on old compilers or compilers with partially conformant implementations. This can help ease the pain of working on the custom architectures common in HPC. Example: #ifdef __cpp_lib_source_location #include <source_location> #endif ... #ifdef auto sl = std::source_location(); std::cerr << "Error at line ... 12 Others have commented on the issue of how difficult it is to write ET programs as well as the complexity of understanding error messages. Let me comment on the issue of compilers: It is true that a while back one of the big issues was finding a compiler that's compliant enough with the C++ standard to make everything work and make it work portably. As a ... 12 Another alternative that may be in line with your train of thought is to use a namespace (or nested namespaces) to properly group constants. An example might be: namespace constants { namespace earth { constexpr double G = 6.67408e-11; constexpr double Mass_Earth = 5.972e24; constexpr double GM = G*Mass_Earth; }// constant properties ... 12 I was able to reproduce the behavior reported in the question, and traced the observed inaccuracies to the following line: return y*sin(pi<Real>()*x)/pi<Real>(); The explicit multiplication with a floating-point approximation of π introduces a small error into the argument to sin, which comprises the representational error in the constant and ... 11 If you want to write something general-purpose, you can do it either with shell scripts if it is something very simple, as Pedro suggests, or aggregate in a higher-level mathematical programming language such as Python or MATLAB. I agree that plain text files are useful for smaller amounts of data, but you should probably switch to binary data for anything ... 11 OPT++ is used internally by Dakota (Sandia), which is much more than an optimization library and is released under the LGPL. You should also take a look at TAO (ANL), released under a BSD-like license. An introduction to both OPT++ and TAO can be found here. Other alternatives are MOOCHO, NOMAD, and HOPSPACK, which are, as far as I know, also LGPL licensed. ... 11 As Misha and Geoff Oxberry pointed out, Mathematica really has a different focus (just because you can pound in a nail with a screwdriver doesn't mean you should). So I take your question as being "If I know Matlab, why should I learn Python?" [Edit: and so, apparently, did you.] For all intents and purposes, Matlab is the English of scientific computing -- ... 11 Two examples of libraries that use modern C++ constructs: Both the eigen and armadillo libraries (linear algebra) use several modern C++ constructs. For instance, they use both expression templates to simplify arithmetic expressions and can sometimes eliminate some temporaries: http://eigen.tuxfamily.org http://arma.sourceforge.net/ http://hpac.rwth-... 11 It is true that compilers are getting better and better at auto-vectorization, and for basic coefficient-wise operations like 2*A-4*B a library like Eigen cannot do much better than recent compilers. However, for slightly more complicated expressions like matrix products, reductions, transposition, powers, etc. the compiler cannot do much. On the other hand, ... Only top voted, non community-wiki answers of a minimum length are eligible
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.47223684191703796, "perplexity": 1093.8447323203707}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370510287.30/warc/CC-MAIN-20200403030659-20200403060659-00354.warc.gz"}
http://www.koreascience.or.kr/article/JAKO201405262619015.page
# METRIC THEOREM AND HAUSDORFF DIMENSION ON RECURRENCE RATE OF LAURENT SERIES • Hu, Xue-Hai ; • Li, Bing ; • Xu, Jian • Published : 2014.01.31 • 74 22 #### Abstract We show that the recurrence rates of Laurent series about continued fractions almost surely coincide with their pointwise dimensions of the Haar measure. Moreover, let $E_{{\alpha},{\beta}}$ be the set of points with lower and upper recurrence rates ${\alpha},{\beta}$, ($0{\leq}{\alpha}{\leq}{\beta}{\leq}{\infty}$), we prove that all the sets $E_{{\alpha},{\beta}}$, are of full Hausdorff dimension. Then the recurrence sets $E_{{\alpha},{\beta}}$ have constant multifractal spectra. #### Keywords recurrence rate;pointwise dimension;continued fractions;Laurent series;Hausdorff dimension #### References 1. E. Artin, Quadratische Korper im Gebiete der hoheren Kongruenzen, I-II, Math. Z. 19 (1924), no. 1, 153-246. https://doi.org/10.1007/BF01181074 2. L. Barreira and B. Saussol, Hausdorff dimension of measures via Poincare recurrence, Comm. Math. Phys. 219 (2001), no. 2, 443-463. https://doi.org/10.1007/s002200100427 3. L. Barreira and B. Saussol, Product structure of Poincare recurrence, Ergodic Theory Dynam. Systems 22 (2002), no. 1, 33-61. 4. V. Berthe and H. Nakada, On continued fraction expansions in positive characteristic: equivalence relations and some metric properties, Expo. Math. 18 (2000), no. 4, 257-284. 5. M. Boshernitzan, Quantitative recurrence results, Invent. Math. 113 (1993), no. 3, 617-631. https://doi.org/10.1007/BF01244320 6. K. J. Falconer, Fractal Geometry, Mathematical Foundations and Application, Wiley, 1990. 7. D. J. Feng and J. Wu, The Hausdorff dimension of recurrent sets in symbolic spaces, Nonlinearity 14 (2001), no. 1, 81-85. https://doi.org/10.1088/0951-7715/14/1/304 8. X. H. Hu, B.W. Wang, J. Wu, and Y. L. Yu, Cantor sets determined by partial quotients of continued fractions of Laurent series, Finite Fields Appl. 14 (2008), no. 2, 417-437. https://doi.org/10.1016/j.ffa.2007.04.002 9. K. S. Lau and L. Shu, The spectrum of Poincare recurrence, Ergodic Theory Dynam. Systems 28 (2008), no. 6, 1917-1943. https://doi.org/10.1017/S0143385707001095 10. H. Niederreiter, The probabilistic theory of linear complexity, Advances in cryptology EUROCRYPT '88 (Davos, 1988), 191209, Lecture Notes in Comput. Sci., 330, Springer, Berlin, 1988. 11. H. Niederreiter and M. Vielhaber, Linear complexity profiles: Hausdorff dimensions for almost perfect profiles and measures for general profiles, J. Complexity. 13 (1997), no. 3, 353-383. https://doi.org/10.1006/jcom.1997.0451 12. D. Ornstein and B. Weiss, Entropy and data compression schemes, IEEE Trans. Inform. Theory 39 (1993), no. 1 78-83. https://doi.org/10.1109/18.179344 13. R. Paysant-Leroux and E. Dubois, Etude metrique de l'algorithme de Jacobi-Perron dans un corps de series formelles, (French) C. R. Acad. Sci. Paris Ser. A-B 275 (1972), A683-A686. 14. B. Saussol and J. Wu, Recurrence spectrum in smooth dynamical system, Nonlinearity 16 (2003), no. 6, 1991-2001. https://doi.org/10.1088/0951-7715/16/6/306 15. W. M. Schmidt, On continued fractions and Diophantine approximation in power series fields, Acta Arith. 95 (2000), no. 2, 139-166. https://doi.org/10.4064/aa-95-2-139-166 16. J. Wu, Hausdorff dimensions of bounded type continued fraction sets of Laurent series, Finite Fields Appl. 13 (2007), no. 1, 20-30. https://doi.org/10.1016/j.ffa.2005.05.003
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.915937602519989, "perplexity": 1736.718926096662}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627998100.52/warc/CC-MAIN-20190616102719-20190616124719-00172.warc.gz"}
http://blog.trifork.com/2009/12/09/mahout-taste-part-one-introduction/
# Mahout – Taste :: Part 1 – Introduction December 9th, 2009 by | This post is the first in a series on Taste, a Java framework for providing personalized recommendations. Taste is part of the larger Mahout framework, which features various scalable machine-learning algorithms. In this post I introduce you to the concepts of personalized recommendations, also known as collaborative filtering. After this introduction, Taste’s architecture and extension points are explained. I finish this post by demonstrating and explaining the TanimotoCoefficientSimilarity, one of Taste’s implementations used for computing recommendations. Personalized Recommendations Today the web is full of services for recommending books, websites, music, applications, movies and so on. Amazon, Last.fm and StumbleUpon, all provide these personalized recommendations for internet users. These features can be quite useful for customers and profitable for many e-commerce sites these days. Collaborative Filtering Let’s first review some basic concepts. The theory that powers all these useful websites mentioned above is the process of Collaborative Filtering, or pattern recognition in large datasets of multiple users. These datasets can contain preferences of users for certain items. For example, Youtube members can rate a video by assigning a number of stars. The number of stars is a user’s preference value, a value from 1 to 5. Based on this collection of personal preferences and a ‘similarity function’ you can recommend videos to users or determine similar users, users with similar taste in videos. In this case, recommending videos is an example of an ‘item-based recommendation’ and determining users with similar tastes is an example of an ‘user-based recommendation’. Introducing Mahout – Taste So how can we use this theory to build recommenders? This is where Mahout – Taste comes in. Mahout is a Java framework for running scalable machine learning algorithms on top of Hadoop. Taste is a sub-framework of Mahout for building recommendation engines. Since April 4th 2008, Taste has become part of Mahout. Below is a figure with Taste’s architecture and the building blocks you need to configure a recommender. Taste architecture diagram The main building block in Taste is the Recommender. The Recommender recommends items based on a given item or it determines users with similar tastes. It works as follows: the recommender applies a similarity function on a subset of pairs of items (or users) in the dataset. A similarity function usually returns a value between 0 and 1, with 1 representing two completely similar items and 0 completely dissimilar items. When the similarity function processes pairs in the dataset the resulting similarity values are collected and are either kept in memory or stored on the filesystem or a database. When the Java application requests a few recommendations for a given item, the Recommender returns the items with the highest similarity. The Recommender retrieves items and users through the DataModel abstraction. Taste contains DataModel implementations for retrieving and storing your dataset through the filesystem or a database. In addition, the DataModel provides methods that count the total number of users, total number of items, number of users that prefer a certain item, and many more functions. Similarity functions use these numbers to compute a similarity value for pairs of items or users. This will be shown later in the example of the TanimotoCoefficientSimilarity, which determines a ratio based on some of these figures. You can build a recommender with Taste by adding a DataModel and a similarity function to a Recommender. You can also define your own similarity function by extending UserSimilarity or ItemSimilarity to recommend users or items, respectively. Tanimoto coefficent similarity Taste contains around a dozen similarity algorithms you can choose from to build a recommender. For this introductory post I will explain Taste’s TanimotoCoefficientSimilarity, a relatively straightforward similarity algorithm that is widely used in chemo-informatics for discovering similarities between molecules. Let’s illustrate the algorithm in the context of a webshop. Suppose there are 3 customers, A, B and C and 5 products, numbered 1 up to 5. Say each customer has bought a few products. For this algorithm it does not matter how many products are purchased, only which products are purchased by which customer. See the table below. Customer purchases Intuitively you may see that the similarity between two products can be expressed by some ratio of purchases of customers. To be more precise, the tanimoto coefficient is computed by the following formula: $$!\frac{c}{a + b – c}$$ $$c =$$ Number of customers that purchased p1 and p2 $$a =$$ Number of customers that purchased p1 $$b =$$ Number of customers that purchased p2 This means that if many customers have bought products, the numerator will be higher and so will be the similarity value. Alternatively, if many people have bought p1 and many have bought p2, but very few people bought both, p1 and p2 are probably dissimilar. Below is a table with calculated tanimoto coefficients for each product pair: Tanimoto coefficients for all product pairs Demonstrating Taste In this section I demonstrate how to express the previous example with Taste.  If you like to try this example at home, download the brand new 0.2 mahout jar here and add it to your classpath. If you are using maven, add the following snippet to your pom.xml <dependency> <groupId>org.apache.mahout</groupId> <artifactId>mahout-core</artifactId> <version>0.2</version> </dependency> Below is a code snippet that shows how to build the DataModel and how to compute similarities with the TanimotoCoefficientSimilarity.  In the setup() method I create a BooleanPreferenceArray for each user. I then fill these arrays, put all of them in a FastByIdMap and put that in the DataModel. Next I create the TanimotoCoefficientSimilarity and pass in the DataModel. I then write a few tests that check whether the similarities computed by Taste return the values I expect, given the formula above. /** * Demonstrates TanimotoCoefficientSimilarity + recommender. * * @author Frank Scholten */ public class TanimotoDemo { private DataModel dataModel; private ItemSimilarity tanimoto; private long CUSTOMER_A = 0; private long CUSTOMER_B = 1; private long CUSTOMER_C = 2; private long productOne = 0; private long productTwo = 1; private long productThree = 2; private long productFour = 3; private long productFive = 4; @Before public void setup() { FastByIDMap<PreferenceArray> userIdMap = new FastByIDMap<PreferenceArray>(); BooleanUserPreferenceArray customerAPrefs = new BooleanUserPreferenceArray(4); customerAPrefs.set(0, new BooleanPreference(CUSTOMER_A, productOne)); customerAPrefs.set(1, new BooleanPreference(CUSTOMER_A, productTwo)); customerAPrefs.set(2, new BooleanPreference(CUSTOMER_A, productFour)); customerAPrefs.set(3, new BooleanPreference(CUSTOMER_A, productFive)); BooleanUserPreferenceArray customerBPrefs = new BooleanUserPreferenceArray(3); customerBPrefs.set(0, new BooleanPreference(CUSTOMER_B, productTwo)); customerBPrefs.set(1, new BooleanPreference(CUSTOMER_B, productThree)); customerBPrefs.set(2, new BooleanPreference(CUSTOMER_B, productFive)); BooleanUserPreferenceArray customerCPrefs = new BooleanUserPreferenceArray(2); customerCPrefs.set(0, new BooleanPreference(CUSTOMER_C, productOne)); customerCPrefs.set(1, new BooleanPreference(CUSTOMER_C, productFive)); userIdMap.put(CUSTOMER_A, customerAPrefs); userIdMap.put(CUSTOMER_B, customerBPrefs); userIdMap.put(CUSTOMER_C, customerCPrefs); dataModel = new GenericDataModel(userIdMap); tanimoto = new TanimotoCoefficientSimilarity(dataModel); } @Test public void testSimilarities() throws TasteException { assertEquals((double) 1, tanimoto.itemSimilarity(productOne, productOne), 0.01); assertEquals((double) 1 / 3, tanimoto.itemSimilarity(productOne, productTwo), 0.01); assertEquals((double) 0, tanimoto.itemSimilarity(productOne, productThree), 0.01); assertEquals((double) 1 / 2, tanimoto.itemSimilarity(productOne, productFour), 0.01); assertEquals((double) 2 / 3, tanimoto.itemSimilarity(productOne, productFive), 0.01); assertEquals((double) 1 / 1, tanimoto.itemSimilarity(productTwo, productTwo), 0.01); assertEquals((double) 1 / 2, tanimoto.itemSimilarity(productTwo, productThree), 0.01); assertEquals((double) 1 / 2, tanimoto.itemSimilarity(productTwo, productFour), 0.01); assertEquals((double) 2 / 3, tanimoto.itemSimilarity(productTwo, productFive), 0.01); assertEquals((double) 1, tanimoto.itemSimilarity(productThree, productThree), 0.01); assertEquals((double) 0, tanimoto.itemSimilarity(productThree, productFour), 0.01); assertEquals((double) 1 / 3, tanimoto.itemSimilarity(productThree, productFive), 0.01); assertEquals((double) 1, tanimoto.itemSimilarity(productFour, productFour), 0.01); assertEquals((double) 1 / 3, tanimoto.itemSimilarity(productFour, productFive), 0.01); assertEquals((double) 1, tanimoto.itemSimilarity(productFive, productFive), 0.01); } @Test public void testRecommendProducts() throws TasteException { ItemBasedRecommender recommender = new GenericItemBasedRecommender(dataModel, tanimoto); List<RecommendedItem> similarToProductThree = recommender.mostSimilarItems(productThree, 2); assertEquals(productTwo, similarToProductThree.get(0).getItemID()); assertEquals(productFive, similarToProductThree.get(1).getItemID()); } } I also added tests for the recommendations themselves. The testRecommendProducts() method uses mostSimilarItems to determine items similar to product 3, in terms of customer preferences. The second parameter of this method is the number of similar items to compute. The result of this method is a list of items ordered by their similarity value, descendingly. We can now predict and see that product two is most similar to product three and product five is second most similar. Note that the code above is only suitable for demonstrative purposes to demonstrate and understand the algorithm. For instance, in an actual a production ready implementation a database backed DataModel can be used, or the algorithm can be ran on Hadoop. This concludes this introduction to Taste. There are a lot of interesting parts that I haven’t explored myself, I am learning this one piece at a time, including the algorithms and mathematics behind it. Some things to check out: the similarity Javadocs for similarity algorithms. And of course there are classes to run your recommender on Hadoop and to evaluate the quality of your recommender with training sets. In the next few posts I will go into some of these subjects and use examples with a larger dataset, look into Hadoop and do a comparison of algorithms. References #### 14 Responses 1. December 9, 2009 at 12:11 by Andrew Phillips For the lazy amongst us…it would be convenient if you added the calculated tanimoto coefficients to your “customer purchases” example 😉 Thanks! 2. […] Mahout – Taste :: Part 1 – Introduction « JTeam Blog / JTeam: Enterprise Java, Open Source, sof…. […] 3. December 9, 2009 at 15:55 by Frank Scholten @Andrew, yes I can do that when I get around to it. 4. December 9, 2009 at 20:00 by Frank Scholten Just added a table with tanimoto coefficients for all product pairs 5. December 9, 2009 at 21:06 by Wouter Heijke Thanks for this informative Mahout introduction! I’ve been wishing to get started with Mahout for quite a while now and your article made me do it today! 6. December 10, 2009 at 09:58 by Frank Scholten @Wouter: Great, let us know what you learned. 7. December 10, 2009 at 13:14 by Wouter Heijke What i’ve noticed that it is easy to get an: org.apache.mahout.cf.taste.common.NoSuchUserException To clarify I think you should change the following: userIdMap.put(0, customerAPrefs); userIdMap.put(1, customerBPrefs); userIdMap.put(2, customerCPrefs); into: userIdMap.put(CUSTOMER_A, customerAPrefs); userIdMap.put(CUSTOMER_B, customerBPrefs); userIdMap.put(CUSTOMER_C, customerCPrefs); 8. December 11, 2009 at 12:57 by Frank Scholten @Wouter: Yes, good point. I updated the code. 9. December 17, 2009 at 12:00 by John Thank you for the detailed write-up. My company is developing a recommendation engine. Do you recommend using Mahout as our core or are there better open source collaborative filtering systems out there? Also, do you know of any websites that use Mahout to provide recommendations? I would like to see it in action. 10. December 18, 2009 at 15:09 by Frank Scholten @John: Mahout is the only open source collaborative filtering system I am researching at the moment. There is an active and growing community behind it, so I’d say give it a try. For a real-life example of Mahout, look into Mippin, a search engine with recommendation features. Check out http://blog.mippin.com/2009/09/may-we-recommend.html 11. January 12, 2010 at 16:41 by Nishant Nice article. Is there any real life dataset representing the boolean preferences? 12. July 9, 2010 at 18:31 by neil kodner Any chance of you updating the document to aid people who are unfamiliar with java? I’ve followed the instructions on downloading and building mahout from svn but am unsure how to get your example code to run. I believe I’m down to classpath issues but I can never be sure. My mahout trunk is in /Users/nkodner/development/mahout/trunk/ I’m not sure how to set up my classpath. I’m also not sure where to save TanimotoDemo.java. Any help is greatly appreciated! I’m looking to try Mahout, specifically TanimotoSimilarity on top of my Hadoop cluster to see how it performs against my Python implementation. 13. July 12, 2010 at 17:11 by Frank Scholten @Neil: It’s best to run this JUnit test from either IntelliJ, http://www.jetbrains.com/idea/download/, or Eclipse, http://www.eclipse.org/downloads/. Create a Java project, create a package taste and a class TanimotoDemo. Now add the mahout jars as dependencies to your project and you can run TanimotoDemo as a JUnit test. It’s possible to run it from the command line with java -cp all:your:dependencies:here org.junit.runner.JunitCore package.to.TanimotoDemo but you need to setup your classpath and an IDE makes that a lot easier. Note btw that this a non-distributed recommendation, it does not run on Hadoop. I will dive into this subject in the future and blog about it. 14. […] Mahout taste part 1 […]
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1845438927412033, "perplexity": 3061.11956460405}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218189242.54/warc/CC-MAIN-20170322212949-00536-ip-10-233-31-227.ec2.internal.warc.gz"}
https://www.varsitytutors.com/high_school_math-help/how-to-find-the-slope-of-a-perpendicular-line
# High School Math : How to find the slope of a perpendicular line ## Example Questions ### Example Question #1 : How To Find The Slope Of A Perpendicular Line A line passes through the points  and .  If a new line is drawn perpendicular to the original line, what will its slope be? Possible Answers: Correct answer: Explanation: The original line has a slope of , a line perpendicular to the original line will have a slope which is the negative reciprocal of this value. ### Example Question #1 : Perpendicular Lines Which of the following is the equation of a line that is perpendicular to the line  ? Possible Answers: Correct answer: Explanation: Perpendicular lines have slopes that are the opposite reciprocals of each other. Thus, we first identify the slope of the given line, which is  (since it is in the form , where  represents slope). Then, we know that any line which is perpendicular to this will have a slope of . Thus, we can determine that  is the only choice with the correct slope. ### Example Question #1 : How To Find The Slope Of A Perpendicular Line What will be the slope of the line perpendicular to ? Possible Answers: Correct answer: Explanation: In standard form, the is the slope. The slope of a perpendicular line is the negative reciprocal of the original line. For our given line, the slope is . Therefore, the slope of the perpendicular line is . ### Example Question #4 : How To Find The Slope Of A Perpendicular Line Which of the following is perpendicular to the line described by Possible Answers: Correct answer: Explanation: The definition of perpendicular lines is that their slopes are inverse reciprocals of one another. Since the slope in the given equation is , this means that the slope of its perpendicular line would be . The answer is the only equation listed that has a slope of . ### Example Question #5 : How To Find The Slope Of A Perpendicular Line Which of the following is perpendicular to the line described by Possible Answers: Correct answer: Explanation: The definition of perpendicular lines is that their slopes are inverse reciprocals of one another. Since the slope in the given equation is , this means that the slope of its perpendicular line would be . The answer is the only equation listed that has a slope of . ### Example Question #6 : How To Find The Slope Of A Perpendicular Line Which of the following gives the slope of a line that is perpendicular to  ? Possible Answers: Correct answer: Explanation: Recall that the slopes of perpendicular lines are opposite reciprocals of one another. As a result, we are looking for the opposite reciprocal of . Thus, we can get that the opposite reciprocal is ### Example Question #7 : How To Find The Slope Of A Perpendicular Line Find the slope of the line perpendicular to . Possible Answers: Correct answer: Explanation: Put this equation into slope-intercept form, y = mx + b, to find the slope, m. Do this by subtracting  from both sides of the equation: The slope of this line is . The slope of the perpendicular line is the negative reciprocal. Switch the numerator and denominator, and then multiply by : ### Example Question #8 : How To Find The Slope Of A Perpendicular Line What is the slope of the line perpendicular to ? Possible Answers: Correct answer: Explanation: In standard form, is the slope. The slope of a perpendicular line is the negative reciprocal of the original line. For our given line, the slope is . Therefore, the slope of the perpendicular line is . ### Example Question #9 : How To Find The Slope Of A Perpendicular Line Find the slope of this line: Possible Answers: Correct answer: Explanation: Isolate for  so that the equation now reads The slope is
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9626665115356445, "perplexity": 380.1814580387091}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250606226.29/warc/CC-MAIN-20200121222429-20200122011429-00081.warc.gz"}
http://math.duke.edu/events/79200-algebraic-de-rham-theory-relative-completion-mathrmsl2mathbbz
# Algebraic de Rham theory for relative completion of $\mathrm{SL}_2(\mathbb{Z})$ #### Ma Luo (Duke University) Wednesday, December 6, 2017 - 3:15pm to 4:15pm Location: 119 Physics In this talk, I will first review relative (unipotent) completions of discrete groups in general, and $\mathrm{SL}_2(\mathbb{Z})$ in particular. We then develop an explicit $\mathbb{Q}$-de Rham theory for the relative completion of $\mathrm{SL}_2(\mathbb{Z})$, which enables us to construct iterated integrals of modular forms of the second kind that provide its periods. Following Francis Brown, these periods are called `multiple modular values'. They contain periods of modular forms. Last updated: 2018/02/24 - 4:21am
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5772004127502441, "perplexity": 2381.0876652726824}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891815843.84/warc/CC-MAIN-20180224152306-20180224172306-00062.warc.gz"}
http://www.science.gov/topicpages/w/waste+licl+salt.html
#### Sample records for waste licl salt 1. Study on LiCl waste salt treatment process by layer melt crystallization SciTech Connect Cho, Yung-Zun; Lee, Tae-Kyo; Choi, Jung-Hoon; Eun, Hee-Chul; Park, Hwan-Seo; Kim, In-Tae; Park, Geun-Il 2013-07-01 Layer melt crystallization operated in a static mode has been applied to separate Group I and II chlorides from surrogate LiCl waste salt. The effects of operating conditions such as crystal growing rate(or flux) and initial impurity concentration on separation (or concentration) of cesium, strontium and barium involved in a LiCl melts were analyzed. In a layer crystallization process, separation was impaired by occlusion of impurities and by residual melt adhering to LiCl crystal after at the end of the process. The crystal growth rate strongly affects the crystal structure, therefore the separation efficiency, while the effect of the initial Cs and Sr concentration in LiCl molten salt was nearly negligible. (authors) 2. Inclusion property of Cs, Sr, and Ba impurities in LiCl crystal formed by layer-melt crystallization SciTech Connect Choi, Jung-Hoon; Cho, Yung-Zun; Lee, Tae-Kyo; Eun, Hee-Chul; Kim, Jun-Hong; Park, Hwan-Seo; Kim, In-Tae; Park, Geun-Il 2013-07-01 Pyroprocessing is one of the promising technologies enabling the recycling of spent nuclear fuels from a commercial light water reactor (LWR). In general, pyroprocessing uses dry molten salts as electrolytes. In particular, LiCl waste salt after pyroprocessing contains highly radioactive I/II group fission products mainly composed of Cs, Sr, and Ba impurities. Therefore, it is beneficial to reuse LiCl salt in the pyroprocessing as an electrolyte for economic and environmental issues. Herein, to understand the inclusion property of impurities within LiCl crystal, the physical properties such as lattice parameter change, bulk modulus, and substitution enthalpy of a LiCl crystal having 0-6 at% Cs{sup +} or Ba{sup 2+} impurities under existence of 1 at% Sr{sup 2+} impurity were calculated via the first-principles density functional theory. The substitution enthalpy of LiCl crystals having 1 at% Sr{sup 2+} showed slightly decreased value than those without Sr{sup 2+} impurity. Therefore, through the substitution enthalpy calculation, it is expected that impurities will be incorporated within LiCl crystal as co-existed form rather than as a single component form. (authors) 3. Management of Salt Waste from Electrochemical Processing of Used Nuclear Fuel SciTech Connect Michael F. Simpson; Michael N. Patterson; Joon Lee; Yifeng Wang; Joshua Versey; Ammon Williams; Supathorn Phongikaroon; James Allensworth; Man-Sung Yim 2013-10-01 Electrochemical processing of used nuclear fuel involves operation of one or more cells containing molten salt electrolyte. Processing of the fuel results in contamination of the salt via accumulation of fission products and transuranic (TRU) actinides. Upon reaching contamination limits, the salt must be removed and either disposed or treated to remove the contaminants and recycled back to the process. During development of the Experimental Breeder Reactor-II spent fuel treatment process, waste salt from the electrorefiner was to be stabilized in a ceramic waste form and disposed of in a high-level waste repository. With the cancellation of the Yucca Mountain high-level waste repository, other options are now being considered. One approach that involves direct disposal of the salt in a geologic salt formation has been evaluated. While waste forms such as the ceramic provide near-term resistance to corrosion, they may not be necessary to ensure adequate performance of the repository. To improve the feasibility of direct disposal, recycling a substantial fraction of the useful salt back to the process equipment could minimize the volume of the waste. Experiments have been run in which a cold finger is used for this purpose to crystallize LiCl from LiCl/CsCl. If it is found to be unsuitable for transportation, the salt waste could also be immobilized in zeolite without conversion to the ceramic waste form. 4. Management of salt waste from electrochemical processing of used nuclear fuel SciTech Connect Simpson, M.F.; Patterson, M.N.; Lee, J.; Wang, Y.; Versey, J.; Phongikaroon, S. 2013-07-01 Electrochemical processing of used nuclear fuel involves operation of one or more cells containing molten salt electrolyte. Processing of the fuel results in contamination of the salt via accumulation of fission products and transuranic (TRU) actinides. Upon reaching contamination limits, the salt must be removed and either disposed or treated to remove the contaminants and recycled back to the process. During development of the Experimental Breeder Reactor-II spent fuel treatment process, waste salt from the electro-refiner was to be stabilized in a ceramic waste form and disposed of in a high-level waste repository. With the cancellation of the Yucca Mountain high-level waste repository, other options are now being considered. One approach that involves direct disposal of the salt in a geologic salt formation has been evaluated. While waste forms such as the ceramic provide near-term resistance to corrosion, they may not be necessary to ensure adequate performance of the repository. To improve the feasibility of direct disposal, recycling a substantial fraction of the useful salt back to the process equipment could minimize the volume of the waste. Experiments have been run in which a cold finger is used for this purpose to crystallize LiCl from LiCl/CsCl. If it is found to be unsuitable for transportation, the salt waste could also be immobilized in zeolite without conversion to the ceramic waste form. (authors) 5. Organic waste processing using molten salt oxidation SciTech Connect Adamson, M. G., LLNL 1998-03-01 Molten Salt Oxidation (MSO) is a thermal means of oxidizing (destroying) the organic constituents of mixed wastes, hazardous wastes, and energetic materials while retaining inorganic and radioactive constituents in the salt. For this reason, MSO is considered a promising alternative to incineration for the treatment of a variety of organic wastes. The U. S. Department of Energys Office of Environmental Management (DOE/EM) is currently funding research that will identify alternatives to incineration for the treatment of organic-based mixed wastes. (Mixed wastes are defined as waste streams which have both hazardous and radioactive properties.) One such project is Lawrence Livermore National Laboratorys Expedited Technology Demonstration of Molten Salt Oxidation (MSO). The goal of this project is to conduct an integrated demonstration of MSO, including off-gas and spent salt treatment, and the preparation of robust solid final forms. Livermore National Laboratory (LLNL) has constructed an integrated pilot-scale MSO treatment system in which tests and demonstrations are presently being performed under carefully controlled (experimental) conditions. The system consists of a MSO process vessel with dedicated off-gas treatment, a salt recycle system, feed preparation equipment, and equipment for preparing ceramic final waste forms. In this paper we describe the integrated system and discuss its capabilities as well as preliminary process demonstration data. A primary purpose of these demonstrations is to identify the most suitable waste streams and waste types for MSO treatment. 6. Alternative Waste Forms for Electro-Chemical Salt Waste SciTech Connect Crum, Jarrod V.; Sundaram, S. K.; Riley, Brian J.; Matyas, Josef; Arreguin, Shelly A.; Vienna, John D. 2009-10-28 This study was undertaken to examine alternate crystalline (ceramic/mineral) and glass waste forms for immobilizing spent salt from the Advanced Fuel Cycle Initiative (AFCI) electrochemical separations process. The AFCI is a program sponsored by U.S. Department of Energy (DOE) to develop and demonstrate a process for recycling spent nuclear fuel (SNF). The electrochemical process is a molten salt process for the reprocessing of spent nuclear fuel in an electrorefiner and generates spent salt that is contaminated with alkali, alkaline earths, and lanthanide fission products (FP) that must either be cleaned of fission products or eventually replaced with new salt to maintain separations efficiency. Currently, these spent salts are mixed with zeolite to form sodalite in a glass-bonded waste form. The focus of this study was to investigate alternate waste forms to immobilize spent salt. On a mole basis, the spent salt is dominated by alkali and Cl with minor amounts of alkaline earth and lanthanides. In the study reported here, we made an effort to explore glass systems that are more compatible with Cl and have not been previously considered for use as waste forms. In addition, alternate methods were explored with the hope of finding a way to produce a sodalite that is more accepting of as many FP present in the spent salt as possible. This study was done to investigate two different options: (1) alternate glass families that incorporate increased concentrations of Cl; and (2) alternate methods to produce a mineral waste form. 7. Disposal of NORM waste in salt caverns SciTech Connect Veil, J.A.; Smith, K.P.; Tomasko, D.; Elcock, D.; Blunt, D.; Williams, G.P. 1998-07-01 Some types of oil and gas production and processing wastes contain naturally occurring radioactive materials (NORM). If NORM is present at concentrations above regulatory levels in oil field waste, the waste requires special disposal practices. The existing disposal options for wastes containing NORM are limited and costly. This paper evaluates the legality, technical feasibility, economics, and human health risk of disposing of NORM-contaminated oil field wastes in salt caverns. Cavern disposal of NORM waste is technically feasible and poses a very low human health risk. From a legal perspective, there are no fatal flaws that would prevent a state regulatory agency from approving cavern disposal of NORM. On the basis of the costs charged by caverns currently used for disposal of nonhazardous oil field waste (NOW), NORM waste disposal caverns could be cost competitive with existing NORM waste disposal methods when regulatory agencies approve the practice. 8. Supplemental Cooling for Nitrate Salt Waste SciTech Connect Goldberg, Mitchell S. 2015-08-19 In July 2015, Los Alamos National Laboratory completed installation of a supplemental cooling system in the structure where remediated nitrate salt waste drums are stored. Although the waste currently is in a safe configuration and is monitored daily,controlling the temperature inside the structure adds another layer of protection for workers, the public,and the environment.This effort is among several layers of precautions designed to secure the waste. 9. Fuel production from wastes using molten salts SciTech Connect Gay, R.L.; Barclay, K.M.; Grantham, L.F.; Yosim, S.J. 1980-01-01 The Rockwell International molten salt process for gasification of wastes with resource recovery has been shown here to be well-suited for the processing of a variety of wastes. A variety of waste forms may be processed, that is, solids, liquids, and solid-liquid mixtures. The process is suitable for applications which involve either small or large throughputs. The gasification medium, sodium carbonate, is stable, non-volatile, inexpensive, and nontoxic. Sulfur-containing pollutants are retained in the melt when sulfur-containing wastes are gasified. In the same manner, halogen-containing pollutants are retained during gasification of halogen-containing wastes. The gasification of a high-nitrogen-content waste (leather scraps) produces very little NO/sub x/ in the off-gas. Valuable minerals may be recovered by processing of the salt after gasification of mineral-laden wastes. In general, the molten salt process is best applied to waste materials involving potential pollutants (such as sulfur or chromium) or to wastes where gasification and resource recovery are important (such as the recovery of silver with simultaneous gasification of x-ray film). 10. Simulation of salt waste evaporation/crystallization SciTech Connect Orebaugh, E.G. 1993-01-22 The database of ProChem software has been enhanced to account for the formation of the mineral, Burkite which can form in alkaline tank wastes during evaporation. This mineral was not suspected until recent evaporation/crystallization studies suggested its presence. The enhanced data base will predict its occurrence and realm of existence. If salt cake temperatures drop below 30{degrees}C the Burkite phase is unstable toward hydrated sodium carbonates and sulfates. ProChem will not predict if this phase is more or less rapidly dissolved than its component salts. The enhanced database improves our ability to simulate waste chemistry. 11. Simulation of salt waste evaporation/crystallization SciTech Connect Orebaugh, E.G. 1993-01-22 The database of ProChem software has been enhanced to account for the formation of the mineral, Burkite which can form in alkaline tank wastes during evaporation. This mineral was not suspected until recent evaporation/crystallization studies suggested its presence. The enhanced data base will predict its occurrence and realm of existence. If salt cake temperatures drop below 30[degrees]C the Burkite phase is unstable toward hydrated sodium carbonates and sulfates. ProChem will not predict if this phase is more or less rapidly dissolved than its component salts. The enhanced database improves our ability to simulate waste chemistry. 12. Molten salt destruction of energetic waste materials DOEpatents Brummond, W.A.; Upadhye, R.S.; Pruneda, C.O. 1995-07-18 A molten salt destruction process is used to treat and destroy energetic waste materials such as high explosives, propellants, and rocket fuels. The energetic material is pre-blended with a solid or fluid diluent in safe proportions to form a fluid fuel mixture. The fuel mixture is rapidly introduced into a high temperature molten salt bath. A stream of molten salt is removed from the vessel and may be recycled as diluent. Additionally, the molten salt stream may be pumped from the reactor, circulated outside the reactor for further processing, and delivered back into the reactor or cooled and circulated to the feed delivery system to further dilute the fuel mixture entering the reactor. 4 figs. 13. Molten salt destruction of energetic waste materials DOEpatents Brummond, William A. (Livermore, CA); Upadhye, Ravindra S. (Pleasanton, CA); Pruneda, Cesar O. (Livermore, CA) 1995-01-01 A molten salt destruction process is used to treat and destroy energetic waste materials such as high explosives, propellants, and rocket fuels. The energetic material is pre-blended with a solid or fluid diluent in safe proportions to form a fluid fuel mixture. The fuel mixture is rapidly introduced into a high temperature molten salt bath. A stream of molten salt is removed from the vessel and may be recycled as diluent. Additionally, the molten salt stream may be pumped from the reactor, circulated outside the reactor for further processing, and delivered back into the reactor or cooled and circulated to the feed delivery system to further dilute the fuel mixture entering the reactor. 14. Cementitious Stabilization of Mixed Wastes with High Salt Loadings SciTech Connect Spence, R.D.; Burgess, M.W.; Fedorov, V.V.; Downing, D.J. 1999-04-01 15. Modeling Solute Thermokinetics in LiCI-KCI Molten Salt for Nuclear Waste Separation SciTech Connect Morgan, Dane; Eapen, Jacob 2013-10-01 Recovery of actinides is an integral part of a closed nuclear fuel cycle. Pyrometallurgical nuclear fuel recycling processes have been developed in the past for recovering actinides from spent metallic and nitride fuels. The process is essentially to dissolve the spent fuel in a molten salt and then extract just the actinides for reuse in a reactor. Extraction is typically done through electrorefining, which involves electrochemical reduction of the dissolved actinides and plating onto a cathode. Knowledge of a number of basic thermokinetic properties of salts and salt-fuel mixtures is necessary for optimizing present and developing new approaches for pyrometallurgical waste processing. The properties of salt-fuel mixtures are presently being studied, but there are so many solutes and varying concentrations that direct experimental investigation is prohibitively time consuming and expensive (particularly for radioactive elements like Pu). Therefore, there is a need to reduce the number of required experiments through modeling of salt and salt-fuel mixture properties. This project will develop first-principles-based molecular modeling and simulation approaches to predict fundamental thermokinetic properties of dissolved actinides and fission products in molten salts. The focus of the proposed work is on property changes with higher concentrations (up to 5 mol%) of dissolved fuel components, where there is still very limited experimental data. The properties predicted with the modeling will be density, which is used to assess the amount of dissolved material in the salt; diffusion coefficients, which can control rates of material transport during separation; and solute activity, which determines total solubility and reduction potentials used during electrorefining. The work will focus on La, Sr, and U, which are chosen to include the important distinct categories of lanthanides, alkali earths, and actinides, respectively. Studies will be performed using LiCl-KCl salt at the eutectic composition (58 mol% LiCl, 42 mol% KCl), which is used for treating spent EBR-II fuel. The same process being used for EBRII fuel is currently being studied for widespread international implementation. The methods will focus on first-principles and first- principles derived interatomic potential based simulations, primarily using molecular dynamics. Results will be validated against existing literature and parallel ongoing experimental efforts. The simulation results will be of value for interpreting experimental results, validating analytical models, and for optimizing waste separation by potentially developing new salt configurations and operating conditions. 16. Delivery system for molten salt oxidation of solid waste DOEpatents Brummond, William A. (Livermore, CA); Squire, Dwight V. (Livermore, CA); Robinson, Jeffrey A. (Manteca, CA); House, Palmer A. (Walnut Creek, CA) 2002-01-01 The present invention is a delivery system for safety injecting solid waste particles, including mixed wastes, into a molten salt bath for destruction by the process of molten salt oxidation. The delivery system includes a feeder system and an injector that allow the solid waste stream to be accurately metered, evenly dispersed in the oxidant gas, and maintained at a temperature below incineration temperature while entering the molten salt reactor. 17. Cerebral salt wasting: pathophysiology, diagnosis, and treatment. PubMed Yee, Alan H; Burns, Joseph D; Wijdicks, Eelco F M 2010-04-01 Cerebral salt wasting (CSW) is a syndrome of hypovolemic hyponatremia caused by natriuresis and diuresis. The mechanisms underlying CSW have not been precisely delineated, although existing evidence strongly implicates abnormal elevations in circulating natriuretic peptides. The key in diagnosis of CSW lies in distinguishing it from the more common syndrome of inappropriate secretion of antidiuretic hormone. Volume status, but not serum and urine electrolytes and osmolality, is crucial for making this distinction. Volume and sodium repletion are the goals of treatment of patients with CSW, and this can be performed using some combination of isotonic saline, hypertonic saline, and mineralocorticoids. PMID:20380974 18. Waste Isolation Pilot Plant Salt Decontamination Testing SciTech Connect Rick Demmer; Stephen Reese 2014-09-01 On February 14, 2014, americium and plutonium contamination was released in the Waste Isolation Pilot Plant (WIPP) salt caverns. At the request of WIPP’s operations contractor, Idaho National Laboratory (INL) personnel developed several methods of decontaminating WIPP salt, using surrogate contaminants and also americium (241Am). The effectiveness of the methods is evaluated qualitatively, and to the extent possible, quantitatively. One of the requirements of this effort was delivering initial results and recommendations within a few weeks. That requirement, in combination with the limited scope of the project, made in-depth analysis impractical in some instances. Of the methods tested (dry brushing, vacuum cleaning, water washing, strippable coatings, and mechanical grinding), the most practical seems to be water washing. Effectiveness is very high, and it is very easy and rapid to deploy. The amount of wastewater produced (2 L/m2) would be substantial and may not be easy to manage, but the method is the clear winner from a usability perspective. Removable surface contamination levels (smear results) from the strippable coating and water washing coupons found no residual removable contamination. Thus, whatever is left is likely adhered to (or trapped within) the salt. The other option that shows promise is the use of a fixative barrier. Bartlett Nuclear, Inc.’s Polymeric Barrier System (PBS) proved the most durable of the coatings tested. The coatings were not tested for contaminant entrapment, only for coating integrity and durability. 19. Salt disposal of heat-generating nuclear waste. SciTech Connect Leigh, Christi D.; Hansen, Francis D. 2011-01-01 This report summarizes the state of salt repository science, reviews many of the technical issues pertaining to disposal of heat-generating nuclear waste in salt, and proposes several avenues for future science-based activities to further the technical basis for disposal in salt. There are extensive salt formations in the forty-eight contiguous states, and many of them may be worthy of consideration for nuclear waste disposal. The United States has extensive experience in salt repository sciences, including an operating facility for disposal of transuranic wastes. The scientific background for salt disposal including laboratory and field tests at ambient and elevated temperature, principles of salt behavior, potential for fracture damage and its mitigation, seal systems, chemical conditions, advanced modeling capabilities and near-future developments, performance assessment processes, and international collaboration are all discussed. The discussion of salt disposal issues is brought current, including a summary of recent international workshops dedicated to high-level waste disposal in salt. Lessons learned from Sandia National Laboratories' experience on the Waste Isolation Pilot Plant and the Yucca Mountain Project as well as related salt experience with the Strategic Petroleum Reserve are applied in this assessment. Disposal of heat-generating nuclear waste in a suitable salt formation is attractive because the material is essentially impermeable, self-sealing, and thermally conductive. Conditions are chemically beneficial, and a significant experience base exists in understanding this environment. Within the period of institutional control, overburden pressure will seal fractures and provide a repository setting that limits radionuclide movement. A salt repository could potentially achieve total containment, with no releases to the environment in undisturbed scenarios for as long as the region is geologically stable. Much of the experience gained from United States repository development, such as seal system design, coupled process simulation, and application of performance assessment methodology, helps define a clear strategy for a heat-generating nuclear waste repository in salt. 20. Mixed Waste Salt Encapsulation Using Polysiloxane - Final Report SciTech Connect Miller, C.M.; Loomis, G.G.; Prewett, S.W. 1997-11-01 A proof-of-concept experimental study was performed to investigate the use of Orbit Technologies polysiloxane grouting material for encapsulation of U.S. Department of Energy mixed waste salts leading to a final waste form for disposal. Evaporator pond salt residues and other salt-like material contaminated with both radioactive isotopes and hazardous components are ubiquitous in the DOE complex and may exceed 250,000,000 kg of material. Current treatment involves mixing low waste percentages (less than 10% by mass salt) with cement or costly thermal treatment followed by cementation to the ash residue. The proposed technology involves simple mixing of the granular salt material (with relatively high waste loadings-greater than 50%) in a polysiloxane-based system that polymerizes to form a silicon-based polymer material. This study involved a mixing study to determine optimum waste loadings and compressive strengths of the resultant monoliths. Following the mixing study, durability testing was performed on promising waste forms. Leaching studies including the accelerated leach test and the toxicity characteristic leaching procedure were also performed on a high nitrate salt waste form. In addition to this testing, the waste form was examined by scanning electron microscope. Preliminary cost estimates for applying this technology to the DOE complex mixed waste salt problem is also given. 1. Effects of Heat Generation on Nuclear Waste Disposal in Salt NASA Astrophysics Data System (ADS) Clayton, D. J. 2008-12-01 Disposal of nuclear waste in salt is an established technology, as evidenced by the successful operations of the Waste Isolation Pilot Plant (WIPP) since 1999. The WIPP is located in bedded salt in southeastern New Mexico and is a deep underground facility for transuranic (TRU) nuclear waste disposal. There are many advantages for placing radioactive wastes in a geologic bedded-salt environment. One desirable mechanical characteristic of salt is that it flows plastically with time ("creeps"). The rate of salt creep is a strong function of temperature and stress differences. Higher temperatures and deviatoric stresses increase the creep rate. As the salt creeps, induced fractures may be closed and eventually healed, which then effectively seals the waste in place. With a backfill of crushed salt emplaced around the waste, the salt creep can cause the crushed salt to reconsolidate and heal to a state similar to intact salt, serving as an efficient seal. Experiments in the WIPP were conducted to investigate the effects of heat generation on the important phenomena and processes in and around the repository (Munson et al. 1987; 1990; 1992a; 1992b). Brine migration towards the heaters was induced from the thermal gradient, while salt creep rates showed an exponential dependence on temperature. The project "Backfill and Material Behavior in Underground Salt Repositories, Phase II" (BAMBUS II) studied the crushed salt backfill and material behavior with heat generation at the Asse mine located near Remlingen, Germany (Bechthold et al. 2004). Increased salt creep rates and significant reconsolidation of the crushed salt were observed at the termination of the experiment. Using the data provided from both projects, exploratory modeling of the thermal-mechanical response of salt has been conducted with varying thermal loading and waste spacing. Increased thermal loading and decreased waste spacing drive the system to higher temperatures, while both factors are desired to reduce costs, as well as decrease the overall footprint of the repository. Higher temperatures increase the rate of salt creep which then effectively seals the waste quicker. Data of the thermal-mechanical response of salt at these higher temperatures is needed to further validate the exploratory modeling and provide meaningful constraints on the repository design. Sandia is a multi program laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy's National Nuclear Security Administration under Contract DE-AC04- 94AL85000. 2. HyponatremiaWhat Is Cerebral Salt Wasting? PubMed Central Momi, Jasminder; Tang, Christopher M; Abcar, Antoine C; Kujubu, Dean A; Sim, John J 2010-01-01 Background: Hyponatremia is a common electrolyte imbalance in hospitalized patients. It is associated with significant morbidity and mortality, especially if the underlying cause is incorrectly diagnosed and not treated appropriately. Often, the hospitalist is faced with a clinical dilemma when a patient presents with hyponatremia of an unclear etiology and with uncertain volume status. Syndrome of inappropriate antidiuretic hormone (SIADH) is frequently diagnosed in this clinical setting, but cerebral salt wasting (CSW) is an important diagnosis to consider. Objective: We wanted to describe the diagnosis, treatment, and history of CSW to provide clinicians with a better understanding of the differential diagnosis for hyponatremia. Conclusion: CSW is a process of extracellular volume depletion due to a tubular defect in sodium transport. Two postulated mechanisms for CSW are the excess secretion of natriuretic peptides and the loss of sympathetic stimulation to the kidney. Making the distinction between CSW and SIADH is important because the treatment for the two conditions is very different. PMID:20740122 3. Vitrification of IFR and MSBR halide salt reprocessing wastes SciTech Connect Siemer, D.D. 2013-07-01 Both of the genuinely sustainable (breeder) nuclear fuel cycles (IFR - Integral Fast Reactor - and MSBR - Molten Salt Breeder Reactor -) studied by the USA's national laboratories would generate high level reprocessing waste (HLRW) streams consisting of a relatively small amount ( about 4 mole %) of fission product halide (chloride or fluoride) salts in a matrix comprised primarily (about 95 mole %) of non radioactive alkali metal halide salts. Because leach resistant glasses cannot accommodate much of any of the halides, most of the treatment scenarios previously envisioned for such HLRW have assumed a monolithic waste form comprised of a synthetic analog of an insoluble crystalline halide mineral. In practice, this translates to making a 'substituted' sodalite ('Ceramic Waste Form') of the IFR's chloride salt-based wastes and fluoroapatite of the MSBR's fluoride salt-based wastes. This paper discusses my experimental studies of an alternative waste management scenario for both fuel cycles that would separate/recycle the waste's halide and immobilize everything else in iron phosphate (Fe-P) glass. It will describe both how the work was done and what its results indicate about how a treatment process for both of those wastes should be implemented (fluoride and chloride behave differently). In either case, this scenario's primary advantages include much higher waste loadings, much lower overall cost, and the generation of a product (glass) that is more consistent with current waste management practices. (author) 4. Expected brine movement at potential nuclear waste repository salt sites SciTech Connect McCauley, V.S.; Raines, G.E. 1987-08-01 The BRINEMIG brine migration code predicts rates and quantities of brine migration to a waste package emplaced in a high-level nuclear waste repository in salt. The BRINEMIG code is an explicit time-marching finite-difference code that solves a mass balance equation and uses the Jenks equation to predict velocities of brine migration. Predictions were made for the seven potentially acceptable salt sites under consideration as locations for the first US high-level nuclear waste repository. Predicted total quantities of accumulated brine were on the order of 1 m/sup 3/ brine per waste package or less. Less brine accumulation is expected at domal salt sites because of the lower initial moisture contents relative to bedded salt sites. Less total accumulation of brine is predicted for spent fuel than for commercial high-level waste because of the lower temperatures generated by spent fuel. 11 refs., 36 figs., 29 tabs. 5. Method for immobilizing mixed waste chloride salts containing radionuclides and other hazardous wastes DOEpatents Lewis, Michele A. (Naperville, IL); Johnson, Terry R. (Wheaton, IL) 1993-01-01 The invention is a method for the encapsulation of soluble radioactive waste chloride salts containing radionuclides such as strontium, cesium and hazardous wastes such as barium so that they may be permanently stored without future threat to the environment. The process consists of contacting the salts containing the radionuclides and hazardous wastes with certain zeolites which have been found to ion exchange with the radionuclides and to occlude the chloride salts so that the resulting product is leach resistant. 6. Method for immobilizing mixed waste chloride salts containing radionuclides and other hazardous wastes DOEpatents Lewis, Michele A.; Johnson, Terry R. 1993-09-07 The invention is a method for the encapsulation of soluble radioactive waste chloride salts containing radionuclides such as strontium, cesium and hazardous wastes such as barium so that they may be permanently stored without future threat to the environment. The process consists of contacting the salts containing the radionuclides and hazardous wastes with certain zeolites which have been found to ion exchange with the radionuclides and to occlude the chloride salts so that the resulting product is leach resistant. 7. Salt Waste Processing Pilot Facility Feed Basis for Design SciTech Connect Pike, J.A. 2001-10-09 This paper presents the characterization of potential feed to the Salt Waste Processing Facility pilot facility for design purposes and supercedes any previous representations of feed to the pilot facility. 8. Characterization of a ceramic waste form encapsulating radioactive electrorefiner salt SciTech Connect Moschetti, T. L.; Sinkler, W.; DiSanto, T.; Noy, M.; Warren, A. R.; Cummings, D. G.; Johnson, S. G.; Goff, K. M.; Bateman, K. J.; Frank, S. M. 1999-11-11 Argonne National Laboratory has developed a ceramic waste form to immobilize radioactive waste salt produced during the electrometallurgical treatment of spent fuel. This study presents the first results from electron microscopy and durability testing of a ceramic waste form produced from that radioactive electrorefiner salt. The waste form consists of two primary phases: sodalite and glass. The sodalite phase appears to incorporate most of the alkali and alkaline earth fission products. Other fission products (rare earths and yttrium) tend to form a separate phase and are frequently associated with the actinides, which form mixed oxides. Seven-day leach test results are also presented. 9. Membrane Treatment of Liquid Salt Bearing Radioactive Wastes SciTech Connect Dmitriev, S. A.; Adamovich, D. V.; Demkin, V. I.; Timofeev, E. M. 2003-02-25 The main fields of introduction and application of membrane methods for preliminary treatment and processing salt liquid radioactive waste (SLRW) can be nuclear power stations (NPP) and enterprises on atomic submarines (AS) utilization. Unlike the earlier developed technology for the liquid salt bearing radioactive waste decontamination and concentrating this report presents the new enhanced membrane technology for the liquid salt bearing radioactive waste processing based on the state-of-the-art membrane unit design, namely, the filtering units equipped with the metal-ceramic membranes of ''TruMem'' brand, as well as the electrodialysis and electroosmosis concentrators. Application of the above mentioned units in conjunction with the pulse pole changer will allow the marked increase of the radioactive waste concentrating factor and the significant reduction of the waste volume intended for conversion into monolith and disposal. Besides, the application of the electrodialysis units loaded with an ion exchange material at the end polishing stage of the radioactive waste decontamination process will allow the reagent-free radioactive waste treatment that meets the standards set for the release of the decontaminated liquid radioactive waste effluents into the natural reservoirs of fish-farming value. 10. Stabilization of high salt wastes by polymer extrusion SciTech Connect Faucette, A.M.; Getty, R.H.; Jantzen, R.C. 1995-12-31 The Rocky Flats Site treats large volumes of nitric acid waste which, upon neutralization, yields a high volume aqueous waste stream containing predominantly nitrate compounds. The cessation of plutonium processing, however, has significantly reduced the concentration of nitrates and increased the concentrations of other anions in the wastes. The changes in the waste stream chemistry have resulted in problems with the current treatment process which mixes spray dried salt with cement and brine. Polymer extrusion technologies are being investigated asan alternative to cementation. Spray dried salts are introduced into a twin screw extruder and mixed with a molten, low density polyethylene. The molten output is cast in a drum and allowed to cool. Toxicity Characteristic Leaching Procedure (TCLP) testing results for 40% and 50% waste loadings using this method are provided. All samples passed Universal Treatment Standard requirements for leachable metal concentrations. Based on these treatability study results, and favorable results from pilot-scale tests using surrogate salts, pilot-scale tests with actual salt waste are being planned. 11. Molten salt processing of mixed wastes with offgas condensation SciTech Connect Cooper, J.F.; Brummond, W.; Celeste, J.; Farmer, J.; Hoenig, C.; Krikorian, O.H.; Upadhye, R. ); Gay, R.L.; Stewart, A.; Yosim, S. . Energy Systems Group) 1991-05-13 We are developing an advanced process for treatment of mixed wastes in molten salt media at temperatures of 700--1000{degrees}C. Waste destruction has been demonstrated in a single stage oxidation process, with destruction efficiencies above 99.9999% for many waste categories. The molten salt provides a heat transfer medium, prevents thermal surges, and functions as an in situ scrubber to transform the acid-gas forming components of the waste into neutral salts and immobilizes potentially fugitive materials by a combination of particle wetting, encapsulation and chemical dissolution and solvation. Because the offgas is collected and assayed before release, and wastes containing toxic and radioactive materials are treated while immobilized in a condensed phase, the process avoids the problems sometimes associated with incineration processes. We are studying a potentially improved modification of this process, which treats oxidizable wastes in two stages: pyrolysis followed by catalyzed molten salt oxidation of the pyrolysis gases at ca. 700{degrees}C. 15 refs., 5 figs., 1 tab. 12. Integrated demonstration of molten salt oxidation with salt recycle for mixed waste treatment SciTech Connect Hsu, P.C. 1997-11-01 Molten Salt Oxidation (MSO) is a thermal, nonflame process that has the inherent capability of completely destroying organic constituents of mixed wastes, hazardous wastes, and energetic materials while retaining inorganic and radioactive constituents in the salt. For this reason, MSO is considered a promising alternative to incineration for the treatment of a variety of organic wastes. Lawrence Livermore National Laboratory (LLNL) has prepared a facility and constructed an integrated pilot-scale MSO treatment system in which tests and demonstrations are performed under carefully controlled (experimental) conditions. The system consists of a MSO processor with dedicated off-gas treatment, a salt recycle system, feed preparation equipment, and equipment for preparing ceramic final waste forms. This integrated system was designed and engineered based on laboratory experience with a smaller engineering-scale reactor unit and extensive laboratory development on salt recycle and final forms preparation. In this paper we present design and engineering details of the system and discuss its capabilities as well as preliminary process demonstration data. A primary purpose of these demonstrations is identification of the most suitable waste streams and waste types for MSO treatment. 13. Hydrological methods preferentially recover cesium from nuclear waste salt cake SciTech Connect Brooke, J.N.; Hamm, L.L. 1997-05-01 The Savannah River Site is treating high level radioactive waste in the form of insoluble solids (sludge), crystallized salt (salt cake), and salt solutions. High costs and operational concerns have prompted DOE to look for ways to improve the salt cake treatment process. A numerical model was developed to evaluate the feasibility of pump and treat technology for extracting cesium from salt cake. A modified version of the VAM3DCG code was used to first establish a steady-state flow field, then to simulate 30 days of operation. Simulation results suggest that efficient cesium extraction can be obtained with low displacement volumes. The actual extraction process will probably be less impressive because of nonuniform properties. 2 refs., 2 figs. 14. Oil field waste disposal in salt caverns: An information website SciTech Connect Tomasko, D.; Veil, J. A. 1999-12-10 Argonne National Laboratory has completed the construction of a Website for the US Department of Energy (DOE) that provides detailed information on salt caverns and their use for disposing of nonhazardous oil field wastes (NOW) and naturally occurring radioactive materials (NORM). Specific topics in the Website include the following: descriptions of salt deposits and salt caverns within the US, salt cavern construction methods, potential types of wastes, waste emplacement, regulatory issues, costs, carcinogenic and noncarcinogenic human health risks associated with postulated cavern release scenarios, new information on cavern disposal (e.g., upcoming meetings, regulatory issues, etc.), other studies supported by the National Petroleum Technology Office (NPTO) (e.g., considerations of site location, cavern stability, development issues, and bedded salt characterization in the Midland Basin), and links to other associated Web sites. In addition, the Website allows downloadable access to reports prepared on the topic that were funded by DOE. Because of the large quantities of NOW and NORM wastes generated annually by the oil industry, information presented on this Website is particularly interesting and valuable to project managers, regulators, and concerned citizens. 15. MIXING MODELING ANALYSIS FOR SRS SALT WASTE DISPOSITION SciTech Connect Lee, S. 2011-01-18 Nuclear waste at Savannah River Site (SRS) waste tanks consists of three different types of waste forms. They are the lighter salt solutions referred to as supernate, the precipitated salts as salt cake, and heavier fine solids as sludge. The sludge is settled on the tank floor. About half of the residual waste radioactivity is contained in the sludge, which is only about 8 percentage of the total waste volume. Mixing study to be evaluated here for the Salt Disposition Integration (SDI) project focuses on supernate preparations in waste tanks prior to transfer to the Salt Waste Processing Facility (SWPF) feed tank. The methods to mix and blend the contents of the SRS blend tanks were evalutaed to ensure that the contents are properly blended before they are transferred from the blend tank such as Tank 50H to the SWPF feed tank. The work consists of two principal objectives to investigate two different pumps. One objective is to identify a suitable pumping arrangement that will adequately blend/mix two miscible liquids to obtain a uniform composition in the tank with a minimum level of sludge solid particulate in suspension. The other is to estimate the elevation in the tank at which the transfer pump inlet should be located where the solid concentration of the entrained fluid remains below the acceptance criterion (0.09 wt% or 1200 mg/liter) during transfer operation to the SWPF. Tank 50H is a Waste Tank that will be used to prepare batches of salt feed for SWPF. The salt feed must be a homogeneous solution satisfying the acceptance criterion of the solids entrainment during transfer operation. The work described here consists of two modeling areas. They are the mixing modeling analysis during miscible liquid blending operation, and the flow pattern analysis during transfer operation of the blended liquid. The modeling results will provide quantitative design and operation information during the mixing/blending process and the transfer operation of the blended liquid in the Salt Disposition Integration (SDI) facility. The results will also help validate the anticipated performance of the pump vendor's design. 16. Engineering Options Assessment Report: Nitrate Salt Waste Stream Processing SciTech Connect Anast, Kurt Roy 2015-11-18 This report examines and assesses the available systems and facilities considered for carrying out remediation activities on remediated nitrate salt (RNS) and unremediated nitrate salt (UNS) waste containers at Los Alamos National Laboratory (LANL). The assessment includes a review of the waste streams consisting of 60 RNS, 29 aboveground UNS, and 79 candidate belowground UNS containers that may need remediation. The waste stream characteristics were examined along with the proposed treatment options identified in the Options Assessment Report . Two primary approaches were identified in the five candidate treatment options discussed in the Options Assessment Report: zeolite blending and cementation. Systems that could be used at LANL were examined for housing processing operations to remediate the RNS and UNS containers and for their viability to provide repackaging support for remaining LANL legacy waste. 17. Engineering Options Assessment Report. Nitrate Salt Waste Stream Processing SciTech Connect Anast, Kurt Roy 2015-11-13 This report examines and assesses the available systems and facilities considered for carrying out remediation activities on remediated nitrate salt (RNS) and unremediated nitrate salt (UNS) waste containers at Los Alamos National Laboratory (LANL). The assessment includes a review of the waste streams consisting of 60 RNS, 29 above-ground UNS, and 79 candidate below-ground UNS containers that may need remediation. The waste stream characteristics were examined along with the proposed treatment options identified in the Options Assessment Report . Two primary approaches were identified in the five candidate treatment options discussed in the Options Assessment Report: zeolite blending and cementation. Systems that could be used at LANL were examined for housing processing operations to remediate the RNS and UNS containers and for their viability to provide repackaging support for remaining LANL legacy waste. 18. Stable and efficient colour enrichment powders of nonpolar nanocrystals in LiCl. PubMed Erdem, Talha; Soran-Erdem, Zeliha; Sharma, Vijay Kumar; Kelestemur, Yusuf; Adam, Marcus; Gaponik, Nikolai; Demir, Hilmi Volkan 2015-11-14 In this work, we propose and develop the inorganic salt encapsulation of semiconductor nanocrystal (NC) dispersion in a nonpolar phase to make a highly stable and highly efficient colour converting powder for colour enrichment in light-emitting diode backlighting. Here the wrapping of the as-synthesized green-emitting CdSe/CdZnSeS/ZnS nanocrystals into a salt matrix without ligand exchange is uniquely enabled by using a LiCl ionic host dissolved in tetrahydrofuran (THF), which simultaneously disperses these nonpolar nanocrystals. We studied the emission stability of the solid films prepared using NCs with and without LiCl encapsulation on blue LEDs driven at high current levels. The encapsulated NC powder in epoxy preserved 95.5% of the initial emission intensity and stabilized at this level while the emission intensity of NCs without salt encapsulation continuously decreased to 34.7% of its initial value after 96 h of operation. In addition, we investigated the effect of ionic salt encapsulation on the quantum efficiency of nonpolar NCs and found the quantum efficiency of the NCs-in-LiCl to be 75.1% while that of the NCs in dispersion was 73.0% and that in a film without LiCl encapsulation was 67.9%. We believe that such ionic salt encapsulated powders of nonpolar NCs presented here will find ubiquitous use for colour enrichment in display backlighting. PMID:26238664 19. Injector nozzle for molten salt destruction of energetic waste materials DOEpatents Brummond, William A. (Livermore, CA); Upadhye, Ravindra S. (Pleasanton, CA) 1996-01-01 An injector nozzle has been designed for safely injecting energetic waste materials, such as high explosives, propellants, and rocket fuels, into a molten salt reactor in a molten salt destruction process without premature detonation or back burn in the injection system. The energetic waste material is typically diluted to form a fluid fuel mixture that is injected rapidly into the reactor. A carrier gas used in the nozzle serves as a carrier for the fuel mixture, and further dilutes the energetic material and increases its injection velocity into the reactor. The injector nozzle is cooled to keep the fuel mixture below the decomposition temperature to prevent spontaneous detonation of the explosive materials before contact with the high-temperature molten salt bath. 20. Injector nozzle for molten salt destruction of energetic waste materials DOEpatents Brummond, W.A.; Upadhye, R.S. 1996-02-13 An injector nozzle has been designed for safely injecting energetic waste materials, such as high explosives, propellants, and rocket fuels, into a molten salt reactor in a molten salt destruction process without premature detonation or back burn in the injection system. The energetic waste material is typically diluted to form a fluid fuel mixture that is injected rapidly into the reactor. A carrier gas used in the nozzle serves as a carrier for the fuel mixture, and further dilutes the energetic material and increases its injection velocity into the reactor. The injector nozzle is cooled to keep the fuel mixture below the decomposition temperature to prevent spontaneous detonation of the explosive materials before contact with the high-temperature molten salt bath. 2 figs. 1. Novel waste printed circuit board recycling process with molten salt PubMed Central Riedewald, Frank; Sousa-Gallagher, Maria 2015-01-01 The objective of the method was to prove the concept of a novel waste PCBs recycling process which uses inert, stable molten salts as the direct heat transfer fluid and, simultaneously, uses this molten salt to separate the metal products in either liquid (solder, zinc, tin, lead, etc.) or solid (copper, gold, steel, palladium, etc.) form at the operating temperatures of 450–470 °C. The PCB recovery reactor is essentially a U-shaped reactor with the molten salt providing a continuous fluid, allowing molten salt access from different depths for metal recovery. A laboratory scale batch reactor was constructed using 316L as suitable construction material. For safety reasons, the inert, stable LiCl–KCl molten salts were used as direct heat transfer fluid. Recovered materials were washed with hot water to remove residual salt before metal recovery assessment. The impact of this work was to show metal separation using molten salts in one single unit, by using this novel reactor methodology. • The reactor is a U-shaped reactor filled with a continuous liquid with a sloped bottom representing a novel reactor concept. • This method uses large PCB pieces instead of shredded PCBs as the reactor volume is 2.2 L. • The treated PCBs can be removed via leg B while the process is on-going. PMID:26150977 2. Novel waste printed circuit board recycling process with molten salt. PubMed Riedewald, Frank; Sousa-Gallagher, Maria 2015-01-01 The objective of the method was to prove the concept of a novel waste PCBs recycling process which uses inert, stable molten salts as the direct heat transfer fluid and, simultaneously, uses this molten salt to separate the metal products in either liquid (solder, zinc, tin, lead, etc.) or solid (copper, gold, steel, palladium, etc.) form at the operating temperatures of 450-470 °C. The PCB recovery reactor is essentially a U-shaped reactor with the molten salt providing a continuous fluid, allowing molten salt access from different depths for metal recovery. A laboratory scale batch reactor was constructed using 316L as suitable construction material. For safety reasons, the inert, stable LiCl-KCl molten salts were used as direct heat transfer fluid. Recovered materials were washed with hot water to remove residual salt before metal recovery assessment. The impact of this work was to show metal separation using molten salts in one single unit, by using this novel reactor methodology. •The reactor is a U-shaped reactor filled with a continuous liquid with a sloped bottom representing a novel reactor concept.•This method uses large PCB pieces instead of shredded PCBs as the reactor volume is 2.2 L.•The treated PCBs can be removed via leg B while the process is on-going. PMID:26150977 3. BLENDING ANALYSIS FOR RADIOACTIVE SALT WASTE PROCESSING FACILITY SciTech Connect Lee, S. 2012-05-10 Savannah River National Laboratory (SRNL) evaluated methods to mix and blend the contents of the blend tanks to ensure the contents are properly blended before they are transferred from the blend tank such as Tank 21 and Tank 24 to the Salt Waste Processing Facility (SWPF) feed tank. The tank contents consist of three forms: dissolved salt solution, other waste salt solutions, and sludge containing settled solids. This paper focuses on developing the computational model and estimating the operation time of submersible slurry pump when the tank contents are adequately blended prior to their transfer to the SWPF facility. A three-dimensional computational fluid dynamics approach was taken by using the full scale configuration of SRS Type-IV tank, Tank 21H. Major solid obstructions such as the tank wall boundary, the transfer pump column, and three slurry pump housings including one active and two inactive pumps were included in the mixing performance model. Basic flow pattern results predicted by the computational model were benchmarked against the SRNL test results and literature data. Tank 21 is a waste tank that is used to prepare batches of salt feed for SWPF. The salt feed must be a homogeneous solution satisfying the acceptance criterion of the solids entrainment during transfer operation. The work scope described here consists of two modeling areas. They are the steady state flow pattern calculations before the addition of acid solution for tank blending operation and the transient mixing analysis during miscible liquid blending operation. The transient blending calculations were performed by using the 95% homogeneity criterion for the entire liquid domain of the tank. The initial conditions for the entire modeling domain were based on the steady-state flow pattern results with zero second phase concentration. The performance model was also benchmarked against the SRNL test results and literature data. 4. Disposal of soluble salt waste from coal gasification SciTech Connect McKnight, C.E. 1980-06-01 This paper addresses pollutants in the form of soluble salts and resource recovery in the form of water and land. A design for disposal of soluble salts has been produced. The interactions of its parameters have been shown by a process design study. The design will enable harmonious compliance with United States Public Laws 92-500 and 94-580, relating to water pollution and resource recovery. In the disposal of waste salt solutions, natural water resources need not be contaminated, because an encapsulation technique is available which will immobilize the salts. At the same time it will make useful landforms available, and water as a resource can be recovered. There is a cost minimum when electrodialysis and evaporation are combined, which is not realizable with evaporation alone, unless very low-cost thermal energy is available or unless very high-cost pretreatment for electrodialysis is required. All the processes making up the proposed disposal process are commercially available, although they are nowhere operating commercially as one process. Because of the commercial availability of the processes, the proposed process may be a candidate 'best commercially available treatment' for soluble salt disposal. 5. Modeling of Sulfate Double-salts in Nuclear Wastes SciTech Connect Toghiani, B. 2000-10-30 Due to limited tank space at Hanford and Savannah River, the liquid nuclear wastes or supernatants have been concentrated in evaporators to remove excess water prior to the hot solutions being transferred to underground storage tanks. As the waste solutions cooled, the salts in the waste exceeded the associated solubility limits and precipitated in the form of saltcakes. The initial step in the remediation of these saltcakes is a rehydration process called saltcake dissolution. At Hanford, dissolution experiments have been conducted on small saltcake samples from five tanks. Modeling of these experimental results, using the Environmental Simulation Program (ESP), are being performed at the Diagnostic Instrumentation and Analysis Laboratory (DIAL) at Mississippi State University. The River Protection Project (RPP) at Hanford will use these experimental and theoretical results to determine the amount of water that will be needed for its dissolution and retrieval operations. A comprehensive effort by the RPP and the Tank Focus Area continues to validate and improve the ESP and its databases for this application. The initial effort focused on the sodium, fluoride, and phosphate system due to its role in the formation of pipeline plugs. In FY 1999, an evaluation of the ESP predictions for sodium fluoride, trisodium phosphate dodecahydrate, and natrophosphate clearly indicated that improvements to the Public database of the ESP were needed. One of the improvements identified was double salts. The inability of any equilibrium thermodynamic model to properly account for double salts in the system can result in errors in the predicted solid-liquid equilibria (SLE) of species in the system. The ESP code is evaluated by comparison with experimental data where possible. However, data does not cover the range of component concentrations and temperatures found in many tank wastes. Therefore, comparison of ESP with another code is desirable, and may illuminate problems with both. For this purpose, the SOLGASMIX code was used in conjunction with a small private database developed at ORNL. This code calculates thermodynamic equilibria through minimization of Gibbs Energy, and utilizes the Pitzer model for activity coefficients. The sodium nitrate-sulfate double salt and the sodium fluoride-sulfate double salt were selected for the FY 2000 validation study of ESP. Even though ESP does not include the sulfate-nitrate double salt, this study found that this omission does not appear to be a major consequence. In this case, the solubility predictions with and without the sulfate-nitrate double salt are comparable. In contrast, even though the sulfate-fluoride double salt is included within the ESP databank, comparison to previous experimental results indicates that ESP underestimates solubility. Thus, the prediction for the sulfate-fluoride system needs to be improved. A main consequence of the inability to accurately predict the SLE of double salts is its impact on the predicted ionic strength of the solution. The ionic strength has been observed to be an important factor in the formation of pipeline plugs. To improve the ESP prediction, solubility tests on the sulfate-fluoride system are underway at DIAL, and these experimental results will be incorporated into the Public database by OLI System, Inc. Preliminary ESP simulations also indicated difficulties with the SLE prediction for anhydrous sodium sulfate. The Public database for the ESP does not include fundamental parameters for this solid in mixed solutions below 32.4 C. The limitation, in the range of anhydrous sodium sulfate, leads to convergence problems in ESP and to inaccurate predictions of solubility near the invariant point when sodium sulfate decahydrate and other salts, such as sodium nitrate, were present. These difficulties were partially corrected through the use of an additional database. In conclusion, these results indicate the need for experimental data at temperatures above 25 C and in solutions containing both nitrate and hydroxide. Furthermore, the validation and do 6. Disposal of oil field wastes and NORM wastes into salt caverns. SciTech Connect Veil, J. A. 1999-01-27 Salt caverns can be formed through solution mining in the bedded or domal salt formations that are found in many states. Salt caverns have traditionally been used for hydrocarbon storage, but caverns have also been used to dispose of some types of wastes. This paper provides an overview of several years of research by Argonne National Laboratory on the feasibility and legality of using salt caverns for disposing of nonhazardous oil field wastes (NOW) and naturally occurring radioactive materials (NORM), the risk to human populations from this disposal method, and the cost of cavern disposal. Costs are compared between the four operating US disposal caverns and other commercial disposal options located in the same geographic area as the caverns. Argonne's research indicates that disposal of NOW into salt caverns is feasible and, in most cases, would not be prohibited by state agencies (although those agencies may need to revise their wastes management regulations). A risk analysis of several cavern leakage scenarios suggests that the risk from cavern disposal of NOW and NORM wastes is below accepted safe risk thresholds. Disposal caverns are economically competitive with other disposal options. 7. Pyrolytic conversion of plastic and rubber waste to hydrocarbons with basic salt catalysts DOEpatents Wingfield, Jr., Robert C.; Braslaw, Jacob; Gealer, Roy L. 1985-01-01 The invention relates to a process for improving the pyrolytic conversion of waste selected from rubber and plastic to low molecular weight olefinic materials by employing basis salt catalysts in the waste mixture. The salts comprise alkali or alkaline earth compounds, particularly sodium carbonate, in an amount of greater than about 1 weight percent based on the waste feed. 8. Stable and efficient colour enrichment powders of nonpolar nanocrystals in LiCl NASA Astrophysics Data System (ADS) Erdem, Talha; Soran-Erdem, Zeliha; Sharma, Vijay Kumar; Kelestemur, Yusuf; Adam, Marcus; Gaponik, Nikolai; Demir, Hilmi Volkan 2015-10-01 In this work, we propose and develop the inorganic salt encapsulation of semiconductor nanocrystal (NC) dispersion in a nonpolar phase to make a highly stable and highly efficient colour converting powder for colour enrichment in light-emitting diode backlighting. Here the wrapping of the as-synthesized green-emitting CdSe/CdZnSeS/ZnS nanocrystals into a salt matrix without ligand exchange is uniquely enabled by using a LiCl ionic host dissolved in tetrahydrofuran (THF), which simultaneously disperses these nonpolar nanocrystals. We studied the emission stability of the solid films prepared using NCs with and without LiCl encapsulation on blue LEDs driven at high current levels. The encapsulated NC powder in epoxy preserved 95.5% of the initial emission intensity and stabilized at this level while the emission intensity of NCs without salt encapsulation continuously decreased to 34.7% of its initial value after 96 h of operation. In addition, we investigated the effect of ionic salt encapsulation on the quantum efficiency of nonpolar NCs and found the quantum efficiency of the NCs-in-LiCl to be 75.1% while that of the NCs in dispersion was 73.0% and that in a film without LiCl encapsulation was 67.9%. We believe that such ionic salt encapsulated powders of nonpolar NCs presented here will find ubiquitous use for colour enrichment in display backlighting.In this work, we propose and develop the inorganic salt encapsulation of semiconductor nanocrystal (NC) dispersion in a nonpolar phase to make a highly stable and highly efficient colour converting powder for colour enrichment in light-emitting diode backlighting. Here the wrapping of the as-synthesized green-emitting CdSe/CdZnSeS/ZnS nanocrystals into a salt matrix without ligand exchange is uniquely enabled by using a LiCl ionic host dissolved in tetrahydrofuran (THF), which simultaneously disperses these nonpolar nanocrystals. We studied the emission stability of the solid films prepared using NCs with and without LiCl encapsulation on blue LEDs driven at high current levels. The encapsulated NC powder in epoxy preserved 95.5% of the initial emission intensity and stabilized at this level while the emission intensity of NCs without salt encapsulation continuously decreased to 34.7% of its initial value after 96 h of operation. In addition, we investigated the effect of ionic salt encapsulation on the quantum efficiency of nonpolar NCs and found the quantum efficiency of the NCs-in-LiCl to be 75.1% while that of the NCs in dispersion was 73.0% and that in a film without LiCl encapsulation was 67.9%. We believe that such ionic salt encapsulated powders of nonpolar NCs presented here will find ubiquitous use for colour enrichment in display backlighting. Electronic supplementary information (ESI) available. See DOI: 10.1039/c5nr02696b 9. Ceramic waste form for residues from molten salt oxidation of mixed wastes SciTech Connect Van Konynenburg, R.A.; Hopper, R.W.; Rard, J.A. 1995-11-01 A ceramic waste form based on Synroc-D is under development for the incorporation of the mineral residues from molten salt oxidation treatment of mixed low-level wastes. Samples containing as many as 32 chemical elements have been fabricated, characterized, and leach-tested. Universal Treatment Standards have been satisfied for all regulated elements except and two (lead and vanadium). Efforts are underway to further improve chemical durability. 10. Overview of Fiscal Year 2002 Research and Development for Savannah River Site's Salt Waste Processing Facility SciTech Connect H. D. Harmon, R. Leugemors, PNNL; S. Fink, M. Thompson, D. Walker, WSRC; P. Suggs, W. D. Clark, Jr 2003-02-26 The Department of Energy's (DOE) Savannah River Site (SRS) high-level waste program is responsible for storage, treatment, and immobilization of high-level waste for disposal. The Salt Processing Program (SPP) is the salt (soluble) waste treatment portion of the SRS high-level waste effort. The overall SPP encompasses the selection, design, construction and operation of treatment technologies to prepare the salt waste feed material for the site's grout facility (Saltstone) and vitrification facility (Defense Waste Processing Facility). Major constituents that must be removed from the salt waste and sent as feed to Defense Waste Processing Facility include actinides, strontium, cesium, and entrained sludge. In fiscal year 2002 (FY02), research and development (R&D) on the actinide and strontium removal and Caustic-Side Solvent Extraction (CSSX) processes transitioned from technology development for baseline process selection to providing input for conceptual design of the Salt Waste Processing Facility. The SPP R&D focused on advancing the technical maturity, risk reduction, engineering development, and design support for DOE's engineering, procurement, and construction (EPC) contractors for the Salt Waste Processing Facility. Thus, R&D in FY02 addressed the areas of actual waste performance, process chemistry, engineering tests of equipment, and chemical and physical properties relevant to safety. All of the testing, studies, and reports were summarized and provided to the DOE to support the Salt Waste Processing Facility, which began conceptual design in September 2002. 11. Treatment of Difficult Wastes with Molten Salt Oxidation SciTech Connect Hsu, P C; Kwak, S 2003-02-21 Molten salt oxidation (MSO) is a good alternative to incineration for the treatment of a variety of organic wastes such as explosives, low-level mixed waste streams, PCB contaminated oils, spent resins and carbon. Since mid-1990s, the U.S. Army Defense Ammunition Center (DAC) and the Department of Energy (DOE) have jointly invested in MSO development at the Lawrence Livermore National Laboratory (LLNL). LLNL first demonstrated the MSO process for the effective destruction of explosives, explosives-contaminated materials, and other wastes on a 1.5-kg/hr bench-scale unit, and then in an integrated MSO facility capable of treating 8 kg/hr of low-level radioactive mixed wastes. Several MSO systems have been built with sizes up to 10 ft in height and 16 inches in diameter. LLNL in 2001 completed a MSO plant for DAC for the destruction of explosives-contaminated sludge and explosives-contaminated carbon. We will present in this paper our latest demonstration data and our operational experience with MSO. 12. Identifying suitable "piercement" salt domes for nuclear waste storage sites SciTech Connect Kehle, R. 1980-08-01 Piercement salt domes of the northern interior salt basins of the Gulf of Mexico are being considered as permanent storage sites for both nuclear and chemically toxic wastes. The suitable domes are stable and inactive, having reached their final evolutionary configuration at least 30 million years ago. They are buried to depths far below the level to which erosion will penetrate during the prescribed storage period and are not subject to possible future reactivation. The salt cores of these domes are themselves impermeable, permitting neither the entry nor exit of ground water or other unwanted materials. In part, a stable dome may be recognized by its present geometric configuration, but conclusive proof depends on establishing its evolutionary state. The evolutionary state of a dome is obtained by reconstructing the growth history of the dome as revealed by the configuration of sedimentary strata in a large area (commonly 3,000 square miles or more) surrounding the dome. A high quality, multifold CDP reflection seismic profile across a candidate dome will provide much of the necessary information when integrated with available subsurface control. Additional seismic profiles may be required to confirm an apparent configuration of the surrounding strata and an interpreted evolutionary history. High frequency seismic data collected in the near vicinity of a dome are also needed as a supplement to the CDP data to permit accurate depiction of the configuration of shallow strata. Such data must be tied to shallow drill hole control to confirm the geologic age at which dome growth ceased. If it is determined that a dome reached a terminal configuration many millions of years ago, such a dome is incapable of reactivation and thus constitutes a stable storage site for nuclear wastes. 13. Method to synthesize dense crystallized sodalite pellet for immobilizing halide salt radioactive waste DOEpatents Koyama, Tadafumi (Tokyo, JP) 1994-01-01 A method for immobilizing waste chloride salts containing radionuclides such as cesium and strontium and hazardous materials such as barium. A sodalite intermediate is prepared by mixing appropriate amounts of silica, alumina and sodium hydroxide with respect to sodalite and heating the mixture to form the sodalite intermediate and water. Heating is continued to drive off the water to form a water-free intermediate. The water-free intermediate is mixed with either waste salt or waste salt which has been contacted with zeolite to concentrate the radionuclides and hazardous material. The waste salt-intermediate mixture is then compacted and heated under conditions of heat and pressure to form sodalite with the waste salt, radionuclides and hazardous material trapped within the sodalite cage structure. This provides a final product having excellent leach resistant capabilities. 14. Method to synthesize dense crystallized sodalite pellet for immobilizing halide salt radioactive waste DOEpatents 1994-08-23 A method is described for immobilizing waste chloride salts containing radionuclides such as cesium and strontium and hazardous materials such as barium. A sodalite intermediate is prepared by mixing appropriate amounts of silica, alumina and sodium hydroxide with respect to sodalite and heating the mixture to form the sodalite intermediate and water. Heating is continued to drive off the water to form a water-free intermediate. The water-free intermediate is mixed with either waste salt or waste salt which has been contacted with zeolite to concentrate the radionuclides and hazardous material. The waste salt-intermediate mixture is then compacted and heated under conditions of heat and pressure to form sodalite with the waste salt, radionuclides and hazardous material trapped within the sodalite cage structure. This provides a final product having excellent leach resistant capabilities. 15. Method to synthesize dense crystallized sodalite pellet for immobilizing halide salt radioactive waste DOEpatents Koyama, T. 1992-01-01 This report describes a method for immobilizing waste chloride salts containing radionuclides such as cesium and strontium and hazardous materials such as barium. A sodalite intermediate is prepared by mixing appropriate amounts of silica, alumina and sodium hydroxide with respect to sodalite and heating the mixture to form the sodalite intermediate and water. Heating is continued to drive off the water to form a water-free intermediate. The water-free intermediate is mixed with either waste salt or waste salt which has been contacted with zeolite to concentrate the radionuclides and hazardous material. The waste salt-intermediate mixture is then compacted and heated under conditions of heat and pressure to form sodalite with the waste salt, radionuclides and hazardous material trapped within the sodalite cage structure. This provides a final product having excellent leach resistant capabilities. 16. Container materials for isolation of radioactive waste in salt SciTech Connect Streicher, M.A.; Andrews, A. 1987-10-01 The workshop reviewed the extensive data on the corrosion resistance of low-carbon steel in simulated salt repository environments, determined whether these data were sufficient to recommend low-carbon steel for fabrication of the container, and assessed the suitability of other materials under consideration in the SRP. The panelists determined the need for testing and research programs, recommended experimental approaches, and recommended materials based on existing technology. On the first day of the workshop, presentations were made on waste package requirements; the expected corrosion environment; degradation processes, including a review of data from corrosion tests on carbon steel; and rationales for container design and materials, modeling studies, and planned future work. The second day was devoted to a panel caucus, presentation of workshop findings, and open discussion. 76 refs., 2 figs., 3 tabs. 17. Electrodialysis-based separation process for salt recovery and recycling from waste water DOEpatents Tsai, S.P. 1997-07-08 A method for recovering salt from a process stream containing organic contaminants is provided, comprising directing the waste stream to a desalting electrodialysis unit so as to create a concentrated and purified salt permeate and an organic contaminants-containing stream, and contacting said concentrated salt permeate to a water-splitting electrodialysis unit so as to convert the salt to its corresponding base and acid. 6 figs. 18. Electrodialysis-based separation process for salt recovery and recycling from waste water DOEpatents Tsai, Shih-Perng 1997-01-01 A method for recovering salt from a process stream containing organic contaminants is provided, comprising directing the waste stream to a desalting electrodialysis unit so as to create a concentrated and purified salt permeate and an organic contaminants containing stream, and contacting said concentrated salt permeate to a water-splitting electrodialysis unit so as to convert the salt to its corresponding base and acid. 19. Test plan for immobilization of salt-containing surrogate mixed wastes using polyester resins SciTech Connect Biyani, R.K.; Douglas, J.C.; Hendrickson, D.W. 1997-07-07 Past operations at many Department of Energy (DOE) sites have resulted in the generation of several waste streams with high salt content. These wastes contain listed and characteristic hazardous constituents and are radioactive. The salts contained in the wastes are primarily chloride, sulfate, nitrate, metal oxides, and hydroxides. DOE has placed these types of wastes under the purview of the Mixed Waste Focus Area (MWFA). The MWFA has been tasked with developing and facilitating the implementation of technologies to treat these wastes in support of customer needs and requirements. The MWFA has developed a Technology Development Requirements Document (TDRD), which specifies performance requirements for technology owners and developers to use as a framework in developing effective waste treatment solutions. This project will demonstrate the use of polyester resins in encapsulating and solidifying DOEs mixed wastes containing salts, as an alternative to conventional and other emerging immobilization technologies. 20. Recovery of salt wastes in the production of propylene oxide SciTech Connect Zyablitseva, M.P.; Tyurin, B.K.; Kudinov, V.I.; Bukbulatov, I.K.; Mazanko, A.F. 1983-02-01 In the production of propylene oxide as much as 40 t dilute calcium chloride solution forms per ton of product in the step of saponification of propylene chlorhydrine with milk of lime. To create a zero-waste technology for production of propylene oxide, there is practical interest in saponification of propylene chlorhydrine with electrolysis brines with recovery of the resultant solution of sodium chloride after purification to remove organic impurities. The possibility of using an electrochemical method to purify wastewater from production of propylene oxide in using the purified solution as starting material for production of electrolysis brines was investigated. Experimental testing of processes of purification and recovery of wastewaters in a regime of industrial electrolysis confirmed the possibility of using purified wastewater from production of propylene oxide as brine for electrolysis. Incorporation of the developed method into industry will permit zero-waste production of propylene oxide with a closed salt cycle. The cost of purification of 1 m/sup 3/ wastewater is 1-1.5 rubles. 1. Preliminary technical and legal evaluation of disposing of nonhazardous oil field waste into salt caverns SciTech Connect Veil, J.; Elcock, D.; Raivel, M.; Caudle, D.; Ayers, R.C. Jr.; Grunewald, B. 1996-06-01 Caverns can be readily formed in salt formations through solution mining. The caverns may be formed incidentally, as a result of salt recovery, or intentionally to create an underground chamber that can be used for storing hydrocarbon products or compressed air or disposing of wastes. The purpose of this report is to evaluate the feasibility, suitability, and legality of disposing of nonhazardous oil and gas exploration, development, and production wastes (hereafter referred to as oil field wastes, unless otherwise noted) in salt caverns. Chapter 2 provides background information on: types and locations of US subsurface salt deposits; basic solution mining techniques used to create caverns; and ways in which salt caverns are used. Later chapters provide discussion of: federal and state regulatory requirements concerning disposal of oil field waste, including which wastes are considered eligible for cavern disposal; waste streams that are considered to be oil field waste; and an evaluation of technical issues concerning the suitability of using salt caverns for disposing of oil field waste. Separate chapters present: types of oil field wastes suitable for cavern disposal; cavern design and location; disposal operations; and closure and remediation. This report does not suggest specific numerical limits for such factors or variables as distance to neighboring activities, depths for casings, pressure testing, or size and shape of cavern. The intent is to raise issues and general approaches that will contribute to the growing body of information on this subject. 2. Disposal of oil field wastes into salt caverns: Feasibility, legality, risk, and costs SciTech Connect Veil, J.A. 1997-10-01 Salt caverns can be formed through solution mining in the bedded or domal salt formations that are found in many states. Salt caverns have traditionally been used for hydrocarbon storage, but caverns have also been used to dispose of some types of wastes. This paper provides an overview of several years of research by Argonne National Laboratory on the feasibility and legality of using salt caverns for disposing of oil field wastes, the risks to human populations from this disposal method, and the cost of cavern disposal. Costs are compared between the four operating US disposal caverns and other commercial disposal options located in the same geographic area as the caverns. Argonnes research indicates that disposal of oil field wastes into salt caverns is feasible and legal. The risk from cavern disposal of oil field wastes appears to be below accepted safe risk thresholds. Disposal caverns are economically competitive with other disposal options. 3. Expedited demonstration of molten salt mixed waste treatment technology. Final report SciTech Connect 1995-02-02 This final report discusses the molten salt mixed waste project in terms of the various subtasks established. Subtask 1: Carbon monoxide emissions; Establish a salt recycle schedule and/or a strategy for off-gas control for MWMF that keeps carbon monoxide emission below 100 ppm on an hourly averaged basis. Subtask 2: Salt melt viscosity; Experiments are conducted to determine salt viscosity as a function of ash composition, ash concentration, temperature, and time. Subtask 3: Determine that the amount of sodium carbonate entrained in the off-gas is minimal, and that any deposited salt can easily be removed form the piping using a soot blower or other means. Subtask 4: The provision of at least one final waste form that meets the waste acceptance criteria of a landfill that will take the waste. This report discusses the progress made in each of these areas. 4. Polyethylene encapsulatin of nitrate salt wastes: Waste form stability, process scale-up, and economics SciTech Connect Kalb, P.D.; Heiser, J.H. III; Colombo, P. 1991-07-01 A polyethylene encapsulation system for treatment of low-level radioactive, hazardous, and mixed wastes has been developed at Brookhaven National Laboratory. Polyethylene has several advantages compared with conventional solidification/stabilization materials such as hydraulic cements. Waste can be encapsulated with greater efficiency and with better waste form performance than is possible with hydraulic cement. The properties of polyethylene relevant to its long-term durability in storage and disposal environments are reviewed. Response to specific potential failure mechanisms including biodegradation, radiation, chemical attack, flammability, environmental stress cracking, and photodegradation are examined. These data are supported by results from extensive waste form performance testing including compressive yield strength, water immersion, thermal cycling, leachability of radioactive and hazardous species, irradiation, biodegradation, and flammability. The bench-scale process has been successfully tested for application with a number of specific problem'' waste streams. Quality assurance and performance testing of the resulting waste form confirmed scale-up feasibility. Use of this system at Rocky Flats Plant can result in over 70% fewer drums processed and shipped for disposal, compared with optimal cement formulations. Based on the current Rocky Flats production of nitrate salt per year, polyethylene encapsulation can yield an estimated annual savings between $1.5 million and$2.7 million, compared with conventional hydraulic cement systems. 72 refs., 23 figs., 16 tabs. 5. Alternative Electrochemical Salt Waste Forms, Summary of FY2010 Results SciTech Connect Riley, Brian J.; Rieck, Bennett T.; Crum, Jarrod V.; Matyas, Josef; McCloy, John S.; Sundaram, S. K.; Vienna, John D. 2010-08-01 In FY2009, PNNL performed scoping studies to qualify two waste form candidates, tellurite (TeO2-based) glasses and halide minerals, for the electrochemical waste stream for further investigation. Both candidates showed promise with acceptable PCT release rates and effective incorporation of the 10% fission product waste stream. Both candidates received reprisal for FY2010 and were further investigated. At the beginning of FY2010, an in-depth literature review kicked off the tellurite glasses study. The review was aimed at ascertaining the state-of-the-art for chemical durability testing and mixed chloride incorporation for tellurite glasses. The literature review led the authors to 4 unique binary and 1 unique ternary systems for further investigation which include TeO2 plus the following: PbO, Al2O3-B2O3, WO3, P2O5, and ZnO. Each system was studied with and without a mixed chloride simulated electrochemical waste stream and the literature review provided the starting points for the baseline compositions as well as starting points for melting temperature, compatible crucible types, etc. The most promising glasses in each system were scaled up in production and were analyzed with the Product Consistency Test, a chemical durability test. Baseline and PCT glasses were analyzed to determine their state, i.e., amorphous, crystalline, phase separated, had undissolved material within the bulk, etc. Conclusions were made as well as the proposed direction for FY2011 plans. Sodalite was successfully synthesized by the sol-gel method. The vast majority of the dried sol-gel consisted of sodalite with small amounts of alumino-silicates and unreacted salt. Upon firing the powders made by sol-gel, the primary phase observed was sodalite with the addition of varying amounts of nepheline, carnegieite, lithium silicate, and lanthanide oxide. The amount of sodalite, nepheline, and carnegieite as well as the bulk density of the fired pellets varied with firing temperature, sol-gel process chemistry, and the amount of glass sintering aid added to the batch. As the firing temperature was increased from 850 C to 950 C, chloride volatility increased, the fraction of sodalite decreased, and the fractions nepheline and carnegieite increased. This indicates that the sodalite structure is not stable and begins to convert to nepheline and carnegieite under these conditions at 950 C. Density has opposite relationship with relation to firing temperature. The addition of a NBS-1, a glass sintering aid, had a positive effect on bulk density and increased the stability of the sodalite structure in a minimal way. 6. Criticality considerations for salt-cake disolution in DOE waste tanks SciTech Connect Trumble, E.F.; Niemer, K.A. 1995-12-31 A large amount of high-level waste is being stored in the form of salt cake at the Savannah River site (SRS) in large (1.3 x 106 gal) underground tanks awaiting startup of the Defense Waste Processing Facility (DWPF). This salt cake will be dissolved with water, and the solution will be fed to DWPF for immobilization in borosilicate glass. Some of the waste that was transferred to the tanks contained enriched uranium and plutonium from chemical reprocessing streams. As water is added to these tanks to dissolve the salt cake, the insoluble portion of this fissile material will be left behind in the tank as the salt solution is pumped out. Because the salt acts as a diluent to the fissile material, the process of repeated water addition, salt dissolution, and salt solution removal will act as a concentrating mechanism for the undissolved fissile material that will remain in the tank. It is estimated that tank 41 H at SRS contains 20 to 120 kg of enriched uranium, varying from 10 to 70% {sup 235}U, distributed nonuniformly throughout the tank. This paper discusses the criticality concerns associated with the dissolution of salt cake in this tank. These concerns are also applicable to other salt cake waste tanks that contain significant quantities of enriched uranium and/or plutonium. 7. Reconsolidation of salt as applied to permanent seals for the Waste Isolation Pilot Plant SciTech Connect Hansen, F.D.; Callahan, G.D.; Van Sembeek, L.L. 1993-07-01 Reconsolidated salt is a fundamental component of the permanent seals for the Waste Isolation Pilot Plant. As regulations are currently understood and seal concepts envisioned, emplaced salt is the sole long-term seal component designed to prevent the shafts from becoming preferred pathways for rating gases or liquids. Studies under way in support of the sealing function of emplaced salt include laboratory testing of crushed salt small-scale in situ tests, constitutive modeling of crushed salt, calculations of the opening responses during operation and closure, and design practicalities including emplacement techniques. This paper briefly summarizes aspects of these efforts and key areas of future work. 8. Salt splitting of sodium-dominated radioactive waste using ceramic membranes SciTech Connect Hollenberg, G.W.; Carlson, C.D.; Virkar, A.; Joshi, A. 1994-08-01 The potential for salt splitting of sodium dominated radioactive wastes by use of a ceramic membrane is reviewed. The technical basis for considering this processing technology is derived from the technology developed for battery and chlor-alkali chemical industry. Specific comparisons are made with the commercial organic membranes which are the standard in nonradioactive salt splitting. Two features of ceramic membranes are expected to be especially attractive: high tolerance to gamma irradiation and high selectivity between sodium and other ions. The objective of the salt splitting process is to separate nonradioactive sodium from contaminated sodium salts prior to other pretreatment processes in order to: (1) concentrate the waste in order to reduce the volume of subsequent additives and capacity of equipment, (2) decrease the pH of the waste in preparation for further processing, and (3) provide sodium with very low radioactivity levels for caustic washing of sludge or low level and mixed waste vitrification. 9. Expected near-field thermal performance for nuclear waste repositories at potential salt sites: Technical report SciTech Connect McNulty, E.G. 1987-08-01 Thermal analyses were made for the environmental assessments of seven potential salt sites for a nuclear waste repository. These analyses predicted that potential repository sites in domal salts located in the Gulf Coast will experience higher temperature than those in bedded salts of Paradox and Palo Duro Basins, mainly because of higher ambient temperatures at depth. The TEMPV5 code, a semi-analytical heat transfer code for finite line sources, calculated temperatures for commercial high-level waste (CHLW) and spent fuel from pressurized-water reactors (SFPWR). Benchmarks with HEATING6, THAC-SIP-3D, STEALTH, and SPECTROM-41 showed that TEMPV5 agreed closely in the very near field around the waste package and approximately in the near-field and far-field regions of the repository. The analyses used site-specific thermal conductivities that were increased by 40% to compensate for reductions caused by testing technique, salt impurities, and other heterogeneities, and sampling disturbance. Analyses showed peak salt temperatures of 236/sup 0/C (CHLW) and 134/sup 0/C (SFPWR) for the bedded salt and 296/sup 0/C (CHLW) and 180/sup 0/C (SFPWR) for the domal salt. Analyses with uncorrected laboratory thermal conductivities would increase peak salt temperatures by about 120/sup 0/C for CHLW and about 60/sup 0/C for SFPWR. These temperature increases would increase the thermally induced flow of brine and accelerate corrosion of the waste package. 30 refs., 35 figs., 48 tabs. 10. Application of molten salt oxidation for the minimization and recovery of plutonium-238 contaminated wastes SciTech Connect Wishau, R. 1998-05-01 Molten salt oxidation (MSO) is proposed as a {sup 238}Pu waste treatment technology that should be developed for volume reduction and recovery of {sup 238}Pu and as an alternative to the transport and permanent disposal of {sup 238}Pu waste to the WIPP repository. In MSO technology, molten sodium carbonate salt at 800--900 C in a reaction vessel acts as a reaction media for wastes. The waste material is destroyed when injected into the molten salt, creating harmless carbon dioxide and steam and a small amount of ash in the spent salt. The spent salt can be treated using aqueous separation methods to reuse the salt and to recover 99.9% of the precious {sup 238}Pu that was in the waste. Tests of MSO technology have shown that the volume of combustible TRU waste can be reduced by a factor of at least twenty. Using this factor the present inventory of 574 TRU drums of {sup 238}Pu contaminated wastes is reduced to 30 drums. Further {sup 238}Pu waste costs of $22 million are avoided from not having to repackage 312 of the 574 drums to a drum total of more than 4,600 drums. MSO combined with aqueous processing of salts will recover approximately 1.7 kilograms of precious {sup 238}Pu valued at 4 million dollars (at$2,500/gram). Thus, installation and use of MSO technology at LANL will result in significant cost savings compared to present plans to transport and dispose {sup 238}Pu TRU waste to the WIPP site. Using a total net present value cost for the MSO project as $4.09 million over a five-year lifetime, the project can pay for itself after either recovery of 1.6 kg of Pu or through volume reduction of 818 drums or a combination of the two. These savings show a positive return on investment. 11. Modified phosphate ceramics for stabilization and solidification of salt mixed wastes. SciTech Connect Singh, D. 1998-06-26 Novel chemically bonded phosphate ceramics have been investigated for stabilization and solidification of chloride and nitrate salt wastes. Using low-temperature processing, we stabilized and solidified chloride and nitrate surrogate salts (with hazardous metals) in magnesium potassium phosphate ceramics up to waste loadings of 70-80 wt.%. A variety of characterizations, including strength, microstructure, and leaching, were then conducted on the waste forms. Leaching tests show that all heavy metals in the leachant are well below the EPAs universal treatment standard limits. Long-term leaching tests, per ANS 16. 1 procedure, yields leachability index for nitrate ions > 12. Chloride ions are expected to have an even higher (i.e., better) leachability index. Structural performance of these final waste forms, as indicated by compression strength and durability in aqueous environments, satisfies the regulatory criteria. Thus, based on the results of this study, it seems that phosphate ceramics are viable option for containment of salt wastes. 12. Risk assessment of nonhazardous oil-field waste disposal in salt caverns. SciTech Connect Elcock, D. 1998-03-10 Salt caverns can be formed in underground salt formations incidentally as a result of mining or intentionally to create underground chambers for product storage or waste disposal. For more than 50 years, salt caverns have been used to store hydrocarbon products. Recently, concerns over the costs and environmental effects of land disposal and incineration have sparked interest in using salt caverns for waste disposal. Countries using or considering using salt caverns for waste disposal include Canada (oil-production wastes), Mexico (purged sulfates from salt evaporators), Germany (contaminated soils and ashes), the United Kingdom (organic residues), and the Netherlands (brine purification wastes). In the US, industry and the regulatory community are pursuing the use of salt caverns for disposal of oil-field wastes. In 1988, the US Environmental Protection Agency (EPA) issued a regulatory determination exempting wastes generated during oil and gas exploration and production (oil-field wastes) from federal hazardous waste regulations--even though such wastes may contain hazardous constituents. At the same time, EPA urged states to tighten their oil-field waste management regulations. The resulting restrictions have generated industry interest in the use of salt caverns for potentially economical and environmentally safe oil-field waste disposal. Before the practice can be implemented commercially, however, regulators need assurance that disposing of oil-field wastes in salt caverns is technically and legally feasible and that potential health effects associated with the practice are acceptable. In 1996, Argonne National Laboratory (ANL) conducted a preliminary technical and legal evaluation of disposing of nonhazardous oil-field wastes (NOW) into salt caverns. It investigated regulatory issues; the types of oil-field wastes suitable for cavern disposal; cavern design and location considerations; and disposal operations, closure and remediation issues. It determined that if caverns are sited and designed well, operated carefully, closed properly, and monitored routinely, they could, from technical and legal perspectives, be suitable for disposing of oil-field wastes. On the basis of these findings, ANL subsequently conducted a preliminary risk assessment on the possibility that adverse human health effects (carcinogenic and noncarcinogenic) could result from exposure to contaminants released from the NOW disposed of in salt caverns. The methodology for the risk assessment included the following steps: identifying potential contaminants of concern; determining how humans could be exposed to these contaminants; assessing contaminant toxicities; estimating contaminant intakes; and estimating human cancer and noncancer risks. To estimate exposure routes and pathways, four postclosure cavern release scenarios were assessed. These were inadvertent cavern intrusion, failure of the cavern seal, failure of the cavern through cracks, failure of the cavern through leaky interbeds, and partial collapse of the cavern roof. Assuming a single, generic, salt cavern and generic oil-field wastes, potential human health effects associated with constituent hazardous substances (arsenic, benzene, cadmium, and chromium) were assessed under each of these scenarios. Preliminary results provided excess cancer risk and hazard index (for noncancer health effects) estimates that were well within the EPA target range for acceptable exposure risk levels. These results lead to the preliminary conclusion that from a human health perspective, salt caverns can provide an acceptable disposal method for nonhazardous oil-field wastes. 13. Options assessment report: Treatment of nitrate salt waste at Los Alamos National Laboratory SciTech Connect Robinson, Bruce Alan; Stevens, Patrice Ann 2015-09-16 This report documents the methodology used to select a method of treatment for the remediated nitrate salt (RNS) and unremediated nitrate salt (UNS) waste containers at Los Alamos National Laboratory (LANL). The method selected should treat the containerized waste in a manner that renders the waste safe and suitable for transport and final disposal in the Waste Isolation Pilot Plant (WIPP) repository, under specifications listed in the WIPP Waste Acceptance Criteria (DOE/CBFO, 2013). LANL recognized that the results must be thoroughly vetted with the New Mexico Environment Department (NMED) and the a modification to the LANL Hazardous Waste Facility Permit is a necessary step before implementation of this or any treatment option. Likewise, facility readiness and safety basis approvals must be received from the Department of Energy (DOE). This report presents LANL's preferred option, and the documentation of the process for reaching the recommended treatment option for RNS and UNS waste, and is presented for consideration by NMED and DOE. 14. Effect of dissolved LiCl on the ionic liquid-Au(111) interface: an in situ STM study NASA Astrophysics Data System (ADS) Borisenko, Natalia; Atkin, Rob; Lahiri, Abhishek; Zein El Abedin, Sherif; Endres, Frank 2014-07-01 The structure of the electrolyte/electrode interface plays a significant role in electrochemical processes. To date, most studies are focusing on understanding the interfacial structure in pure ionic liquids. In this paper in situ scanning tunnelling microscopy (STM) has been employed to elucidate the structure of the charged Au(111)-ionic liquid (1-butyl-1-methylpyrrolidinium tris(pentafluoroethyl)trifluorophosphate, [Py1,4]FAP) interface in the presence of 0.1 M LiCl. The addition of the Li salt to the ionic liquid has a strong influence on the interfacial structure. In the first STM scan in situ measurements reveal that Au(111) undergoes the (22 \\times \\surd 3) herringbone reconstruction in a certain potential regime, and there is strong evidence that the gold surface dissolves at negative electrode potentials in [Py1,4]FAP containing LiCl. Bulk deposition of Li is obtained at -2.9 V in the second STM scan. 15. Effect of dissolved LiCl on the ionic liquid-Au(111) interface: an in situ STM study. PubMed Borisenko, Natalia; Atkin, Rob; Lahiri, Abhishek; El Abedin, Sherif Zein; Endres, Frank 2014-07-16 The structure of the electrolyte/electrode interface plays a significant role in electrochemical processes. To date, most studies are focusing on understanding the interfacial structure in pure ionic liquids. In this paper in situ scanning tunnelling microscopy (STM) has been employed to elucidate the structure of the charged Au(111)-ionic liquid (1-butyl-1-methylpyrrolidinium tris(pentafluoroethyl)trifluorophosphate, [Py1,4]FAP) interface in the presence of 0.1 M LiCl. The addition of the Li salt to the ionic liquid has a strong influence on the interfacial structure. In the first STM scan in situ measurements reveal that Au(111) undergoes the (22 x ?3) 'herringbone' reconstruction in a certain potential regime, and there is strong evidence that the gold surface dissolves at negative electrode potentials in [Py1,4]FAP containing LiCl. Bulk deposition of Li is obtained at -2.9 V in the second STM scan. PMID:24919647 16. Brine migration in salt and its implications in the geologic disposal of nuclear waste SciTech Connect Jenks, G.H.; Claiborne, H.C. 1981-12-01 This report respresents a comprehensive review and analysis of available information relating to brine migration in salt surrounding radioactive waste in a salt repository. The topics covered relate to (1) the characteristics of salt formations and waste packages pertinent to considerations of rates, amounts, and effects of brine migration, (2) experimental and theoretical information on brine migration, and (3) means of designing to minimize any adverse effects of brine migration. Flooding, brine pockets, and other topics were not considered, since these features will presumably be eliminated by appropriate site selection and repository design. 115 references. 17. Study on a regeneration process of LiCl-KCl eutectic based waste salt generated from the pyrochemical process SciTech Connect Eun, H.C.; Cho, Y.Z.; Choi, J.H.; Kim, J.H.; Lee, T.K.; Park, H.S.; Kim, I.T.; Park, G.I. 2013-07-01 A regeneration process of LiCl-KCl eutectic waste salt generated from the pyrochemical process of spent nuclear fuel has been studied. This regeneration process is composed of a chemical conversion process and a vacuum distillation process. Through the regeneration process, a high efficiency of renewable salt recovery can be obtained from the waste salt and rare earth nuclides in the waste salt can be separated as oxide or phosphate forms. Thus, the regeneration process can contribute greatly to a reduction of the waste volume and a creation of durable final waste forms. (authors) 18. Using Aspen simulation package to determine solubility of mixed salts in TRU waste evaporator bottoms SciTech Connect Hatchell, J.L. 1998-03-01 Nitric acid from plutonium process waste is a candidate for waste minimization by recycling. Process simulation software packages, such as Aspen, are valuable tools to estimate how effective recovery processes can be, however, constants in equations of state for many ionic components are not in their data libraries. One option is to combine single salt solubilitys in the Aspen model for mixed salt system. Single salt solubilities were regressed in Aspen within 0.82 weight percent of literature values. These were combined into a single Aspen model and used in the mixed salt studies. A simulated nitric acid waste containing mixed aluminum, calcium, iron, magnesium and sodium nitrate was tested to determine points of solubility between 25 and 100 C. Only four of the modeled experimental conditions, at 50 C and 75 C, produced a saturated solution. While experimental results indicate that sodium nitrate is the first salt to crystallize out, the Aspen computer model shows that the most insoluble salt, magnesium nitrate, the first salt to crystallize. Possible double salt formation is actually taking place under experimental conditions, which is not captured by the Aspen model. 19. Alternative Electrochemical Salt Waste Forms, Summary of FY/CY2011 Results SciTech Connect Riley, Brian J.; McCloy, John S.; Crum, Jarrod V.; Rodriguez, Carmen P.; Windisch, Charles F.; Lepry, William C.; Matyas, Josef; Westman, Matthew P.; Rieck, Bennett T.; Lang, Jesse B.; Pierce, David A. 2011-12-01 This report summarizes the 2011 fiscal+calendar year efforts for developing waste forms for a spent salt generated in reprocessing nuclear fuel with an electrochemical separations process. The two waste forms are tellurite (TeO2-based) glasses and sol-gel-derived high-halide mineral analogs to stable minerals found in nature. 20. Secondary Aluminum Processing Waste: Salt Cake Characterization and Reactivity EPA Science Inventory Thirty-nine salt cake samples were collected from 10 SAP facilities across the U.S. The facilities were identified by the Aluminum Association to cover a wide range of processes. Results suggest that while the percent metal leached from salt cake was relatively low, the leachabl... 1. Secondary Aluminum Processing Waste: Salt Cake Characterization and Reactivity EPA Science Inventory Thirty-nine salt cake samples were collected from 10 SAP facilities across the U.S. The facilities were identified by the Aluminum Association to cover a wide range of processes. Results suggest that while the percent metal leached from the salt cake was relatively low, the leac... 2. Test procedures for polyester immobilized salt-containing surrogate mixed wastes SciTech Connect Biyani, R.K.; Hendrickson, D.W. 1997-07-18 These test procedures are written to meet the procedural needs of the Test Plan for immobilization of salt containing surrogate mixed waste using polymer resins, HNF-SD-RE-TP-026 and to ensure adequacy of conduct and collection of samples and data. This testing will demonstrate the use of four different polyester vinyl ester resins in the solidification of surrogate liquid and dry wastes, similar to some mixed wastes generated by DOE operations. 3. Radioactive waste isolation in salt: peer review of the Office of Nuclear Waste Isolation's report on Functional Design Criteria for a Repository for High-Level Radioactive Waste SciTech Connect Hambley, D.F.; Russell, J.E.; Busch, J.S.; Harrison, W.; Edgar, D.E.; Tisue, M.W. 1984-08-01 This report summarizes Argonne's review of the Office of Nuclear Waste Isolation's (ONWI's) draft report entitled Functional Design Criteria for High-Level Nuclear Waste Repository in Salt, dated January 23, 1984. Recommendations are given for improving the ONWI draft report. 4. Potential vertical movement of large heat-generating waste packages in salt. SciTech Connect Clayton, Daniel James; Martinez, Mario J.; Hardin, Ernest L. 2013-05-01 With renewed interest in disposal of heat-generating waste in bedded or domal salt formations, scoping analyses were conducted to estimate rates of waste package vertical movement. Vertical movement is found to result from thermal expansion, from upward creep or heave of the near-field salt, and from downward buoyant forces on the waste package. A two-pronged analysis approach was used, with thermal-mechanical creep modeling, and coupled thermal-viscous flow modeling. The thermal-mechanical approach used well-studied salt constitutive models, while the thermal-viscous approach represented the salt as a highly viscous fluid. The Sierra suite of coupled simulation codes was used for both approaches. The waste package in all simulations was a right-circular cylinder with the density of steel, in horizontal orientation. A time-decaying heat generation function was used to represent commercial spent fuel with typical burnup and 50-year age. Results from the thermal-mechanical base case showed approximately 27 cm initial uplift of the package, followed by gradual relaxation closely following the calculated temperature history. A similar displacement history was obtained with the package density set equal to that of salt. The slight difference in these runs is attributable to buoyant displacement (sinking) and is on the order of 1 mm in 2,000 years. Without heat generation the displacement stabilizes at a fraction of millimeter after a few hundred years. Results from thermal-viscous model were similar, except that the rate of sinking was constant after cooldown, at approximately 0.15 mm per 1,000 yr. In summary, all calculations showed vertical movement on the order of 1 mm or less in 2,000 yr, including calculations using well-established constitutive models for temperature-dependent salt deformation. Based on this finding, displacement of waste packages in a salt repository is not a significant repository performance issue. 5. Conceptual waste package interim product specifications and data requirements for disposal of borosilicate glass defense high-level waste forms in salt geologic repositories SciTech Connect Not Available 1983-06-01 The conceptual waste package interim product specifications and data requirements presented are applicable specifically to the normal borosilicate glass product of the Defense Waste Processing Facility (DWPF). They provide preliminary numerical values for the defense high-level waste form parameters and properties identified in the waste form performance specification for geologic isolation in salt repositories. Subject areas treated include containment and isolation, operational period safety, criticality control, waste form/production canister identification, and waste package performance testing requirements. This document was generated for use in the development of conceptual waste package designs in salt. It will be revised as additional data, analyses, and regulatory requirements become available. 6. Glovebox design requirements for molten salt oxidation processing of transuranic waste SciTech Connect Ramsey, K.B.; Acosta, S.V.; Wernly, K.D. 1998-12-31 This paper presents an overview of potential technologies for stabilization of {sup 238}Pu-contaminated combustible waste. Molten salt oxidation (MSO) provides a method for removing greater than 99.999% of the organic matrix from combustible waste. Implementation of MSO processing at the Los Alamos National Laboratory (LANL) Plutonium Facility will eliminate the combustible matrix from {sup 238}Pu-contaminated waste and consequently reduce the cost of TRU waste disposal operations at LANL. The glovebox design requirements for unit operations including size reduction and MSO processing will be presented. 7. Reduction of salt cake waste by removing low value oxide fines SciTech Connect Skoch, J.T.; Collins, R.L. 1995-12-31 With the ongoing pressure on the secondary aluminum business to limit the amount of waste generated from processing aluminum dross, it becomes very advantageous to preprocess the dross to reduce waste. The advantage of preprocessing is that it significantly reduces the oxides and salt cake generated when melted. Various products can be produced from the oxides for the steel industry. The paper will demonstrate that removal of oxide fines before melting will result in large volumes of material not being landfilled. The authors will show that the end result of this technology is to significantly reduce the amount of salt cake sent to a landfill while maximizing the recyclability of the aluminum dross. 8. Long-term isolation of high-level radioactive waste in salt repositories containing brine SciTech Connect Schweitzer, D.G.; Sastre, C.A. . Nuclear Energy Dept.) 1989-09-01 At present, only one concept, the Swedish design utilizing a thick-walled copper waste package, has been accepted as being capable of isolating high-level waste for hundreds of thousands of years in a granite-type repository. Theoretical arguments show that after the relatively short times required for salt consolidation self-shielded thin-walled copper waste packages have no significant failure or degradation reactions in anoxic neutral and acid brines. Thermodynamic analysis of reactions after consolidation (constant-volume reactions under lithostatic pressures in the absence of oxygen) show that miniscule amounts of metal reacting with brine can produce very large hydrogen pressures. Analyses of gamma radiolysis and diffusion processes show that in an array of thousands of waste packages removal of hydrogen from the outer packages should be negligible for a properly selected salt repository. 9. Alternative Electrochemical Salt Waste Forms, Summary of FY11-FY12 Results SciTech Connect Riley, Brian J.; Mccloy, John S.; Crum, Jarrod V.; Lepry, William C.; Rodriguez, Carmen P.; Windisch, Charles F.; Matyas, Josef; Westman, Matthew P.; Rieck, Bennett T.; Lang, Jesse B.; Olszta, Matthew J.; Pierce, David A. 2014-03-26 The Fuel Cycle Research and Development Program, sponsored by the U.S. Department of Energy Office of Nuclear Energy, is currently investigating alternative waste forms for wastes generated from nuclear fuel processing. One such waste results from an electrochemical separations process, called the “Echem” process. The Echem process utilizes a molten KCl-LiCl salt to dissolve the fuel. This process results in a spent salt containing alkali, alkaline earth, lanthanide halides and small quantities of actinide halides, where the primary halide is chloride with a minor iodide fraction. Pacific Northwest National Laboratory (PNNL) is concurrently investigating two candidate waste forms for the Echem spent-salt: high-halide minerals (i.e., sodalite and cancrinite) and tellurite (TeO2)-based glasses. Both of these candidates showed promise in fiscal year (FY) 2009 and FY2010 with a simplified nonradioactive simulant of the Echem waste. Further testing was performed on these waste forms in FY2011 and FY2012 to assess the possibility of their use in a sustainable fuel cycle. This report summarizes the combined results from FY2011 and FY2012 efforts. 10. LiCl inhibits PRRSV infection by enhancing Wnt/?-catenin pathway and suppressing inflammatory responses. PubMed Hao, Hong-ping; Wen, Li-bin; Li, Jia-rong; Wang, Yue; Ni, Bo; Wang, Rui; Wang, Xin; Sun, Ming-xia; Fan, Hong-jie; Mao, Xiang 2015-05-01 Lithium chloride (LiCl) has been used as a mood stabilizer in the manic depressive disorders treatment. Recent studies show that LiCl is also a potent inhibitor for some DNA and RNA viruses. Porcine reproductive and respiratory syndrome virus (PRRSV) is an important viral pathogen in modern pig industry. In this study, we assessed the inhibitory effect of LiCl on PRRSV infection using plaque-formation assay, Q-PCR and Western blot analysis. Our results showed that LiCl could inhibit PRRSV infection in MARC-145 and PAM-CD163 cells. Previous reports have shown that LiCl could induce the Wnt pathway in the absence of Wnt ligands. In our studies, we demonstrated that LiCl activates the Wnt pathway in PRRSV infected cells. Additionally, the knockdown of ?-catenin or the Wnt/?-catenin pathway inhibitor PNU74654 was able to reverse the antiviral effect of LiCl, which suggested that the inhibitory effect of LiCl against PRRSV replication might be associated with the activation of the Wnt/?-catenin pathway. We also found that lower viral replication after LiCl treatment was associated with the reduced mRNA levels of pro-inflammatory IL-8, IL-6, IL-1 ?, tumor necrosis factor ? and decreased NF-?B nuclear translocation. Collectively, our data demonstrated that LiCl inhibited PRRSV infection by enhancing Wnt/?-catenin pathway and suppressing pro-inflammatory responses. PMID:25746333 11. Caustic Recycle from Hanford Tank Waste Using NaSICON Ceramic Membrane Salt Splitting Process SciTech Connect Fountain, Matthew S.; Kurath, Dean E.; Sevigny, Gary J.; Poloski, Adam P.; Pendleton, J.; Balagopal, S.; Quist, M.; Clay, D. 2009-02-20 A family of inorganic ceramic materials, called sodium (Na) Super Ion Conductors (NaSICON), has been studied at Pacific Northwest National Laboratory (PNNL) to investigate their ability to separate sodium from radioactively contaminated sodium salt solutions for treating U.S. Department of Energy (DOE) tank wastes. Ceramatec Inc. developed and fabricated a membrane containing a proprietary NAS-GY material formulation that was electrochemically tested in a bench-scale apparatus with both a simulant and a radioactive tank-waste solution to determine the membrane performance when removing sodium from DOE tank wastes. Implementing this sodium separation process can result in significant cost savings by reducing the disposal volume of low-activity wastes and by producing a NaOH feedstock product for recycle into waste treatment processes such as sludge leaching, regenerating ion exchange resins, inhibiting corrosion in carbon-steel tanks, or retrieving tank wastes. 12. Fluid Transport Driven by Heat-Generating Nuclear Waste in Bedded Salt NASA Astrophysics Data System (ADS) Jordan, A.; Harp, D. R.; Stauffer, P. H.; Ten Cate, J. A.; Labyed, Y.; Boukhalfa, H.; Lu, Z.; Person, M. A.; Robinson, B. A. 2013-12-01 The question of where to safely dispose high-level nuclear waste (HLW) provides ample motivation for scientific research on deep geologic disposal options. The goal of this study is to model the dominant heat and mass transport processes that would be driven by heat generating nuclear waste buried in bedded salt. The interaction between liquid brine flow towards the heat source, establishment of a heat pipe in the mine-run salt backfill, boiling, and vapor condensation leads to changes in porosity, permeability, saturation, thermal conductivity, and rheology of the salt surrounding potential waste canisters. The Finite Element Heat and Mass transfer code (FEHM) was used to simulate these highly coupled thermal, hydrological, and chemical processes. The numerical model has been tested against recent and historical experimental data to develop and improve the salt material model. We used the validated numerical model to make predictions of temperature gradients, porosity changes, and tracer behavior that will be testable in a future 2-year field-scale heater experiment to be carried out in an experimental test bed at the Waste Isolation Pilot Plant (WIPP) site near Carlsbad, NM. 13. Mineralocorticoid replacement during infancy for salt wasting congenital adrenal hyperplasia due to 21-hydroxylase deficiency PubMed Central Gomes, Larissa G.; Madureira, Guiomar; Mendonca, Berenice B.; Bachega, Tania A. S. S. 2013-01-01 OBJECTIVE: The protocols for glucocorticoid replacement in children with salt wasting 21-hydroxylase deficiency are well established; however, the current recommendation for mineralocorticoid replacement is general and suggests individualized dose adjustments. This study aims to retrospectively review the 9-?-fludrocortisone dose regimen in salt wasting 21-hydroxylase deficient children who have been adequately treated during infancy. METHODS: Twenty-three salt wasting 21-hydroxylase deficient patients with good anthropometric and hormonal control were followed in our center since diagnosis. The assessments of cortisone acetate and 9-?-fludrocortisone doses, anthropometric parameters, and biochemical and hormonal levels were rigorously evaluated in pre-determined intervals from diagnosis to two years of age. RESULTS: The 9-?-fludrocortisone doses decreased over time during the first and second years of life; the median fludrocortisone doses were 200 ?g at 0-6 months, 150 ?g at 7-18 months and 125 ?g at 19-24 months. The cortisone acetate dose per square meter was stable during follow-up (median?=?16.8 mg/m2/day). The serum sodium, potassium and plasma rennin activity levels during treatment were normal, except in the first month of life, when periodic 9-?-fludrocortisone dose adjustments were made. CONCLUSIONS: The mineralocorticoid needs of salt wasting 21-hydroxylase deficient patients are greater during early infancy and progressively decrease during the first two years of life, which confirms that a partial aldosterone resistance exists during this time. Our study proposes a safety regiment for mineralocorticoid replacement during this critical developmental period. PMID:23525308 14. Preliminary Technical and Legal Evaluation of Disposing of Nonhazardous Oil Field Waste into Salt Caverns SciTech Connect Ayers, Robert C.; Caudle, Dan; Elcock, Deborah; Raivel, Mary; Veil, John; and Grunewald, Ben 1999-01-21 This report presents an initial evaluation of the suitability, feasibility, and legality of using salt caverns for disposal of nonhazardous oil field wastes. Given the preliminary and general nature of this report, we recognize that some of our findings and conclusions maybe speculative and subject to change upon further research on this topic. 15. Direct Grout Stabilization of High Cesium Salt Waste: Cesium Leaching Studies SciTech Connect Langton, C.A. 1999-09-19 'The direct grout alternative is a viable option for treatment/stabilization and disposal of salt waste containing Cs-137 concentrations of 1-3 Ci/gal. The significant difference between these waste solutions is that the high cesium salt solution will contain between 1 and 3 Curies of Cs-137 per gallon compared to a negligible amount in the current salt solution. This difference will require special engineering and shielding for a direct grout processing facility and disposal units to achieve acceptable radiation exposure conditions. The higher cesium concentrations in the direct grout also require that the cesium leaching be evaluated as a function of curing temperature. ANS 16.1 leaching results and distribution ratios (approximations of distribution coefficients) as a function of temperature are presented in this report.' 16. Disposal of NORM-Contaminated Oil Field Wastes in Salt Caverns SciTech Connect Blunt, D.L.; Elcock, D.; Smith, K.P.; Tomasko, D.; Viel, J.A.; and Williams, G.P. 1999-01-21 In 1995, the U.S. Department of Energy (DOE), Office of Fossil Energy, asked Argonne National Laboratory (Argonne) to conduct a preliminary technical and legal evaluation of disposing of nonhazardous oil field waste (NOW) into salt caverns. That study concluded that disposal of NOW into salt caverns is feasible and legal. If caverns are sited and designed well, operated carefully, closed properly, and monitored routinely, they can be a suitable means of disposing of NOW (Veil et al. 1996). Considering these findings and the increased U.S. interest in using salt caverns for NOW disposal, the Office of Fossil Energy asked Argonne to conduct further research on the cost of cavern disposal compared with the cost of more traditional NOW disposal methods and on preliminary identification and investigation of the risks associated with such disposal. The cost study (Veil 1997) found that disposal costs at the four permitted disposal caverns in the United States were comparable to or lower than the costs of other disposal facilities in the same geographic area. The risk study (Tomasko et al. 1997) estimated that both cancer and noncancer human health risks from drinking water that had been contaminated by releases of cavern contents were significantly lower than the accepted risk thresholds. Since 1992, DOE has funded Argonne to conduct a series of studies evaluating issues related to management and disposal of oil field wastes contaminated with naturally occurring radioactive material (NORM). Included among these studies were radiological dose assessments of several different NORM disposal options (Smith et al. 1996). In 1997, DOE asked Argonne to conduct additional analyses on waste disposal in salt caverns, except that this time the wastes to be evaluated would be those types of oil field wastes that are contaminated by NORM. This report describes these analyses. Throughout the remainder of this report, the term ''NORM waste'' is used to mean ''oil field waste contaminated by NORM''. 17. Engineering study of the potential uses of salts from selective crystallization of Hanford tank wastes SciTech Connect Hendrickson, D.W. 1996-04-30 The Clean Salt Process (CSP) is the fractional crystallization of nitrate salts from tank waste stored on the Hanford Site. This study reviews disposition options for a CSP product made from Hanford Site tank waste. These options range from public release to onsite low-level waste disposal to no action. Process, production, safety, environment, cost, schedule, and the amount of CSP material which may be used are factors considered in each option. The preferred alternative is offsite release of clean salt. Savings all be generated by excluding the material from low-level waste stabilization. Income would be received from sales of salt products. Savings and income from this alternative amount to$1,027 million, excluding the cost of CSP operations. Unless public sale of CSP products is approved, the material should be calcined. The carbonate form of the CSP could then be used as ballast in tank closure and stabilization efforts. Not including the cost of CSP operations, savings of $632 million would be realized. These savings would result from excluding the material from low-level waste stabilization and reducing purchases of chemicals for caustic recycle and stabilization and closure. Dose considerations for either alternative are favorable. No other cost-effective alternatives that were considered had the capacity to handle significant quantities of the CSP products. If CSP occurs, full-scale tank-waste stabilization could be done without building additional treatment facilities after Phase 1 (DOE 1996). Savings in capital and operating cost from this reduction in waste stabilization would be in addition to the other gains described. 18. Radioactive waste isolation in salt: geochemistry of brine in rock salt in temperature gradients and gamma-radiation fields - a selective annotated bibliography SciTech Connect Hull, A.B.; Williams, L.B. 1985-07-01 Evaluation of the extensive research concerning brine geochemistry and transport is critically important to successful exploitation of a salt formation for isolating high-level radioactive waste. This annotated bibliography has been compiled from documents considered to provide classic background material on the interactions between brine and rock salt, as well as the most important results from more recent research. Each summary elucidates the information or data most pertinent to situations encountered in siting, constructing, and operating a mined repository in salt for high-level radioactive waste. The research topics covered include the basic geology, depositional environment, mineralogy, and structure of evaporite and domal salts, as well as fluid inclusions, brine chemistry, thermal and gamma-radiation effects, radionuclide migration, and thermodynamic properties of salts and brines. 4 figs., 6 tabs. 19. The advantages of a salt/bentonite backfill for Waste Isolation Pilot Plant disposal rooms SciTech Connect Butcher, B.M.; Novak, C.F. ); Jercinovic, M. ) 1991-04-01 A 70/30 wt% salt/bentonite mixture is shown to be preferable to pure crushed salt as backfill for disposal rooms in the Waste Isolation Pilot Plant (WIPP). This report discusses several selection criteria used to arrive at this conclusion: the need for low permeability and porosity after closure, chemical stability with the surroundings, adequate strength to avoid shear erosion from human intrusion, ease of emplacement, and sorption potential for brine and radionuclides. Both salt and salt/bentonite are expected to consolidate to a final state of impermeability (i.e., {le} 10{sup {minus}18}m{sup 2}) adequate for satisfying federal nuclear regulations. Any advantage of the salt/bentonite mixture is dependent upon bentonite's potential for sorbing brine and radionuclides. Estimates suggest that bentonite's sorption potential for water in brine is much less than for pure water. While no credit is presently taken for brine sorption in salt/bentonite backfill, the possibility that some amount of inflowing brine would be chemically bound is considered likely. Bentonite may also sorb much of the plutonium, americium, and neptunium within the disposal room inventory. Sorption would be effective only if a major portion of the backfill is in contact with radioactive brine. Brine flow from the waste out through highly localized channels in the backfill would negate sorption effectiveness. Although the sorption potentials of bentonite for both brine and radionuclides are not ideal, they are distinctly beneficial. Furthermore, no detrimental aspects of adding bentonite to the salt as a backfill have been identified. These two observations are the major reasons for selecting salt/bentonite as a backfill within the WIPP. 39 refs., 16 figs., 6 tabs. 20. Can nonhazardous oil field wastes be disposed of in salt caverns? SciTech Connect Veil, J.A. 1996-10-01 Solution-mined salt caverns have been used for many years for storing hydrocarbon products. This paper summarizes an Argonne National Laboratory report that reviews the legality, technical suitability, and feasibility of disposing of nonhazardous oil and gas exploration and production wastes in salt caverns. An analysis of regulations indicated that there are no outright regulatory prohibitions on cavern disposal -of oil field wastes at either the federal level or in the 11 oil-producing states that were studied. There is no actual field experience on the long-term impacts that might arise following closure of waste disposal caverns. Although research has found that pressures will build up in a closed cavern, none has specifically addressed caverns filled with oil field wastes. More field research on pressure build up in closed caverns is needed. On the basis of preliminary investigations, we believe that disposal of oil field wastes in salt caverns is legal and feasible. The technical suitability of the practice depends on whether the caverns are well-sited and well-designed, carefully operated, properly closed, and routinely monitored. 1. New information on disposal of oil field wastes in salt caverns SciTech Connect Veil, J.A. 1996-10-01 Solution-mined salt caverns have been used for many years for storing hydrocarbon products. This paper summarizes an Argonne National Laboratory report that reviews the legality, technical suitability, and feasibility of disposing of nonhazardous oil and gas exploration and production wastes in salt caverns. An analysis of regulations indicated that there are no outright regulatory prohibitions on cavern disposal of oil field wastes at either the federal level or in the 11 oil-producing states that were studied. There is no actual field experience on the long-term impacts that might arise following closure of waste disposal caverns. Although research has found that pressures will build-up in a closed cavern, none has specifically addressed caverns filled with oil field wastes. More field research on pressure build-up in closed caverns is needed. On the basis of preliminary investigations, we believe that disposal of oil field wastes in salt caverns is legal and feasible. The technical suitability of the practice depends on whether the caverns are well-sited and well-designed, carefully operated, properly closed, and routinely monitored. 2. Options Assessment Report: Treatment of Nitrate Salt Waste at Los Alamos National Laboratory SciTech Connect Robinson, Bruce Alan; Stevens, Patrice Ann 2015-12-17 This report documents the methodology used to select a method of treatment for the remediated nitrate salt (RNS) and unremediated nitrate salt (UNS) waste containers at Los Alamos National Laboratory (LANL). The method selected should treat the containerized waste in a manner that renders the waste safe and suitable for transport and final disposal in the Waste Isolation Pilot Plant (WIPP) repository, under specifications listed in the WIPP Waste Acceptance Criteria (DOE/CBFO, 2013). LANL recognizes that the results must be thoroughly vetted with the New Mexico Environment Department (NMED) and that a modification to the LANL Hazardous Waste Facility Permit is a necessary step before implementation of this or any treatment option. Likewise, facility readiness and safety basis approvals must be received from the Department of Energy (DOE). This report presents LANL’s preferred option, and the documentation of the process for reaching the recommended treatment option for RNS and UNS waste, and is presented for consideration by NMED and DOE. 3. Molten salt oxidation of mixed waste: Preliminary bench-scale experiments without radioactivity SciTech Connect Haas, P.A.; Rudolph, J.C.; Bell, J.T. 1994-06-01 Molten salt oxidation (MSO) is a process in which organic wastes are oxidized by sparging them with air through a bed of molten sodium carbonate (bp 851 {degrees}C) at {ge} 900{degrees}C. This process is readily applicable to the mixed waste because acidic products from Cl, S, P, etc., in the waste, along with most metals and most radionuclides, are retained within the melt as oxides or salts. Rockwell International has studied the application of MSO to various wastes, including some mixed waste. A unit used by Rockwell to study the mixed waste treatment is presently in use at Oak Ridge National Laboratory (ORNL). ORNLs studies to date have concentrated on chemical flowsheet questions. Concerns that were studied included carbon monoxide (CO) emissions, NO{sub x}, emissions, and metal retention under a variety of conditions. Initial experiments show that CO emissions increase with increasing NaCl content in the melt, increasing temperature, and increasing airflow. Carbon monoxide content is especially high (> 2000 ppm) with high chlorine content (> 10%). Thermal NO{sub x}, emissions are relatively low ( < 5 ppm) at temperatures < 1000{degrees}C. However, most (85--100%) of the nitrogen in the feed as organic nitrate or amine was released as NO{sub x}, The metal contents of the melt and of knockout pot samples of condensed salt show high volatilities of Cs as CsCl. Average condensed salt concentrations were 60% for barium and 100% for strontium and cobalt. The cerium disappeared -- perhaps from deposition on the alumina reactor walls. 4. Ion Recognition Approach to Volume Reduction of Alkaline Tank Waste by Separation of Sodium Salts SciTech Connect Levitskaia, Tatiana G.; Lumetta, Gregg J.; Moyer, Bruce A.; Bonnesen, Peter V. 2006-06-01 The purpose of this research involving collaboration between Oak Ridge National Laboratory (ORNL) and Pacific Northwest National Laboratory (PNNL) is to explore new approaches to the separation of sodium hydroxide, sodium nitrate, and other sodium salts from high-level alkaline tank waste. The principal potential benefit is a major reduction in disposed waste volume, obviating the building of expensive new waste tanks and reducing the costs of low-activity waste immobilization. Principles of ion recognition are being researched toward discovery of liquid extraction systems that selectively separate sodium hydroxide and sodium nitrate from other waste components. The successful concept of pseudohydroxide extraction using fluorinated alcohols and phenols is being developed at ORNL and PNNL toward a greater understanding of the controlling equilibria, role of solvation, and of synergistic effects involving crown ethers. Studies at PNNL are directed toward new solvent formulation for the practical sodium pseudohydroxide extraction systems. 5. Subsurface geology of a potential waste emplacement site, Salt Valley Anticline, Grand County, Utah USGS Publications Warehouse Hite, R.J. 1977-01-01 The Salt Valley anticline, which is located about 32 km northeast of Moab, Utah, is perhaps one of the most favorable waste emplacement sites in the Paradox basin. The site, which includes about 7.8 km 2, is highly accessible and is adjacent to a railroad. The anticline is one of a series of northwest-trending salt anticlines lying along the northeast edge of the Paradox basin. These anticlines are cored by evaporites of the Paradox Member of the Hermosa Formation of Middle Pennsylvanian age. The central core of the Salt Valley anticline forms a ridgelike mass of evaporites that has an estimated amplitude of 3,600 m. The evaporite core consists of about 87 percent halite rock, which includes some potash deposits; the remainder is black shale, silty dolomite, and anhydrite. The latter three lithologies are referred to as 'marker beds.' Using geophysical logs from drill holes on the anticline, it is possible to demonstrate that the marker beds are complexly folded and faulted. Available data concerning the geothermal gradient and heatflow at the site indicate that heat from emplaced wastes should be rapidly dissipated. Potentially exploitable resources of potash and petroleum are present at Salt Valley. Development of these resources may conflict with use of the site for waste emplacement. 6. Molten salt destruction of energetic material wastes as an alternative to open burning. Revision 1 SciTech Connect Upadhye, R.S.; Brummond, W.A.; Pruneda, C.O.; Watkins, B.E. 1994-11-02 As a result of the end of the Cold War and the shift in emphasis to a smaller stockpile, many munitions, both conventional and nuclear, are scheduled for retirement and rapid dismantlement and demilitarization. Major components of these munitions are the explosives and propellants, or energetic materials. The Molten Salt Destruction (MSD) Process has been demonstrated for the destruction of HE and HE-containing wastes. MSD converts the organic constituents of the waste into non-hazardous substances such as carbon dioxide, nitrogen and water. Any inorganic constituents of the waste, such as binders and metallic particles, are retained in the molten salt. LLNL has built a small-scale (about 1 kg/hr throughput) unit to test the destruction of HE using the MSD process. The authors have demonstrated that HEs and liquid propellants can be safely and fully destroyed using the molten salt destruction process. The authors are currently working on a number of improvements to the process. They are modifying the design of unit to obtain more throughput without any increase in salt entrainment. They are implementing an advanced nozzle design for injection of larger particles. They are defining operating envelopes for a number of high explosives and formulations. They are developing models to study the temperature profile of a top-feed nozzle for feeding larger particles into the unit. 7. Room closure response to gas generation and mechanical strength of different waste forms in a bedded salt repository SciTech Connect Mendenhall, F.T.; Stone, C.M. 1993-05-01 Finite element calculations of the porosity history of a nuclear waste disposal room in a bedded salt formation have been completed. The analyses include an elastic/secondary creep model for the host halite and a nonlinear consolidation model for the crushed salt backfill. Separate gas generation and constitutive models were used for three distinct waste forms, (1) unaltered defense related CH-TRU waste, (2) shredded and cemented CH-TRU waste, and (3) incinerated and vitrified CH-TRU waste. Solutions were determined for a 2000 year time period starting from the decommissioning of the repository. The resulting room porosities varied from roughly 55% to less than 10%. 8. Non-Radiological Air Quality Modeling for the High-Level Waste Salt Disposition SciTech Connect Hunter, C.H. 1999-11-29 Dispersion modeling of non-radiological airborne emissions associated with the construction and operation of three alternatives for high-level waste salt disposition at the Savannah River Site has been completed. The results will be used by Department of Energy-Savannah River in the preparation of the salt disposition supplemental environmental impact statement. Estimated maximum ground-level concentrations of applicable regulated air pollutants of the site boundary and at the distance to a hypothetical, co-located onsite worker are summarized in tables. In all cases, model estimated ambient concentrations are less than regulatory standards. 9. Characteristics of wasteform composing of phosphate and silicate to immobilize radioactive waste salts. PubMed Park, Hwan-Seo; Cho, In-Hak; Eun, Hee Chul; Kim, In-Tae; Cho, Yong Zun; Lee, Han-Soo 2011-03-01 In the radioactive waste management, metal chloride wastes from a pyrochemical process is one of problematic wastes not directly applicable to a conventional solidification process. Different from a use of minerals or a specific phosphate glass for immobilizing radioactive waste salts, our research group applied an inorganic composite, SAP (SiO(2)-Al(2)O(3)-P(2)O(5)), to stabilize them by dechlorination. From this method, a unique wasteform composing of phosphate and silicate could be fabricated. This study described the characteristic of the wasteform on the morphology, chemical durability, and some physical properties. The wasteform has a unique "domain-matrix" structure which would be attributed to the incompatibility between silicate and phosphate glass. At higher amounts of chemical binder, "P-rich phase encapsulated by Si-rich phase" was a dominant morphology, but it was changed to be Si-rich phase encapsulated by P-rich phase at a lower amount of binder. The domain and subdomain size in the wasteform was about 0.5-2 ?m and hundreds of nm, respectively. The chemical durability of wasteform was confirmed by various leaching test methods (PCT-A, ISO dynamic leaching test, and MCC-1). From the leaching tests, it was found that the P-rich phase had ten times lower leach-resistance than the Si-rich phase. The leach rates of Cs and Sr in the wasteform were about 10(-3)g/m(2) day, and the leached fractions of them were about 0.04% and 0.06% at 357 days, respectively. Using this method, we could stabilize and solidify the waste salt to form a monolithic wasteform with good leach-resistance. Also, the decrease of waste volume by the dechlorination approach would be beneficial in the final disposal cost, compared with the present immobilization methods for waste salt. PMID:21288037 10. Waste segregation analysis for salt well pumping in the 200 W Area -- Task 3.4 SciTech Connect Reynolds, D.A. 1995-04-28 There is an estimated 7 million liters (1.9 million gallons) of potentially complexed waste that need to be pumped from single-shell tanks (SST) in the 200 West Area. This represents up to 40% of the salt well liquor that needs to be pumped in the 200 West Area. There are three double-shell (DST) tanks in the 241-SY tank farm in the 200 West Area. Tank 241-SY-101 is full and not usable. Tank 241-SY-102 has a transuranic (TRU) sludge in the bottom. Current rules prohibit mixing complexed waste with TRU waste. Tank 241-SY-103 has three major problems. First, 241-SY-103 is on the Flammable Watch list. Second, adding waste to tank 241-SY-103 has the potential for an episodic release of hydrogen gas. Third, 241-SY-103 will not hold all of the potentially complexed waste from the SSTs. This document looks at more details regarding the salt well pumping of the 200 West Area tank farm. Some options are considered but it is beyond the scope of this document to provide an in-depth study necessary to provide a defensible solution to the complexed waste problem. 11. Hydrous mineral dehydration around heat-generating nuclear waste in bedded salt formations. PubMed Jordan, Amy B; Boukhalfa, Hakim; Caporuscio, Florie A; Robinson, Bruce A; Stauffer, Philip H 2015-06-01 Heat-generating nuclear waste disposal in bedded salt during the first two years after waste emplacement is explored using numerical simulations tied to experiments of hydrous mineral dehydration. Heating impure salt samples to temperatures of 265 C can release over 20% by mass of hydrous minerals as water. Three steps in a series of dehydration reactions are measured (65, 110, and 265 C), and water loss associated with each step is averaged from experimental data into a water source model. Simulations using this dehydration model are used to predict temperature, moisture, and porosity after heating by 750-W waste canisters, assuming hydrous mineral mass fractions from 0 to 10%. The formation of a three-phase heat pipe (with counter-circulation of vapor and brine) occurs as water vapor is driven away from the heat source, condenses, and flows back toward the heat source, leading to changes in porosity, permeability, temperature, saturation, and thermal conductivity of the backfill salt surrounding the waste canisters. Heat pipe formation depends on temperature, moisture availability, and mobility. In certain cases, dehydration of hydrous minerals provides sufficient extra moisture to push the system into a sustained heat pipe, where simulations neglecting this process do not. PMID:25965632 12. Equipment evaluation for low density polyethylene encapsulated nitrate salt waste at the Rocky Flats Plant SciTech Connect Yamada, W.I.; Faucette, A.M.; Jantzen, R.C.; Logsdon, B.W.; Oldham, J.H.; Saiki, D.M.; Yudnich, R.J. 1993-08-30 Mixed wastes at the Rocky Flats Plant (RFP) are subject to regulation by the Resource Conservation and Recovery Act (RCRA). Polymer solidification is being developed as a final treatment technology for several of these mixed wastes, including nitrate salts. Encapsulation nitrate salts with low density polyethylene (LDPE) has been the preliminary focus of the RFP polymer solidification effort. Literature reviews, industry surveys, and lab-scale and pilot-scale tests have been conducted to evaluate several options for encapsulating nitrate salts with LDPE. Most of the effort has focused on identifying compatible drying and extrusion technologies. Other processing options, specifically meltration and non-heated compounding machines, were also investigated. The best approach appears to be pretreatment of the nitrate salt waste brine in either a vertical or horizontal thin film evaporator followed by compounding of the dried waste with LDPE in an intermeshing, co-rotating, twin-screw extruder. Additional pilot-scale tests planned for the fall of 1993 should further support this recommendation. Preliminary evaluation work indicates that meltration is not possible at atmospheric pressure with the LDPE (Chevron PE-1409) provided by RFP. However, meltration should be possible at atmospheric pressure using another LDPE formulation with altered physical and rheological properties: Lower molecular weight and lower viscosity (Epoline C-15). Contract modifications are now in process to allow a follow-on pilot scale demonstration. Questions regarding changed safety and physical properties of the resultant LDPE waste form due to use of the Epoline C-15 will be addressed. No additional work with non-heated mixer compounder machines is planned at this time. 13. Phase Equilibrium Studies of Savannah River Tanks and Feed Streams for the Salt Waste Processing Facility SciTech Connect Weber, C.F. 2001-06-19 A chemical equilibrium model is developed and used to evaluate supersaturation of tanks and proposed feed streams to the Salt Waste Processing Facility. The model uses Pitzer's model for activity coefficients and is validated by comparison with a variety of thermodynamic data. The model assesses the supersaturation of 13 tanks at the Savannah River Site (SRS), indicating that small amounts of gibbsite and or aluminosilicate may form. The model is also used to evaluate proposed feed streams to the Salt Waste Processing Facility for 13 years of operation. Results indicate that dilutions using 3-4 M NaOH (about 0.3-0.4 L caustic per kg feed solution) should avoid precipitation and reduce the Na{sup +} ion concentration to 5.6 M. 14. Conditioning matrices from high level waste resulting from pyrochemical processing in fluorine salt SciTech Connect Grandjean, Agnes; Advocat, Thierry; Bousquet, Nicolas; Jegou, Christophe 2007-07-01 Separating the actinides from the fission products through reductive extraction by aluminium in a LiF/AlF{sub 3} medium is a process investigated for pyrometallurgical reprocessing of spent fuel. The process involves separation by reductive salt-metal extraction. After dissolving the fuel or the transmutation target in a salt bath, the noble metal fission products are first extracted by contacting them with a slightly reducing metal. After extracting the metal fission products, then the actinides are selectively separated from the remaining fission products. In this hypothesis, all the unrecoverable fission products would be conditioned as fluorides. Therefore, this process will generate first a metallic waste containing the 'reducible' fission products (Pd, Mo, Ru, Rh, Tc, etc.) and a fluorine waste containing alkali-metal, alkaline-earth and rare earth fission products. Immobilization of these wastes in classical borosilicate glasses is not feasible due to the very low solubility of noble metals, and of fluoride in these hosts. Alternative candidates have therefore been developed including silicate glass/ceramic system for fluoride fission products and metallic ones for noble metal fission products. These waste-forms were evaluated for their confinement properties like homogeneity, waste loading, volatility during the elaboration process, chemical durability, etc. using appropriate techniques. (authors) 15. Estimated human health risks of disposing of nonhazardous oil field waste in salt caverns SciTech Connect Tomasko, D.; Elcock, D.; Veil, J. 1997-09-01 Argonne National Laboratory (ANL) has completed an evaluation of the possibility that adverse human health effects (carcinogenic and noncarcinogenic) could result from exposure to contaminants released from nonhazardous oil field wastes (NOW) disposed in domal salt caverns. In this assessment, several steps were used to evaluate potential human health risks: identifying potential contaminants of concern, determining how humans could be exposed to these contaminants, assessing the contaminants toxicities, estimating contaminant intakes, and, finally, calculating human cancer and noncancer risks. 16. Risk assessment of nonhazardous oil-field waste disposal in salt caverns. SciTech Connect Elcock, D. 1998-03-05 In 1996, Argonne National Laboratory (ANL) conducted a preliminary technical and legal evaluation of disposing of nonhazardous oil-field wastes (NOW) into salt caverns. Argonne determined that if caverns are sited and designed well, operated carefully, closed properly, and monitored routinely, they could be suitable for disposing of oil-field wastes. On the basis of these findings, Argonne subsequently conducted a preliminary evaluation of the possibility that adverse human health effects (carcinogenic and noncarcinogenic) could result from exposure to contaminants released from the NOW disposed of in domal salt caverns. Steps used in this evaluation included the following: identifying potential contaminants of concern, determining how humans could be exposed to these contaminants, assessing contaminant toxicities, estimating contaminant intakes, and calculating human cancer and noncancer risk estimates. Five postclosure cavern release scenarios were assessed. These were inadvertent cavern intrusion, failure of the cavern seal, failure of the cavern through cracks, failure of the cavern through leaky interbeds, and a partial collapse of the cavern roof. Assuming a single, generic, salt cavern and generic oil-field wastes, potential human health effects associated with constituent hazardous substances (arsenic, benzene, cadmium, and chromium) were assessed under each of these scenarios. Preliminary results provided excess cancer risk and hazard index (referring to noncancer health effects) estimates that were well within the US Environmental Protection Agency (EPA) target range for acceptable exposure risk levels. These results led to the preliminary conclusion that from a human health perspective, salt caverns can provide an acceptable disposal method for nonhazardous oil-field wastes. 17. Risk analyses for disposing nonhazardous oil field wastes in salt caverns SciTech Connect Tomasko, D.; Elcock, D.; Veil, J.; Caudle, D. 1997-12-01 Salt caverns have been used for several decades to store various hydrocarbon products. In the past few years, four facilities in the US have been permitted to dispose nonhazardous oil field wastes in salt caverns. Several other disposal caverns have been permitted in Canada and Europe. This report evaluates the possibility that adverse human health effects could result from exposure to contaminants released from the caverns in domal salt formations used for nonhazardous oil field waste disposal. The evaluation assumes normal operations but considers the possibility of leaks in cavern seals and cavern walls during the post-closure phase of operation. In this assessment, several steps were followed to identify possible human health risks. At the broadest level, these steps include identifying a reasonable set of contaminants of possible concern, identifying how humans could be exposed to these contaminants, assessing the toxicities of these contaminants, estimating their intakes, and characterizing their associated human health risks. The contaminants of concern for the assessment are benzene, cadmium, arsenic, and chromium. These were selected as being components of oil field waste and having a likelihood to remain in solution for a long enough time to reach a human receptor. 18. Aspects of the thermal and transport properties of crystalline salt in designing radioactive waste storages in halogen formations SciTech Connect Nikitin, A. N. Pocheptsova, O. A.; Matthies, S. 2010-05-15 Some of the properties of natural rock salt are described. This rock is of great practical interest, because, along with its conventional applications in the chemical and food industries, it is promising for use in engineering underground radioactive waste storages and natural gas reservoirs. The results of structural and texture studies of rock salt by neutron diffraction are discussed. The nature of the salt permeability under temperature and stress gradients is theoretically estimated. 19. Aspects of the thermal and transport properties of crystalline salt in designing radioactive waste storages in halogen formations NASA Astrophysics Data System (ADS) Nikitin, A. N.; Pocheptsova, O. A.; Matthies, S. 2010-05-01 Some of the properties of natural rock salt are described. This rock is of great practical interest, because, along with its conventional applications in the chemical and food industries, it is promising for use in engineering underground radioactive waste storages and natural gas reservoirs. The results of structural and texture studies of rock salt by neutron diffraction are discussed. The nature of the salt permeability under temperature and stress gradients is theoretically estimated. 20. USING MINED SPACE FOR LONG-TERM RETENTION OF NONRADIOACTIVE HAZARDOUS WASTE. VOLUME 2. SOLUTION MINED SALT CAVERNS EPA Science Inventory This two-volume report assesses the current status of using mined-space for long-term retention of nonradioactive hazardous waste. Volume 2 expands the definition of mined space to include that created by solution mining of salt. This report examines the extent of salt deposits i... 1. Crystallization of rhenium salts in a simulated low-activity waste borosilicate glass SciTech Connect Riley, Brian J.; McCloy, John S.; Goel, Ashutosh; Liezers, Martin; Schweiger, Michael J.; Liu, Juan; Rodriguez, Carmen P.; Kim, Dong-Sang 2013-04-01 This study presents a new method for looking at the solubility of volatile species in simulated low-activity waste glass. The present study looking at rhenium salts is also applicable to real applications involving radioactive technetium salts. In this synthesis method, oxide glass powder is mixed with the volatiles species, vacuum-sealed in a fused quartz ampoule, and then heat-treated under vacuum in a furnace. This technique restricts the volatile species to the headspace above the melt but still within the sealed ampoule, thus maximizing the volatile concentration in contact with the glass. Various techniques were used to measure the solubility of rhenium in glass and include energy dispersive spectroscopy, wavelength dispersive spectroscopy, laser ablation inductively-coupled plasma mass spectroscopy, and inductively-coupled plasma optical emission spectroscopy. The Re-solubility in this glass was determined to be ~3004 parts per million Re atoms. Above this concentration, the salts separated out of the melt as inclusions and as a low viscosity molten salt phase on top of the melt observed during and after cooling. This salt phase was analyzed with X-ray diffraction, scanning electron microscopy as well as some of the other aforementioned techniques and identified to be composed of alkali perrhenate and alkali sulfate. 2. TEM investigation of a ceramic waste form for immobilization of process salts generated during electrometallurgical treatment of spent nuclear fuel. SciTech Connect Esh, D. W.; Frank, S. M.; Goff, K. M.; Johnson, S. G.; Moschetti, T. L.; O'Holleran, T. P.; Sinkler, W. 1999-05-06 Transmission electron microscopy (TEM) examination is presented of the microstructure of a ceramic waste form developed at Argonne National Lab - West for immobilization of actinides and fission products present in an electrorefiner salt. The material is produced by occluding the salt in zeolite granules, followed by hot isostatic pressing of the occluded zeolite in a mixture with a borosilicate glass. The paper presents results from a cold surrogate ceramic waste form, as well as {sup 239}Pu and {sup 238}Pu loaded samples. 3. Dynamics of water clusters in solution with LiCl NASA Astrophysics Data System (ADS) Corsaro, Carmelo; Mallamace, Domenico; Cicero, Nicola; Vasi, Sebastiano; Dugo, Giacomo; Mallamace, Francesco 2016-01-01 In this work we study by means of Nuclear Magnetic Resonance spectroscopy the dynamics of the different water clusters that form within a solution with LiCl at eutectic concentration in the temperature range 320-205 K. This solution is considered a model system allowing the investigation of water properties in the deep supercooled regime in its bulk phase. Our data reveal two important dynamical changes occurring at two relevant temperatures for water: the highest temperature coincides with that of the water density maximum (277 K) and the lowest with that of the so-called dynamical crossover (≃ 225 K). We interpret our data in terms of the different influence that the ions exert on water by lowering the temperature and of the tendency that water displays to develop its characteristic hydrogen bond network. 4. RESULTS OF THE EXTRACTION-SCRUB-STRIP TESTING USING AN IMPROVED SOLVENT FORMULATION AND SALT WASTE PROCESSING FACILITY SIMULATED WASTE SciTech Connect Peters, T.; Washington, A.; Fink, S. 2012-01-09 The Office of Waste Processing, within the Office of Technology Innovation and Development, is funding the development of an enhanced solvent - also known as the next generation solvent (NGS) - for deployment at the Savannah River Site to remove cesium from High Level Waste. The technical effort is a collaborative effort between Oak Ridge National Laboratory (ORNL) and Savannah River National Laboratory (SRNL). As part of the program, the Savannah River National Laboratory (SRNL) has performed a number of Extraction-Scrub-Strip (ESS) tests. These batch contact tests serve as first indicators of the cesium mass transfer solvent performance with actual or simulated waste. The test detailed in this report used simulated Tank 49H material, with the addition of extra potassium. The potassium was added at 1677 mg/L, the maximum projected (i.e., a worst case feed scenario) value for the Salt Waste Processing Facility (SWPF). The results of the test gave favorable results given that the potassium concentration was elevated (1677 mg/L compared to the current 513 mg/L). The cesium distribution value, DCs, for extraction was 57.1. As a comparison, a typical D{sub Cs} in an ESS test, using the baseline solvent formulation and the typical waste feed, is {approx}15. The Modular Caustic Side Solvent Extraction Unit (MCU) uses the Caustic-Side Solvent Extraction (CSSX) process to remove cesium (Cs) from alkaline waste. This process involves the use of an organic extractant, BoBCalixC6, in an organic matrix to selectively remove cesium from the caustic waste. The organic solvent mixture flows counter-current to the caustic aqueous waste stream within centrifugal contactors. After extracting the cesium, the loaded solvent is stripped of cesium by contact with dilute nitric acid and the cesium concentrate is transferred to the Defense Waste Processing Facility (DWPF), while the organic solvent is cleaned and recycled for further use. The Salt Waste Processing Facility (SWPF), under construction, will use the same process chemistry. The Office of Waste Processing (EM-31) expressed an interest in investigating the further optimization of the organic solvent by replacing the BoBCalixC6 extractant with a more efficient extractant. This replacement should yield dividends in improving cesium removal from the caustic waste stream, and in the rate at which the caustic waste can be processed. To that end, EM-31 provided funding for both the Savannah River National Laboratory (SRNL) and the Oak Ridge National Laboratory (ORNL). SRNL wrote a Task Technical Quality and Assurance Plan for this work. As part of the envisioned testing regime, it was decided to perform an ESS test using a simulated waste that simulated a typical envisioned SWPF feed, but with added potassium to make the waste more challenging. Potassium interferes in the cesium removal, and its concentration is limited in the feed to <1950 mg/L. The feed to MCU has typically contained <500 mg/L of potassium. 5. NRC Monitoring of Salt Waste Disposal at the Savannah River Site - 13147 SciTech Connect Pinkston, Karen E.; Ridge, A. Christianne; Alexander, George W.; Barr, Cynthia S.; Devaser, Nishka J.; Felsher, Harry D. 2013-07-01 As part of monitoring required under Section 3116 of the Ronald W. Reagan National Defense Authorization Act for Fiscal Year 2005 (NDAA), the NRC staff reviewed an updated DOE performance assessment (PA) for salt waste disposal at the Saltstone Disposal Facility (SDF). The NRC staff concluded that it has reasonable assurance that waste disposal at the SDF meets the 10 CFR 61 performance objectives for protection of individuals against intrusion (chap.61.42), protection of individuals during operations (chap.61.43), and site stability (chap.61.44). However, based on its evaluation of DOE's results and independent sensitivity analyses conducted with DOE's models, the NRC staff concluded that it did not have reasonable assurance that DOE's disposal activities at the SDF meet the performance objective for protection of the general population from releases of radioactivity (chap.61.41) evaluated at a dose limit of 0.25 mSv/yr (25 mrem/yr) total effective dose equivalent (TEDE). NRC staff also concluded that the potential dose to a member of the public is expected to be limited (i.e., is expected to be similar to or less than the public dose limit in chap.20.1301 of 1 mSv/yr [100 mrem/yr] TEDE) and is expected to occur many years after site closure. The NRC staff used risk insights gained from review of the SDF PA, its experience monitoring DOE disposal actions at the SDF over the last 5 years, as well as independent analysis and modeling to identify factors that are important to assessing whether DOE's disposal actions meet the performance objectives. Many of these factors are similar to factors identified in the NRC staff's 2005 review of salt waste disposal at the SDF. Key areas of interest continue to be waste form and disposal unit degradation, the effectiveness of infiltration and erosion controls, and estimation of the radiological inventory. Based on these factors, NRC is revising its plan for monitoring salt waste disposal at the SDF in coordination with South Carolina Department of Health and Environmental Control (SCDHEC). DOE has completed or begun additional work related to salt waste disposal to address these factors. NRC staff continues to evaluate information related to the performance of the SDF and has been working with DOE and SCDHEC to resolve NRC staff's technical concerns. (authors) 6. A Preliminary Performance Assessment for Salt Disposal of High-Level Nuclear Waste - 12173 SciTech Connect Lee, Joon H.; Clayton, Daniel; Jove-Colon, Carlos; Wang, Yifeng 2012-07-01 A salt repository is one of the four geologic media currently under study by the U.S. DOE Office of Nuclear Energy to support the development of a long-term strategy for geologic disposal of commercial used nuclear fuel (UNF) and high-level radioactive waste (HLW). The immediate goal of the generic salt repository study is to develop the necessary modeling tools to evaluate and improve the understanding of the repository system response and processes relevant to long-term disposal of UNF and HLW in a salt formation. The current phase of this study considers representative geologic settings and features adopted from previous studies for salt repository sites. For the reference scenario, the brine flow rates in the repository and underlying interbeds are very low, and transport of radionuclides in the transport pathways is dominated by diffusion and greatly retarded by sorption on the interbed filling materials. I-129 is the dominant annual dose contributor at the hypothetical accessible environment, but the calculated mean annual dose is negligibly small. For the human intrusion (or disturbed) scenario, the mean mass release rate and mean annual dose histories are very different from those for the reference scenario. Actinides including Pu-239, Pu-242 and Np-237 are major annual dose contributors, and the calculated peak mean annual dose is acceptably low. A performance assessment model for a generic salt repository has been developed incorporating, where applicable, representative geologic settings and features adopted from literature data for salt repository sites. The conceptual model and scenario for radionuclide release and transport from a salt repository were developed utilizing literature data. The salt GDS model was developed in a probabilistic analysis framework. The preliminary performance analysis for demonstration of model capability is for an isothermal condition at the ambient temperature for the near field. The capability demonstration emphasizes key attributes of a salt repository that are potentially important to the long-term safe disposal of UNF and HLW. The analysis presents and discusses the results showing repository responses to different radionuclide release scenarios (undisturbed and human intrusion). For the reference (or nominal or undisturbed) scenario, the brine flow rates in the repository and underlying interbeds are very low, and transport of radionuclides in the transport pathways is dominated by diffusion and greatly retarded by sorption on the interbed filling materials. I-129 (non-sorbing and unlimited solubility with a very long half-life) is the dominant annual dose contributor at the hypothetical accessible environment, but the calculated mean annual dose is negligibly small that there is no meaningful consequence for the repository performance. For the human intrusion (or disturbed) scenario analysis, the mean mass release rate and mean annual dose histories are very different from those for the reference scenario analysis. Compared to the reference scenario, the relative annual dose contributions by soluble, non-sorbing fission products, particularly I-129, are much lower than by actinides including Pu-239, Pu-242 and Np-237. The lower relative mean annual dose contributions by the fission product radionuclides are due to their lower total inventory available for release (i.e., up to five affected waste packages), and the higher mean annual doses by the actinides are the outcome of the direct release of the radionuclides into the overlying aquifer having high water flow rates, thereby resulting in an early arrival of higher concentrations of the radionuclides at the biosphere drinking water well prior to their significant decay. The salt GDS model analysis has also identified the following future recommendations and/or knowledge gaps to improve and enhance the confidence of the future repository performance analysis. - Repository thermal loading by UNF and HLW, and the effect on the engineered barrier and near-field performance. - Closure and consolidation of salt rocks by creep deformation under the influence of thermal perturbation, and the effect on the engineered barrier and near-field performance. - Brine migration and radionuclide transport under the influence of thermal perturbation in generic salt repository environment, and the effect on the engineered barrier and near-field performance and far-field performance. - Near-field geochemistry and radionuclide mobility in generic salt repository environment (high ionic strength brines, elevated temperatures and chemically reducing condition). - Degradation of engineer barrier components (waste package, waste canister, waste forms, etc.) in a generic salt repository environment (high ionic strength brines, elevated temperatures and chemically reducing condition). - Waste stream types and inventory estimates, particularly for reprocessing high-level waste. (authors) 7. Ion Recognition Approach to Volume Reduction of Alkaline Tank Waste by Separation of Sodium Salts SciTech Connect Levitskaia, Tatiana G.; Lumetta, Gregg J.; Moyer, Bruce A.; Bonnesen, Peter V. 2005-06-01 The purpose of this research involving collaboration between Oak Ridge National Laboratory (ORNL) and Pacific Northwest National Laboratory (PNNL) is to explore new approaches to the separation of sodium hydroxide, sodium nitrate, and other sodium salts from high-level alkaline tank waste. The principal potential benefit is a major reduction in disposed waste volume, obviating the building of expensive new waste tanks and reducing the costs of low-activity waste immobilization. Principles of ion recognition are being researched toward discovery of liquid-liquid extraction systems that selectively separate sodium hydroxide and sodium nitrate from other waste components. The successful concept of pseudohydroxide extraction using fluorinated alcohols and phenols is being developed at ORNL and PNNL toward a greater understanding of the controlling equilibria, role of solvation, and of synergistic effects involving crown ethers. Synthesis efforts are being directed toward enhanced sodium binding by crown ethers, both neutral and proton-ionizable. Studies with real tank waste at PNNL will provide feedback toward solvent compositions that have promising properties. 8. Analytical Chemistry and Materials Characterization Results for Debris Recovered from Nitrate Salt Waste Drum S855793 SciTech Connect Martinez, Patrick Thomas; Chamberlin, Rebecca M.; Schwartz, Daniel S.; Worley, Christopher Gordon; Garduno, Katherine; Lujan, Elmer J. W.; Borrego, Andres Patricio; Castro, Alonso; Colletti, Lisa Michelle; Fulwyler, James Brent; Holland, Charlotte S.; Keller, Russell C.; Klundt, Dylan James; Martinez, Alexander; Martin, Frances Louise; Montoya, Dennis Patrick; Myers, Steven Charles; Porterfield, Donivan R.; Schake, Ann Rene; Schappert, Michael Francis; Soderberg, Constance B.; Spencer, Khalil J.; Stanley, Floyd E.; Thomas, Mariam R.; Townsend, Lisa Ellen; Xu, Ning 2015-09-16 Solid debris was recovered from the previously-emptied nitrate salt waste drum S855793. The bulk sample was nondestructively assayed for radionuclides in its as-received condition. Three monoliths were selected for further characterization. Two of the monoliths, designated Specimen 1 and 3, consisted primarily of sodium nitrate and lead nitrate, with smaller amounts of lead nitrate oxalate and lead oxide by powder x-ray diffraction. The third monolith, Specimen 2, had a complex composition; lead carbonate was identified as the predominant component, and smaller amounts of nitrate, nitrite and carbonate salts of lead, magnesium and sodium were also identified. Microfocused x-ray fluorescence (MXRF) mapping showed that lead was ubiquitous throughout the cross-sections of Specimens 1 and 2, while heteroelements such as potassium, calcium, chromium, iron, and nickel were found in localized deposits. MXRF examination and destructive analysis of fragments of Specimen 3 showed elevated concentrations of iron, which were broadly distributed through the sample. With the exception of its high iron content and low carbon content, the chemical composition of Specimen 3 was within the ranges of values previously observed in four other nitrate salt samples recovered from emptied waste drums. 9. Preliminary thermomechanical analyses of a conceptual nuclear waste repository at four salt sites SciTech Connect Wagner, R.A.; Loken, M.C.; Tammemagi, H.Y. 1986-07-01 This report describes preliminary thermal and thermomechanical numerical analyses of a potential nuclear waste repository at four salt sites: Paradox Basin, Utah; Permian Basin, Texas; Richton Dome, Mississippi; and Vacherie Dome, Louisiana. The analysis is subdivided according to three distinct geometric regions: canister, disposal room, and repository. An evaluation of the thermal and thermomechanical responses is performed to assess the suitability of each site to host a nuclear waste repository, under a given set of conceptual design assumptions that have not been optimized. In order to compare the thermal and thermomechanical responses of the rock, the same base case repository conditions were assumed to be used as a ruler to measure salt response relative to each site. Absolute values have not been determined; this report presents a relative comparison of differences that exist between sites due to depth and salt creep characteristics. This report represents the first attempt at calculating relative differences in rock responses and in no way represents values which should be used in a specific design. 44 refs., 21 figs., 11 tabs. 10. THERMAL DESTRUCTION OF HIGHLY CHLORINATED MIXED WASTES WITHOUT GENERATING CORROSIVE OFF-GASES USING MOLTEN SALT OXIDATION (1,2) SciTech Connect Smith, W.; Feizollahi, F. 2002-02-25 A pilot-scale MSO (Molten Salt Oxidation) system was used to process 45-gallons of a halogenated mixed waste that is difficult to treat with other thermal systems. The mixed waste was a halogenated solvent that consisted mostly of methylchloroform. The 80 weight percent of waste consisting of highly corrosive chlorine was captured in the first process vessel as sodium chloride. The sodium chloride leached chrome from that process vessel and the solidified salt exhibited the toxicity characteristic for chrome as measured by TCLP (Toxicity Characteristic Leaching Procedure) testing. The operating ranges for parameters such as salt bed temperature, off-gas temperature, and feed rate that enable sustained operation were identified. At feed rates below the sustainable limit, both processing capacity and maintenance requirements increased with feed rate. Design and operational modifications to increase the sustainable feed rate limit and reduce maintenance requirements reduced both salt carryover and volumetric gas flows. 11. Hydrostatic and shear consolidation tests with permeability measurements on Waste Isolation Pilot Plant crushed salt SciTech Connect Brodsky, N.S. 1994-03-01 Crushed natural rock salt is a primary candidate for use as backfill and barrier material at the Waste Isolation Pilot Plant (WIPP) and therefore Sandia National Laboratories (SNL) has been pursuing a laboratory program designed to quantify its consolidation properties and permeability. Variables that influence consolidation rate that have been examined include stress state and moisture content. The experimental results presented in this report complement existing studies and work in progress conducted by SNL. The experiments described in this report were designed to (1) measure permeabilities of consolidated specimens of crushed salt, (2) determine the influence of brine saturation on consolidation under hydrostatic loads, and 3) measure the effects of small applied shear stresses on consolidation properties. The laboratory effort consisted of 18 individual tests: three permeability tests conducted on specimens that had been consolidated at Sandia, six hydrostatic consolidation and permeability tests conducted on specimens of brine-saturated crushed WIPP salt, and nine shear consolidation and permeability tests performed on crushed WIPP salt specimens containing 3 percent brine by weight. For hydrostatic consolidation tests, pressures ranged from 1.72 MPa to 6.90 MPa. For the shear consolidation tests, confining pressures were between 3.45 MPa and 6.90 MPa and applied axial stress differences were between 0.69 and 4.14 MPa. All tests were run under drained conditions at 25{degrees}C. 12. [A Case of Severe Hyponatremia Caused by Renal Salt Wasting Syndrome in Oropharyngeal Cancer]. PubMed Fujikawa, Taro; Shirakura, Satoru; Hatanaka, Akio; Okano, Wataru; Tokumaru, Takao; Yamada, Masato; Saito, Yoshihiro; Beppu, Takeshi 2015-08-01 Hyponatremia is one of the electrolyte abnormalities frequently encountered in cancer therapy. Cisplatin is a well-known drug which can raise various adverse events, including hyponatremia. A male with advanced oropharyngeal cancer is presented in the present report, who was treated with radiotherapy with concurrent administration of cisplatin and who underwent a total of three episodes of severe hyponatremia in the course of therapy. The first two attacks of hyponatremia following cisplatin administration were accompanied by dehydration and excessive urination, and the patient recovered in one week with rehydration and salt supplementation. Excessive loss of salt in urine confirmed that these events were caused by renal salt wasting syndrome after cisplatin administration. On the other hand, the third attack was due to the syndrome of inappropriate antidiuretic hormone secretion after surgery for a bone fracture. Estimation of the extracellular fluid volume and salt intake/output balance is always believed to be necessary for the diagnosis and proper management of severe hyponatremia after chemotherapy-based treatment with cisplatin. PMID:26548098 13. Tank Waste Transport Stability: Summaries of Hanford Slurry and Salt-Solution Studies in FY 2000 SciTech Connect Welch, T.D. 2002-07-08 This report is a collection of summary articles on FY 2000 studies of slurry transport and salt-well pumping related to Hanford tank waste transfers. These studies are concerned with the stability (steady, uninterrupted flow) of tank waste transfers, a subset of the Department of Energy (DOE) Tanks Focus Area Tank (TFA) Waste Chemistry effort. This work is a collaborative effort of AEA Technology plc, the Diagnostic Instrumentation and Analysis Laboratory at Mississippi State University (DIAL-MSU), the Hemispheric Center for Environmental Technology at Florida International University (HCET-FIU), Numatec Hanford Corporation (NHC), and the Oak Ridge National Laboratory (ORNL). The purpose of this report is to provide, in a single document, an overview of these studies to help the reader identify contacts and resources for obtaining more detailed information and to help promote useful interchanges between researchers and users. Despite over 50 years of experience in transporting radioactive tank wastes to and from equipment and tanks at the Department of Energy's Hanford, Savannah River, and Oak Ridge sites, waste slurry transfer pipelines and process piping become plugged on occasion. At Hanford, several tank farm pipelines are no longer in service because of plugs. At Savannah River, solid deposits in the outlet line of the 2H evaporator have resulted in an unplanned extended downtime. Although waste transfer criteria and guidelines intended to prevent pipeline plugging are in place, they are not always adequate. To avoid pipeline plugging in the future, other factors that are not currently embodied in the transfer criteria may need to be considered. The work summarized here is being conducted to develop a better understanding of the chemical and waste flow dynamics during waste transfer. The goal is to eliminate pipeline plugs by improving analysis and engineering tools in the field that incorporate this understanding. 14. Disposal of NORM-contaminated oil field wastes in salt caverns -- Legality, technical feasibility, economics, and risk SciTech Connect Veil, J.A.; Smith, K.P.; Tomasko, D.; Elcock, D.; Blunt, D.; Williams, G.P. 1998-07-01 Some types of oil and gas production and processing wastes contain naturally occurring radioactive materials (NORM). If NORM is present at concentrations above regulatory levels in oil field waste, the waste requires special disposal practices. The existing disposal options for wastes containing NORM are limited and costly. This paper evaluates the legality, technical feasibility, economics, and human health risk of disposing of NORM-contaminated oil field wastes in salt caverns. Cavern disposal of NORM waste is technically feasible and poses a very low human health risk. From a legal perspective, there are no fatal flaws that would prevent a state regulatory agency from approaching cavern disposal of NORM. On the basis of the costs charged by caverns currently used for disposal of nonhazardous oil field waste (NOW), NORM waste disposal caverns could be cost competitive with existing NORM waste disposal methods when regulatory agencies approve the practice. 15. Salt Processing at the Savannah River Site: Results of Technology Down-Selection and Research and Development to Support New Salt Waste Processing Facility SciTech Connect Lang, K.; Gerdes, K.; Picha, K.; Spader, W.; McCullough, J.; Reynolds, J.; Morin, J. P.; Harmon, H. D. 2002-02-26 The Department of Energy's (DOE) Savannah River Site (SRS) high-level waste (HLW) program is responsible for storage, treatment, and immobilization of HLW for disposal. The Salt Processing Project (SPP) is the salt waste (water-soluble) treatment portion of this effort. The overall SPP encompasses the selection, design, construction, and operation of technologies to prepare the salt-waste feed material for immobilization at the site's Saltstone Production Facility (SPF) and vitrification facility (Defense Waste Processing Facility [DWPF]). Major constituents that must be removed from the salt waste and sent as feed to DWPF include cesium (Cs), strontium (Sr), and actinides. In April 2000, the DOE Deputy Secretary for Project Completion (EM-40) established the SRS Salt Processing Project Technical Working Group (TWG) to manage technology development of treatment alternatives for SRS high-level salt wastes. The separation alternatives investigated included three candidate Cs-removal processes selected, as well as actinide and Sr removal that are also required as a part of each process. The candidate Cs-removal processes are: crystalline Silicotitanate Non-Elutable Ion Exchange (CST); caustic Side Solvent Extraction (CSSX); and small Tank Tetraphenylborate Precipitation (STTP). The Tanks Focus Area was asked to assist DOE by managing the SPP research and development (R&D), revising roadmaps, and developing down-selection criteria. The down-selection decision process focused its analysis on three levels: (a) identification of goals that the selected technology should achieve, (b) selection criteria that are a measure of performance of the goal, and (c) criteria scoring and weighting for each technology alternative. After identifying the goals and criteria, the TWG analyzed R&D results and engineering data and scored the technology alternatives versus the criteria. Based their analysis and scoring, the TWG recommended CSSX as the preferred alternative. This recommendation was formalized in July 2001 when DOE published the Savannah River Site Salt Processing Alternatives Final Supplemental Environmental Impact Statement (SEIS) and was finalized in the DOE Record of Decision issued in October 2001. 16. Ion Recognition Approach to Volume Reduction of Alkaline Tank Waste by Separation of Sodium Salts SciTech Connect Moyer, Bruce A.; Bonnesen, Peter V.; Custelcean, Radu; Delmau, Laetitia H.; Engle, Nancy L.; Kang, Hyun-Ah; Keever, Tamara J.; Marchand, Alan P.; Gadthula, Srinivas; Gore, Vinayak K.; Huang, Zilin; Sivappa, Rasapalli; Tirunahari, Pavan K.; Levitskaia, Tatiana G.; Lumetta, Gregg J. 2005-09-26 The purpose of this research involving collaboration between Oak Ridge National Laboratory (ORNL) and Pacific Northwest National Laboratory (PNNL) is to explore new approaches to the separation of sodium hydroxide, sodium nitrate, and other sodium salts from high-level alkaline tank waste. The principal potential benefit is a major reduction in disposed waste volume, obviating the building of expensive new waste tanks and reducing the costs of vitrification. Principles of ion recognition are being researched toward discovery of liquid-liquid extraction systems that selectively separate sodium hydroxide and sodium nitrate from other waste components. The successful concept of pseudo hydroxide extraction using fluorinated alcohols and phenols is being developed at ORNL and PNNL toward a greater understanding of the controlling equilibria, role of solvation, and of synergistic effects involving crown ethers. Synthesis efforts are being directed toward enhanced sodium binding by crown ethers, both neutral and proton-ionizable. Studies with real tank waste at PNNL will provide feedback toward solvent compositions that have promising properties. 17. DEGRADED TBP SOLVENT REGENERATION TECHNOLOGY USING BUTYLAMINE AS A SOLVENT WASHING TO REDUCE SOLID SALT WASTE SciTech Connect Asakura, T.; Itoh, Y.; Hotoku, S.; Morita, Y.; Uchiyama, G. 2003-02-27 Normal butylamine compounds are studied as salt-free wash reagents for degraded solvent used in PUREX process in spent fuel reprocessing. The solvent wash tests were carried out with two types of butylamine compounds, n-butylamine oxalate and n-butylamine bicarbonate, by counter-current mode using a small size mixer-settler composed of two 4-stage wash steps. Di-n-butyl phosphoric acid (HDBP), the main degradation product from TBP, was removed from real degraded solvent with decontamination factor of 2.5 {approx} 7.9. The study on electrolytic decomposition of butylamine compounds was also conducted for waste treatment. 18. Central Diabetes Insipidus and Cisplatin-Induced Renal Salt Wasting Syndrome: A Challenging Combination. PubMed Cortina, Gerard; Hansford, Jordan R; Duke, Trevor 2016-05-01 We describe a 2-year-old female with a suprasellar primitive neuroectodermal tumor and central diabetes insipidus (DI) who developed polyuria with natriuresis and subsequent hyponatremia 36 hr after cisplatin administration. The marked urinary losses of sodium in combination with a negative sodium balance led to the diagnosis of cisplatin-induced renal salt wasting syndrome (RSWS). The subsequent clinical management is very challenging. Four weeks later she was discharged from ICU without neurological sequela. The combination of cisplatin-induced RSWS with DI can be confusing and needs careful clinical assessment as inaccurate diagnosis and management can result in increased neurological injury. PMID:26928867 19. Review of geochemical measurement techniques for a nuclear waste repository in bedded salt SciTech Connect Knauss, K.G.; Steinborn, T.L. 1980-05-22 A broad, general review is presented of geochemical measurement techniques that can provide data necessary for site selection and repository effectiveness assessment for a radioactive waste repository in bedded salt. The available measurement techniques are organized according to the parameter measured. The list of geochemical parameters include all those measurable geochemical properties of a sample whole values determine the geochemical characteristics or behavior of the system. For each technique, remarks are made pertaining to the operating principles of the measurement instrument and the purpose for which the technique is used. Attention is drawn to areas where further research and development are needed. 20. Independent Assessment of the Savannah River Site High-Level Waste Salt Disposition Alternatives Evaluation SciTech Connect J. T. Case; M. L. Renfro 1998-12-01 This report presents the results of the Independent Project Evaluation (IPE) Team assessment of the Westinghouse Savannah River Company High-Level Waste Salt Disposition Systems Engineering (SE) Team's deliberations, evaluations, and selections. The Westinghouse Savannah River Company concluded in early 1998 that production goals and safety requirements for processing SRS HLW salt to remove Cs-137 could not be met in the existing In-Tank Precipitation Facility as currently configured for precipitation of cesium tetraphenylborate. The SE Team was chartered to evaluate and recommend an alternative(s) for processing the existing HLW salt to remove Cs-137. To replace the In-Tank Precipitation process, the Savannah River Site HLW Salt Disposition SE Team downselected (October 1998) 140 candidate separation technologies to two alternatives: Small-Tank Tetraphenylborate (TPB) Precipitation (primary alternative) and Crystalline Silicotitanate (CST) Nonelutable Ion Exchange (backup alternative). The IPE Team, commissioned by the Department of Energy, concurs that both alternatives are technically feasible and should meet all salt disposition requirements. But the IPE Team judges that the SE Team's qualitative criteria and judgments used in their downselection to a primary and a backup alternative do not clearly discriminate between the two alternatives. To properly choose between Small-Tank TPB and CST Ion Exchange for the primary alternative, the IPE Team suggests the following path forward: Complete all essential R and D activities for both alternatives and formulate an appropriate set of quantitative decision criteria that will be rigorously applied at the end of the R and D activities. Concurrent conceptual design activities should be limited to common elements of the alternatives. 1. Salt tectonics SciTech Connect Talbot, C.J.; Jackson, M.P.A. 1988-01-01 Salt deposits have economic significance because of their importance as oil and gas traps and their potential as radioactive waste disposal sites. This article reviews the formation of salt domes, beginning with a description of the formation of salt deposits as evaporites and a discussion of early attempts to model the development of salt domes. Current work on tectonics of salt dome formation and related tectonics is then discussed in detail. 2. Waste SciTech Connect Gay, R.L.; Barclay, K.M.; Grantham, L.F.; Yosim, S.J. 1981-09-01 A process for converting wastes in molten salts into usable fuels is described. The molten salt acts as a reaction medium and potential acidic pollutants are retained in the melt. The waste is converted to a fuel gas by reacting it with insufficient air for complete conversion to CO/sub 2/ and H/sub 2/O. The product gas is cleared of particles using a baghouse or venturi scrubber and it is then burned in a boiler to produce steam. The results for waste streams containing a high-sulfur oil refinery waste, rubber, wood, leather scraps, and waste x-ray film are presented in this article. 3. A Retrospective Analysis of the Growth Pattern in Patients with Salt-wasting 21-Hydroxylase Deficiency PubMed Central Kawano, Atsuko; Kohno, Hitoshi; Miyako, Kenichi 2014-01-01 Abstract The objective of this study was to investigate the growth pattern of children with the salt-wasting form of congenital adrenal hyperplasia caused by 21-hydroxylase deficiency (21-OHD). We reviewed the medical records of 13 patients in whom salt-wasting 21-OHD was diagnosed during the first 2 mo of life at our hospital from 1980 through 2008. Six reached adult height. Growth patterns, bone age, biochemical data, and the hydrocortisone dose at each growth stage were analyzed retrospectively. The mean adult height was 155.1 6.5 cm (mean SD) in females and 158.1 7.1 cm in males. Although length at birth was normal or longer than the national mean in almost all patients, the mean height SD score of both boys and girls decreased to below 0 SD during infancy. Subsequently, both boys and girls transiently showed growth acceleration and reached their peak growth velocity at 310 yr of age. In conclusion, in addition to suppression of growth during infancy, there was inappropriate growth acceleration during childhood. Especially from 3 mo to 3 yr of age, decreasing the hydrocortisone dose in patients who exhibit slower growth may lead to satisfactory height outcomes. Also, strict adjustment of the hydrocortisone dose to avoid accelerated growth from childhood to adolescence might improve adult height outcomes of patients with 21-OHD. PMID:24790384 4. Partial molar volumes and viscosities of aqueous hippuric acid solutions containing LiCl and MnCl2 4H2O at 303.15 K NASA Astrophysics Data System (ADS) Deosarkar, S. D.; Tawde, P. D.; Zinjade, A. B.; Shaikh, A. I. 2015-09-01 Density (?) and viscosity (?) of aqueous hippuric acid (HA) solutions containing LiCl and MnCl2 4H2O have been studied at 303.15 K in order to understand volumetric and viscometric behavior of these systems. Apparent molar volume (?v) of salts were calculated from density data and fitted to Massons relation and partial molar volumes (?{v/0}) at infinite dilution were determined. Relative viscosity data has been used to determine viscosity A and B coefficients using Jones-Dole relation. Partial molar volume and viscosity coefficients have been discussed in terms of ion-solvent interactions and overall structural fittings in solution. 5. Estimate of the risks of disposing nonhazardous oil field wastes into salt caverns SciTech Connect Tomasko, D.; Elcock, D.; Veil, J. 1997-12-31 Argonne National Laboratory (ANL) has completed an evaluation of the possibility that adverse human health effects (carcinogenic and noncarcinogenic) could result from exposure to contaminants released from nonhazardous oil field wastes (NOW) disposed in domal salt caverns. Potential human health risks associated with hazardous substances (arsenic, benzene, cadmium, and chromium) in NOW were assessed under four postclosure cavern release scenarios: inadvertent cavern intrusion, failure of the cavern seal, failure of the cavern through cracks or leaky interbeds, and a partial collapse of the cavern roof. To estimate potential human health risks for these scenarios, contaminant concentrations at the receptor were calculated using a one-dimensional solution to an advection/dispersion equation that included first order degradation. Assuming a single, generic salt cavern and generic oil-field wastes, the best-estimate excess cancer risks ranged from 1.7 {times} 10{sup {minus}12} to 1.1 {times} 10{sup {minus}8} and hazard indices (referring to noncancer health effects) ranged from 7 {times} 10{sup {minus}9} to 7 {times} 10{sup {minus}4}. Under worse-case conditions in which the probability of cavern failure is 1.0, excess cancer risks ranged from 4.9 {times} 10{sup {minus}9} to 1.7 {times} 10{sup {minus}5} and hazard indices ranged from 7.0 {times} 10{sup {minus}4} to 0.07. Even under worst-case conditions, the risks are within the US Environmental Protection Agency (EPA) target range for acceptable exposure levels. From a human health risk perspective, salt caverns can, therefore, provide an acceptable disposal method for NOW. 6. Expected environments in high-level nuclear waste and spent fuel repositories in salt SciTech Connect Claiborne, H.C.; Rickertsen, L.D., Graham, R.F. 1980-08-01 The purpose of this report is to describe the expected environments associated with high-level waste (HLW) and spent fuel (SF) repositories in salt formations. These environments include the thermal, fluid, pressure, brine chemistry, and radiation fields predicted for the repository conceptual designs. In this study, it is assumed that the repository will be a room and pillar mine in a rock-salt formation, with the disposal horizon located approx. 2000 ft (610 m) below the surface of the earth. Canistered waste packages containing HLW in a solid matrix or SF elements are emplaced in vertical holes in the floor of the rooms. The emplacement holes are backfilled with crushed salt or other material and sealed at some later time. Sensitivity studies are presented to show the effect of changing the areal heat load, the canister heat load, the barrier material and thickness, ventilation of the storage room, and adding a second row to the emplacement configuration. The calculated thermal environment is used as input for brine migration calculations. The vapor and gas pressure will gradually attain the lithostatic pressure in a sealed repository. In the unlikely event that an emplacement hole will become sealed in relatively early years, the vapor space pressure was calculated for three scenarios (i.e., no hole closure - no backfill, no hole closure - backfill, and hole closure - no backfill). It was assumed that the gas in the system consisted of air and water vapor in equilibrium with brine. A computer code (REPRESS) was developed assuming that these changes occur slowly (equilibrium conditions). The brine chemical environment is outlined in terms of brine chemistry, corrosion, and compositions. The nuclear radiation environment emphasized in this report is the stored energy that can be released as a result of radiation damage or crystal dislocations within crystal lattices. 7. Radioactive waste isolation in salt: peer review of Office of Nuclear Waste Isolation's Socioeconomic Program Plan SciTech Connect Winter, R.; Fenster, D.; O'Hare, M.; Zillman, D.; Harrison, W.; Tisue, M. 1984-07-01 The following recommendations have been abstracted from the body of this report. The Office of Nuclear Waste Isolation's Socioeconomic Program Plan for the Establishment of Mined Geologic Repositories to Isolate Nuclear Waste should be modified to: (1) encourage active public participation in the decision-making processes leading to repository site selection; (2) clearly define mechanisms for incorporating the concerns of local residents, state and local governments, and other potentially interested parties into the early stages of the site selection process. In addition, the Office of Nuclear Waste Isolation should carefully review the overall role that these persons and groups, including local pressure groups organized in the face of potential repository development, will play in the siting process; (3) place significantly greater emphasis on using primary socioeconomic data during the site selection process, reversing the current overemphasis on secondary data collection, description of socioeconomic conditions at potential locations, and development of analytical methodologies; (4) include additional approaches to solving socioeconomic problems. For example, a reluctance to acknowledge that solutions to socioeconomic problems need to be found jointly with interested parties is evident in the plan; (5) recognize that mitigation mechanisms other than compensation and incentives may be effective; (6) as soon as potential sites are identified, the US Department of Energy (DOE) should begin discussing impact mitigation agreements with local officials and other interested parties; and (7) comply fully with the pertinent provisions of NWPA. 8. Radioactive waste isolation in salt: Peer review of the Golder Associates draft test plan for in situ testing in an exploratory shaft in salt SciTech Connect Hambley, D.F.; Mraz, D.Z.; Unterberter, R.R.; Stormont, J.C.; Neuman, S.P.; Russell, J.E.; Jacoby, C.H.; Hull, A.B.; Brady, B.H.G.; Ditmars, J.D. 1987-01-01 This report documents the peer review conducted by Argonne National Laboratory of a document entitled ''Draft Test Plan for In Situ Testing in an Exploratory Shaft in Salt,'' prepared for Battelle Memorial Institute's Office of Nuclear Waste Isolation by Golder Associates, Inc. In general, the peer review panelists found the test plan to be technically sound, although some deficiencies were identified. Recommendations for improving the test plan are presented in this review report. A microfiche copy of the following unpublished report is attached to the inside back cover of this report: ''Draft Test Plan for In Situ Testing in an Exploratory Shaft in Salt,'' prepared by Golder Associates, Inc., for Office of Nuclear Waste Isolation, Battelle Memorial Institute, Columbus, Ohio (March 1985). 9. Removal of salt from high-level waste tanks by density-driven circulation or mechanical agitation SciTech Connect Kiser, D.L. 1981-01-01 Twenty-two high-level waste storage tanks at the Savannah River Plant are to be retired in the tank replacement/waste transfer program. The salt-removal portion of this program requires dissolution of about 19 million liters of salt cake. Steam circulation jets were originally proposed to dissolve the salt cake. However, the jets heated the waste tank to 80 to 90/sup 0/C. This high temperature required a long cooldown period before transfer of the supernate by jet, and increased the risk of stress-corrosion cracking in these older tanks. A bench-scale investigation at the Savannah River Laboratory developed two alternatives to steam-jet circulation. One technique was density-driven circulation, which in bench tests dissolved salt at the same rate as a simulated steam circulation jet but at a lower temperature. The other technique was mechanical agitation, which dissolved the salt cake faster and required less fresh water than either density-driven circulation or the simulated steam circulation jet. Tests in an actual waste tank verified bench-scale results and demonstrated the superiority of mechanical agitation. 10. Summary strategy for compliance with postclosure requirements for the waste package for the Salt Repository Project: Final report SciTech Connect Not Available 1988-03-01 This document presents a summary of the strategy of the Salt Repository Project (SRP) to show compliance with the requirements for the waste package after permanent closure of the repository at the site in Deaf Smith County, Texas. The postclosure requirements that govern the performance of the waste package are those in 10 CFR 60.113 for substantially complete containment of the waste and for gradual release of radionuclides after the containment period, and for the postclosure design requirements in 10 CFR 60.135. Also, the waste package plays a role in showing compliance with the total system release requirement in 40 CFR 191.13. 12 refs. 11. A woman with salt-wasting congenital adrenal hyperplasia presenting with a mucinous ovarian cystadenoma during pregnancy. PubMed Kamata, Yuji; Hayashi, Akinori; Ogawa, Akifumi; Ichikawa, Raishi; Moriya, Tatsumi; Shichiri, Masayoshi 2011-01-01 Women with congenital adrenal hyperplasia (CAH) caused by steroid 21-hydroxylase deficiency show reduced fertility, especially with the salt-wasting form. A 27-year-old pregnant woman with this disease underwent laparotomy and oophorectomy to remove a multilocular ovarian tumor at 14 weeks of pregnancy. This proved to be a mucinous cystadenoma. Toward the third trimester, she presented with marked elevations of 17?-hydroxyprogesterone and plasma renin activity. Careful management of endocrine and body fluid homeostasis allowed her to give birth to a healthy female infant with normal external genitalia. This case illustrates endocrinological parameters during pregnancy in a woman with classical salt-wasting CAH. PMID:21921381 12. Dechlorination and stabilization of radioactive chloride salt waste in a molten state SciTech Connect In-Tae Kim; Hwan-Seo Park; Yong-Jun Cho; Hwan-Young Kim; Seong-Won Park; Eung-Ho Kim 2007-07-01 This study suggests a new method to stabilize the molten salt wastes generated from he pyro-processing of a LWR spent fuel. Using a conventional sol-gel process, an inorganic material (SiO{sub 2}-Al{sub 2}O{sub 3}-P{sub 2}O{sub 5}, SAP) reactive to metal chlorides was prepared. In this paper, the reactivity of the SAP on the metal chlorides at 650-850 deg. C, the thermal stability of the reaction products and their leach-resistance under the PCT-A leach test were investigated. In the SAP, three different kinds of chains are available; Si-O-Si (main chain), Si-O-Al (side chain) and Al-O-P/P-O-P (reactive chain). Alkali metal chlorides were converted into metal aluminosilicate (Li{sub x}Al{sub x}Si{sub 1-x}O{sub 2-x}) and metal phosphate(Li{sub 3}PO{sub 4} and Cs{sub 2}AlP{sub 3}O{sub 10}) while the alkaline earth and rare earth chlorides were changed into only metal phosphates (Sr{sub 5}(PO{sub 4}){sub 3}Cl and CePO{sub 4}). The conversion rate was about 96% at a salt waste/SAP weight ratio of 0.5 and a weight loss up to 1100 deg. C measured by the thermo-gravimetric analysis was below 1 Wt%. The leach rates of Cs and Sr under the PCT-A leaching condition were about 10{sup -2} and 10{sup -4} g/m{sup 3}.day, respectively. From these results, it could be concluded that the SAP developed in this study can be considered as an effective stabilizer for metal chlorides and the method of using the SAP could provide a chance to minimize the final waste volume to be disposed off. (authors) 13. Radioactive waste isolation in salt: peer review of Office of Nuclear Waste Isolation's Socioeconomic Program Plan SciTech Connect Winter, R.; Fenster, D.; O'Hare, M.; Zillman, D.; Harrison, W.; Tisue, M. 1984-02-01 The ONWI Socioeconomic Program Plan spells out DOE's approach to analyzing the socioeconomic impacts from siting, constructing, and operating radioactive waste repositories and discusses mitigation strategies. The peer review indicated the following modifications should be made to the Plan: encourage active public participation in the decision-making processes leading to repository site selection; clearly define mechanisms for incorporating the concerns of local residents, state and local governments, and other potentially interested parties into the early stages of the site selection process; place significantly greater emphasis on using primary socioeconomic data during the site selection process, reversing the current overemphasis on secondary data collection, description of socioeconomic conditions at potential locations, and development of analytical methodologies; recognize that mitigation mechanisms other than compensation and incentives may be effective; as soon as potential sites are identified, the US Department of Energy (DOE) should begin discussing impact mitigation agreements with local officials and other interested parties; and comply fully with the pertinent provisions of NWPA. 14. Permeability of natural rock salt from the Waste Isolation Pilot Plant (WIPP) during damage evolution and healing SciTech Connect Pfeifle, T.W. 1998-06-01 The US Department of Energy has developed the Waste Isolation Pilot Plant (WIPP) in the bedded salt of southeastern New Mexico to demonstrate the safe disposal of radioactive transuranic wastes. Four vertical shafts provide access to the underground workings located at a depth of about 660 meters. These shafts connect the underground facility to the surface and potentially provide communication between lithologic units, so they will be sealed to limit both the release of hazardous waste from and fluid flow into the repository. The seal design must consider the potential for fluid flow through a disturbed rock zone (DRZ) that develops in the salt near the shafts. The DRZ, which forms initially during excavation and then evolves with time, is expected to have higher permeability than the native salt. The closure of the shaft openings (i.e., through salt creep) will compress the seals, thereby inducing a compressive back-stress on the DRZ. This back-stress is expected to arrest the evolution of the DRZ, and with time will promote healing of damage. This paper presents laboratory data from tertiary creep and hydrostatic compression tests designed to characterize damage evolution and healing in WIPP salt. Healing is quantified in terms of permanent reduction in permeability, and the data are used to estimate healing times based on considerations of first-order kinetics. 15. Cerebral salt-wasting syndrome after hematopoietic stem cell transplantation in adolescents: 3 case reports PubMed Central Jeon, Yeon Jin; Lee, Hyun Young; Jung, In Ah; Cho, Won-Kyoung; Cho, Bin 2015-01-01 Cerebral salt-wasting syndrome (CSWS) is a rare disease characterized by a extracellular volume depletion and hyponatremia induced by marked natriuresis. It is mainly reported in patients who experience a central nervous system insult, such as cerebral hemorrhage or encephalitis. The syndrome of inappropriate antidiuretic hormone secretion is a main cause of severe hyponatremia after hematopoietic stem cell transplantation, whereas CSWS is rarely reported. We report 3 patients with childhood acute leukemia who developed CSWS with central nervous system complication after hematopoietic stem cell transplantation. The diagnosis of CSW was made on the basis of severe hyponatremia accompanied by increased urine output with clinical signs of dehydration. All patients showed elevated natriuretic peptide and normal antidiuretic hormone. Aggressive water and sodium replacement treatment was instituted in all 3 patients and 2 of them were effectively recovered, the other one was required to add fludrocortisone administration. PMID:26817009 16. Cerebral salt-wasting syndrome after hematopoietic stem cell transplantation in adolescents: 3 case reports. PubMed Jeon, Yeon Jin; Lee, Hyun Young; Jung, In Ah; Cho, Won-Kyoung; Cho, Bin; Suh, Byung-Kyu 2015-12-01 Cerebral salt-wasting syndrome (CSWS) is a rare disease characterized by a extracellular volume depletion and hyponatremia induced by marked natriuresis. It is mainly reported in patients who experience a central nervous system insult, such as cerebral hemorrhage or encephalitis. The syndrome of inappropriate antidiuretic hormone secretion is a main cause of severe hyponatremia after hematopoietic stem cell transplantation, whereas CSWS is rarely reported. We report 3 patients with childhood acute leukemia who developed CSWS with central nervous system complication after hematopoietic stem cell transplantation. The diagnosis of CSW was made on the basis of severe hyponatremia accompanied by increased urine output with clinical signs of dehydration. All patients showed elevated natriuretic peptide and normal antidiuretic hormone. Aggressive water and sodium replacement treatment was instituted in all 3 patients and 2 of them were effectively recovered, the other one was required to add fludrocortisone administration. PMID:26817009 17. Risk analyses for disposing of nonhazardous oil field wastes in salt caverns SciTech Connect Tomasko, D.; Elcock, D.; Veil, J. 1997-09-01 Argonne National Laboratory (ANL) has completed an evaluation of the possibility that adverse human health effects (carcinogenic and noncarcinogenic) could result from exposure to contaminants released from nonhazardous oil field wastes (NOW) disposed of in domal salt caverns. In this assessment, several steps were used to evaluate potential human health risks: identifying potential contaminants of concern; determining how humans could be exposed to these contaminants; assessing the contaminants toxicities; estimating contaminant intakes; and, finally, calculating human cancer and noncancer risks. Potential human health risks associated with hazardous substances (arsenic, benzene, cadmium, and chromium) in NOW were assessed under four postclosure cavern release scenarios: inadvertent cavern intrusion, failure of the cavern seal, failure of the cavern through cracks or leaky interbeds, and a partial collapse of the cavern roof. To estimate potential human health risks for these scenarios, contaminant concentrations at the receptor were calculated using a one-dimensional solution to an advection/dispersion equation that included first order degradation. Even under worst-case conditions, the risks have been found to be within the US EPA target range for acceptable exposure levels. From a human health risk perspective, salt caverns can provide an acceptable disposal method for NOW. 18. Radioactive Waste Isolation in Salt: Peer review of documents dealing with geophysical investigations SciTech Connect McGinnis, L.D.; Bowen, R.H. 1987-03-01 The Salt Repository Project, a US Department of Energy program to develop a mined repository in salt for high-level radioactive waste, is governed by a complex and sometimes inconsistent array of laws, administrative regulations, guidelines, and position papers. In conducting multidisciplinary peer reviews of contractor documents in support of this project, Argonne National Laboratory has needed to inform its expert reviewers of these governmental mandates, with particular emphasis on the relationship between issues and the technical work undertaken. This report acquaints peer review panelists with the regulatory framework as it affects their reviews of site characterization plans and related documents, including surface-based and underground test plans. Panelists will be asked to consider repository performance objectives and issues as they judge the adequacy of proposed geophysical testing. All site-specific discussions relate to the Deaf Smith County site in Texas, which was approved for site characterization by the President in May 1986. Natural processes active at the Deaf Smith County site and the status of geophysical testing near the site are reviewed briefly. 25 refs., 4 figs., 5 tabs. 19. Creep tests on clean and argillaceous salt from the Waste Isolation Pilot Plant SciTech Connect Mellegard, K.D.; Pfeifle, T.W. 1993-05-01 Fifteen triaxial compression creep tests were performed on clean and argillaceous salt from the Waste Isolation Pilot Plant (WIPP). The temperatures in the tests were either 25{degrees}C or 100{degrees}C while the stress difference ranged from 3.5 MPa to 21.0 MPa. In all tests, the confining pressure was 15 MPa. Test duration ranged from 23 to 613 days with an average duration of 300 days. The results of the creep tests supplemented earlier testing and were used to estimate two parameters in the Modified Munson-Dawson constitutive law for the creep behavior of salt. The two parameters determined from each test were the steady-state strain rate and the transient strain limit. These estimates were combined with parameter estimates determined from previous testing to study the dependence of both transient and steady-state creep deformation on stress difference. The exponents on stress difference determined in this study were found to be consistent with revised estimates of the exponents reported by other investigators. 20. Splicing mutation in CYP21 associated with delayed presentation of salt-wasting congenital adrenal hyperplasia SciTech Connect Kohn, B.; Patel, S.V.; Pelczar, J.V. 1995-07-03 Patients with salt-wasting congenital adrenal hyperplasia (SW-CAH) most commonly carry an A-G transition at nucleotide 656 (nt 656 A{r_arrow}G), causing abnormal splicing of exons 2 and 3 in CYP21, the gene encoding active steroid 21-hydroxylase. Affected infants are severely deficient in cortisol and aldosterone, and usually come to medical attention during the neonatal period. We report on 2 affected boys, homozygous for the nt 656 mutation, who thrived in early infancy, but suffered salt-wasting crises unusually late in infancy, at 3.5 and 5.5 months, respectively. Laboratory studies at presentation showed hyponatremia, hyperkalemia, dehydration, and acidosis; serum aldosterone was low in spite of markedly elevated plasma renin activity. Basal 17-hydroxyprogesterone levels were only moderately elevated, yet the stimulated levels were more typical of severe, classic CAH due to 21-hydroxylase deficiency. Genomic DNA from the patients was analyzed. Southern blot showed no major deletions or rearrangements. CYP21-specific amplification by polymerase chain reaction, coupled with allele-specific hybridization using wild-type and mutant probes at each of 9 sites for recognized disease-causing mutations, revealed a single, homozygous mutation in each patient: nt 656 A{r_arrow}G. These results were confirmed by sequence analysis. We conclude that the common nt 656 A{r_arrow}G mutation is sometimes associated with delayed phenotypic expression of SW-CAH. We speculate that variable splicing of the mutant CYP21 may modify the clinical manifestation of this disease. 22 refs., 1 fig., 1 tab. 1. Modeling of the T S D E Heater Test to Investigate Crushed Salt Reconsolidation and Rock Salt Creep for the Underground Disposal of High-Level Nuclear Waste NASA Astrophysics Data System (ADS) Blanco Martin, L.; Rutqvist, J.; Birkholzer, J. T.; Wolters, R.; Lux, K. H. 2014-12-01 Rock salt is a potential medium for the underground disposal of nuclear waste because it has several assets, in particular its water and gas tightness in the undisturbed state, its ability to heal induced fractures and its high thermal conductivity as compared to other shallow-crustal rocks. In addition, the run-of-mine, granular salt, may be used to backfill the mined open spaces. We present simulation results associated with coupled thermal, hydraulic and mechanical processes in the TSDE (Thermal Simulation for Drift Emplacement) experiment, conducted in the Asse salt mine in Germany [1]. During this unique test, conceived to simulate reference repository conditions for spent nuclear fuel, a significant amount of data (temperature, stress changes and displacements, among others) was measured at 20 cross-sections, distributed in two drifts in which a total of six electrical heaters were emplaced. The drifts were subsequently backfilled with crushed salt. This test has been modeled in three-dimensions, using two sequential simulators for flow (mass and heat) and geomechanics, TOUGH-FLAC and FLAC-TOUGH [2]. These simulators have recently been updated to accommodate large strains and time-dependent rheology. The numerical predictions obtained by the two simulators are compared within the framework of an international benchmark exercise, and also with experimental data. Subsequently, a re-calibration of some parameters has been performed. Modeling coupled processes in saliniferous media for nuclear waste disposal is a novel approach, and in this study it has led to the determination of some creep parameters that are very difficult to assess at the laboratory-scale because they require extremely low strain rates. Moreover, the results from the benchmark are very satisfactory and validate the capabilities of the two simulators used to study coupled thermal, mechanical and hydraulic (multi-component, multi-phase) processes relative to the underground disposal of high-level nuclear waste in rock salt. References: [1] Bechthold et al., 1999. BAMBUS-I Project. Euratom, Report EUR19124-EN. [2] Blanco Martín et al., 2014. Comparison of two sequential simulators to investigate thermal-hydraulic-mechanical processes related to nuclear waste isolation in saliniferous formations. In preparation. 2. Design evaluation: Structural calculations for the construction and salt handling shaft and the waste handling shaft at the Waste Isolation Pilot Plant (WIPP) SciTech Connect Torres, T.M. 1988-02-01 One fundamental strategy under consideration for sealing the Waste Isolation Pilot Plant (WIPP) shafts involves the consolidation of the salt backfill in specified portions of the shaft. The criterion for effective salt consolidation was defined to occur when the fractional density of the crushed salt reaches 95%. Thus, the shaft is sealed with a backfill material (crushed salt and/or precompacted salt blocks) that eventually develops properties considered favorable for long-term sealing purposes. This report describes the structural analyses performed to address how much deformation will occur in the shaft as a function of time and depth and how long will it take for the backfill to consolidate to 95% relative density. The two-dimensional axisymmetric shaft model was allowed to creep for a period of 30 years before instantaneous backfill emplacement. Results indicate that the backfill will consolidate to 95% density in less than 300 years at the repository horizon for an initial relative density of 0.60. Most of the backfill in the shaft reaches 95% density in less than 1000 years if the backfill is emplaced at an initial relative density of 0.85. Thus, the calculations show that salt seal materials can reconsolidate in the WIPP shafts to densities nearly equal to those for the intact host rock salt in a relatively short time span. 10 refs., 17 figs. 3. Efficacy of backfilling and other engineered barriers in a radioactive waste repository in salt SciTech Connect Claiborne, H.C. 1982-09-01 In the United States, investigation of potential host geologic formations was expanded in 1975 to include hard rocks. Potential groundwater intrusion is leading to very conservative and expensive waste package designs. Recent studies have concluded that incentives for engineered barriers and 1000-year canisters probably do not exist for reasonable breach scenarios. The assumption that multibarriers will significantly increase the safety margin is also questioned. Use of a bentonite backfill for surrounding a canister of exotic materials was developed in Sweden and is being considered in the US. The expectation that bentonite will remain essentially unchanged for hundreds of years for US repository designs may be unrealistic. In addition, thick bentonite backfills will increase the canister surface temperature and add much more water around the canister. The use of desiccant materials, such as CaO or MgO, for backfilling seems to be a better method of protecting the canister. An argument can also be made for not using backfill material in salt repositories since the 30-cm-thick space will provide for hole closure for many years and will promote heat transfer via natural convection. It is concluded that expensive safety systems are being considered for repository designs that do not necessarily increase the safety margin. It is recommended that the safety systems for waste repositories in different geologic media be addressed individually and that cost-benefit analyses be performed. 4. Evaluation of Sludge Batch 5 Qualification with ISDP Salt Batch 1 Compliance to DWPF Waste Acceptance Criteria SciTech Connect Shafer, A. 2010-05-05 The purpose of this report is to document the acceptability of Sludge Batch 5 with the initial macrobatch operation of the Interim Salt Disposition Project (ISDP) waste to the Defense Waste Processing Facility (DWPF). This report was prepared to comply with the requirements listed in the Waste Acceptance Criteria for Sludge, Actinide Removal Process (ARP), and Modular Caustic Side Solvent Extraction Unit (MCU) Process Transfers to 512-S and DWPF. The requirements for transfers to 512-S were evaluated during ISDP Salt Batch 1 qualification. The calculations of sludge concentrations are based entirely on the Tank 51 sample processed at SRNL. This is conservative because Tank 51 is blended with the dilute feed in the DWPF Feed Tank (Tank 40). This report documents the acceptability of sludge only as well as Sludge Batch 5 sludge slurry combined with ARP/MCU products for feed to DWPF. All criteria were met for unblended Tank 51 material. 5. Conceptual design of retrieval systems for emplaced transuranic waste containers in a salt bed depository. Final report SciTech Connect Fogleman, S.F. 1980-04-01 The US Department of Energy and the Nuclear Regulatory Commission have jurisdiction over the nuclear waste management program. Design studies were previously made of proposed repository site configurations for the receiving, processing, and storage of nuclear wastes. However, these studies did not provide operational designs that were suitable for highly reliable TRU retrieval in the deep geologic salt environment for the required 60-year period. The purpose of this report is to develop a conceptual design of a baseline retrieval system for emplaced transuranic waste containers in a salt bed depository. The conceptual design is to serve as a working model for the analysis of the performance available from the current state-of-the-art equipment and systems. Suggested regulations would be based upon the results of the performance analyses. 6. Development of Spheroidal Inorganic Sorbents for Treatment of Acidic Salt-Bearing Liquid Waste SciTech Connect Collins, J.L. 2001-09-07 A spheroidal composite inorganic sorbent was developed for U.S. Department of Energy-Efficient Separations and Processing Crosscutting Program (USDOE-ESP) for potential use in removing radioactive cesium isotopes from acidic high-salt waste streams such as those at Idaho National Engineering and Environmental Laboratory (INEEL). The sorbent, zirconium monohydrogen phosphate (ZrHP) embedded with fine powder of ammonium molybdophosphate (AMP), was prepared using a unique internal gelation process which can be used to make porous reproducible microspheres that are structurally strong, have a low tendency for surface erosion, and improve the flow dynamics for column operations. Both ZrHP and AMP are excellent sorbent materials and, being inorganic, are stable in high radiation fields. AMP is a very effective sorbent for removing cesium from salt-bearing waste streams for a wide range of acidity. In the pH range of 2 to 10, ZrHP is also a very effective sorbent for removing Cs, Sr, Th, U(VI), Pu(IV), Am(III), Hg, and Pb from streams of lower ionic concentrations. Crucial to developing the spheroidal AMP-ZrHP sorbent was to determine the ideal weight percentage of AMP that could be embedded in the ZrHP microspheres in order to maintain the structural integrity of the microspheres and also achieve a good cesium separation. A total of 12 preparations were made. The dry weight percentage of AMP ranged from 30 to 60. Overall, the best composite microspheres prepared contained 50% AMP (by dry weight measurement). Another composite microsphere, which was composed of titanium monohydrogen phosphate (TiHP) embedded with 18 wt % (air-dried weight) potassium cobalt hexacyanoferrate (KCoCF) and developed for a different separations application, was also batch tested for comparison. It proved to be as effective in removing,the cesium as the air-dried AMP (50 wt %)-ZrHP. Granular KCoCF was also prepared and was very effective. Large samples of each of these materials were sent to INEEL for small-column testing with real waste. 7. SAVANNAH RIVER SITE INCIPIENT SLUDGE MIXING IN RADIOACTIVE LIQUID WASTE STORAGE TANKS DURING SALT SOLUTION BLENDING SciTech Connect Leishear, R.; Poirier, M.; Lee, S.; Steeper, T.; Fowley, M.; Parkinson, K. 2011-01-12 This paper is the second in a series of four publications to document ongoing pilot scale testing and computational fluid dynamics (CFD) modeling of mixing processes in 85 foot diameter, 1.3 million gallon, radioactive liquid waste, storage tanks at Savannah River Site (SRS). Homogeneous blending of salt solutions is required in waste tanks. Settled solids (i.e., sludge) are required to remain undisturbed on the bottom of waste tanks during blending. Suspension of sludge during blending may potentially release radiolytically generated hydrogen trapped in the sludge, which is a safety concern. The first paper (Leishear, et. al. [1]) presented pilot scale blending experiments of miscible fluids to provide initial design requirements for a full scale blending pump. Scaling techniques for an 8 foot diameter pilot scale tank were also justified in that work. This second paper describes the overall reasons to perform tests, and documents pilot scale experiments performed to investigate disturbance of sludge, using non-radioactive sludge simulants. A third paper will document pilot scale CFD modeling for comparison to experimental pilot scale test results for both blending tests and sludge disturbance tests. That paper will also describe full scale CFD results. The final paper will document additional blending test results for stratified layers in salt solutions, scale up techniques, final full scale pump design recommendations, and operational recommendations. Specifically, this paper documents a series of pilot scale tests, where sludge simulant disturbance due to a blending pump or transfer pump are investigated. A principle design requirement for a blending pump is UoD, where Uo is the pump discharge nozzle velocity, and D is the nozzle diameter. Pilot scale test results showed that sludge was undisturbed below UoD = 0.47 ft{sup 2}/s, and that below UoD = 0.58 ft{sup 2}/s minimal sludge disturbance was observed. If sludge is minimally disturbed, hydrogen will not be released. Installation requirements were also determined for a transfer pump which will remove tank contents, and which is also required to not disturb sludge. Testing techniques and test results for both types of pumps are presented. 8. Highlights of the Salt Extraction Process NASA Astrophysics Data System (ADS) Abbasalizadeh, Aida; Seetharaman, Seshadri; Teng, Lidong; Sridhar, Seetharaman; Grinder, Olle; Izumi, Yukari; Barati, Mansoor 2013-11-01 This article presents the salient features of a new process for the recovery of metal values from secondary sources and waste materials such as slag and flue dusts. It is also feasible in extracting metals such as nickel and cobalt from ores that normally are difficult to enrich and process metallurgically. The salt extraction process is based on extraction of the metals from the raw materials by a molten salt bath consisting of NaCl, LiCl, and KCl corresponding to the eutectic composition with AlCl3 as the chlorinating agent. The process is operated in the temperature range 973 K (700C) to 1173 K (900C). The process was shown to be successful in extracting Cr and Fe from electric arc furnace (EAF) slag. Electrolytic copper could be produced from copper concentrate based on chalcopyrite in a single step. Conducting the process in oxygen-free atmosphere, sulfur could be captured in the elemental form. The method proved to be successful in extracting lead from spent cathode ray tubes. In order to prevent the loss of AlCl3 in the vapor form and also chlorine gas emission at the cathode during the electrolysis, liquid aluminum was used. The process was shown to be successful in extracting Nd and Dy from magnetic scrap. The method is a highly promising process route for the recovery of strategic metals. It also has the added advantage of being environmentally friendly. 9. Diagnosis and Treatment of Cerebral Salt Wasting Syndrome With Cryptococcal Meningitis in HIV Patient. PubMed Lee, Sunggeun; Collado, Anitsira; Singla, Montish; Carbajal, Roger; Chaudhari, Ashok; Baumstein, Donald 2016-01-01 Hyponatremia is one of the most common electrolyte imbalances in HIV patients. The differential diagnosis may include hypovolemic hyponatremia, syndrome of inappropriate antidiuretic hormone secretion (SIADH), and adrenal insufficiency. Here, we describe a case of hyponatremia secondary to cerebral salt wasting syndrome (CSWS) in an HIV patient with cryptococcal meningitis. A 52-year-old man with a history of diabetes and HIV was admitted for headache and found to have cryptococcal meningitis. He was also found to have asymptomatic hyponatremia. He had signs of hypovolemia, such as orthostatic hypotension, dry mucosa, decreased skin turgor, hemoconcentration, contraction alkalosis, and high BUN/Cr ratio. The laboratory findings revealed sodium of 125 mmol/L, potassium of 5.5 mmol/L, urine osmolality of 522 mOsm/kg, urine sodium of 162 mmol/L, and urine chloride of 162 mmol/L. We started normal saline for hypovolemia, each 1 L prior and after amphotericin therapy. However, hypovolemia did not improve significantly despite IV fluid. Cosyntropin stimulation test was negative, and renin level was 0.25 ng·mL·h, with the aldosterone level of <1 ng/dL, the serum brain natriuretic peptide of 15 pg/mL, and serum uric acid of 2.8 mg/dL. The diagnosis of CSWS was suspected, fludrocortisone was tried, and hypovolemia and hyponatremia improved. Cryptococcal meningitis in HIV patients can present with CSWS, and the distinction between CSWS and SIADH is important because the treatment for CSWS is different than that of SIADH. Both share a similar clinical picture except that CSWS presents with constant hypovolemia despite volume replacement. Salt tablets, normal saline, or fludrocortisone can be used for treatment. PMID:25569595 10. Accumulation of COGEMA-La Hague-derived reprocessing wastes in French salt marsh sediments. PubMed Cundy, Andrew B; Croudace, Ian W; Warwick, Phillip E; Oh, Jung-Suk; Haslett, Simon K 2002-12-01 Over the past five decades, authorized low-level discharges from coastal nuclear facilities have released significant quantities of artificial radionuclides into the marine environment. In northwest Europe, the majority of the total discharge has derived from nuclear reprocessing activities at Sellafield in the United Kingdom and COGEMA-La Hague in France. At the Sellafield site, a significant amount of the discharges has been trapped in offshore fine sediment deposits, and notably in local coastal and estuarine sediments, and much research has been focused on understanding the distribution, accumulation, and reworking of long-lived radionuclides in these deposits. In contrast, there are few high-resolution published data on the vertical distribution of radionuclides in fine-grained estuarine sediments near, and downstream of, COGEMA-La Hague. This paper therefore examines the vertical distribution of a range of anthropogenic radionuclides in dated salt marsh cores from two estuaries, one adjacent to, and the other downstream of, the COGEMA-La Hague discharge point (the Havre de Carteret at Barneville-Carteret and the Baie de Somme, respectively). The radionuclides examined show a vertical distribution which predominantly reflects variations in input from COGEMA-La Hague (albeit much more clearly at Barneville-Carteret than at the Baie de Somme site), and Pu isotopic ratios are consistent with a La Hague, rather than weapons' fallout, source. Because of sediment mixing, the marshes apparently retain an integrated record of the La Hague discharges, rather than an exact reproduction of the discharge history. Sorption of radionuclides increases in the order 90Sr < 137Cs < 60Co < 239,240Pu, which is consistent with Kd values reported in the literature. In general, the radionuclide activities observed at the sites studied are low (particularly in comparison with salt marsh sediments near the Sellafield facility), but are similar to those found in areas of fine sedimentation in the central Channel. These marshes are not major sinks for discharged reprocessing wastes. PMID:12523411 11. Radioactive waste isolation in salt: special advisory report on the status of the Office of Nuclear Waste Isolation's plans for repository performance assessment SciTech Connect Ditmars, J.D.; Walbridge, E.W.; Rote, D.M.; Harrison, W.; Herzenberg, C.L. 1983-10-01 Repository performance assessment is analysis that identifies events and processes that might affect a repository system for isolation of radioactive waste, examines their effects on barriers to waste migration, and estimates the probabilities of their occurrence and their consequences. In 1983 Battelle Memorial Institute's Office of Nuclear Waste Isolation (ONWI) prepared two plans - one for performance assessment for a waste repository in salt and one for verification and validation of performance assessment technology. At the request of the US Department of Energy's Salt Repository Project Office (SRPO), Argonne National Laboratory reviewed those plans and prepared this report to advise SRPO of specific areas where ONWI's plans for performance assessment might be improved. This report presents a framework for repository performance assessment that clearly identifies the relationships among the disposal problems, the processes underlying the problems, the tools for assessment (computer codes), and the data. In particular, the relationships among important processes and 26 model codes available to ONWI are indicated. A common suggestion for computer code verification and validation is the need for specific and unambiguous documentation of the results of performance assessment activities. A major portion of this report consists of status summaries of 27 model codes indicated as potentially useful by ONWI. The code summaries focus on three main areas: (1) the code's purpose, capabilities, and limitations; (2) status of the elements of documentation and review essential for code verification and validation; and (3) proposed application of the code for performance assessment of salt repository systems. 15 references, 6 figures, 4 tables. 12. Resistance of Coatings for Boiler Components of Waste-to-Energy Plants to Salt Melts Containing Copper Compounds NASA Astrophysics Data System (ADS) Galetz, Mathias Christian; Bauer, Johannes Thomas; Schütze, Michael; Noguchi, Manabu; Cho, Hiromitsu 2013-06-01 The accelerating effect of heavy metal compounds on the corrosive attack of boiler components like superheaters poses a severe problem in modern waste-to-energy plants (WTPs). Coatings are a possible solution to protect cheap, low alloyed steel substrates from heavy metal chloride and sulfate salts, which have a relatively low melting point. These salts dissolve many alloys, and therefore often are the limiting factor as far as the lifetime of superheater tubes is concerned. In this work the corrosion performance under artificial salt deposits of different coatings, manufactured by overlay welding, thermal spraying of self-fluxing as well as conventional systems was investigated. The results of our studies clearly demonstrate the importance of alloying elements such as molybdenum or silicon. Additionally, the coatings have to be dense and of a certain thickness in order to resist the corrosive attack under these severe conditions. 13. Melting of the precipitated ice IV in LiCl aqueous solution and polyamorphism of water. PubMed Mishima, Osamu 2011-12-01 Melting of the precipitated ice IV in supercooled LiCl-H(2)O solution was studied in the range of 0-0.6 MPa and 160-270 K. Emulsified solution was used to detect this metastable transition. Ice IV was precipitated from the aqueous solution of 2.0 mol % LiCl (or 4.8 mol % LiCl) in each emulsion particle at low-temperature and high-pressure conditions, and the emulsion was decompressed at different temperatures. The melting of ice IV was detected from the temperature change of the emulsified sample during the decompression. There was an apparently sudden change in the slope of the ice IV melting curve (liquidus) in the pressure-temperature diagram. At the high-pressure and high-temperature side of the change, the solute-induced freezing point depression was observed. At the low-pressure and low-temperature side, ice IV transformed into ice Ih on the decompression, and the transition was almost unrelated to the concentration of LiCl. These experimental results were roughly explained by the presumed existence of two kinds of liquid water (low-density liquid water and high-density liquid water), or polyamorphism in water, and by the simple assumption that LiCl dissolved maily in high-density liquid water. PMID:21736291 14. Costs for off-site disposal of nonhazardous oil field wastes: Salt caverns versus other disposal methods SciTech Connect Veil, J.A. 1997-09-01 According to an American Petroleum Institute production waste survey reported on by P.G. Wakim in 1987 and 1988, the exploration and production segment of the US oil and gas industry generated more than 360 million barrels (bbl) of drilling wastes, more than 20 billion bbl of produced water, and nearly 12 million bbl of associated wastes in 1985. Current exploration and production activities are believed to be generating comparable quantities of these oil field wastes. Wakim estimates that 28% of drilling wastes, less than 2% of produced water, and 52% of associated wastes are disposed of in off-site commercial facilities. In recent years, interest in disposing of oil field wastes in solution-mined salt caverns has been growing. This report provides information on the availability of commercial disposal companies in oil-and gas-producing states, the treatment and disposal methods they employ, and the amounts they charge. It also compares cavern disposal costs with the costs of other forms of waste disposal. 15. Radioactive waste isolation in salt: Peer review of the Office of Nuclear Waste Isolation's draft report on an issues hierarchy and data needs for site characterization SciTech Connect Harrison, W.; Fenster, D.F.; Ditmars, J.D.; Paddock, R.A.; Rote, D.M.; Hambley, D.F.; Seitz, M.G.; Hull, A.B. 1986-12-01 At the request of the Salt Repository Project (SRPO), Argonne National Laboratory conducted an independent peer review of a report by the Battelle Office of Nuclear Waste Isolation entitled ''Salt Repository Project Issues Hierarchy and Data Needs for Site Characterization (Draft).'' This report provided a logical structure for evaluating the outstanding questions (issues) related to selection and licensing of a site as a high-level waste repository. It also provided a first estimate of the information and data necessary to answer or resolve those questions. As such, this report is the first step in developing a strategy for site characterization. Microfiche copies of ''Draft Issues Hierarchy, Resolution Strategy, and Information Needs for Site Characterization and Environmental/Socioeconomic Evaluation - July, 1986'' and ''Issues Hierarchy and Data Needs for Site Characterization - February, 1985'' are included in the back pocket of this report. 16. Syndrome of Inappropriate Antidiuretic Hormone Secretion and Cerebral/Renal Salt Wasting Syndrome: Similarities and Differences PubMed Central Oh, Ji Young; Shin, Jae Il 2015-01-01 Hyponatremia (sodium levels of <135?mEq/L) is one of the most common electrolyte imbalances in clinical practice, especially in patients with neurologic diseases. Hyponatremia can cause cerebral edema and brain herniation; therefore, prompt diagnosis and proper treatment is important in preventing morbidity and mortality. Among various causes of hyponatremia, diagnosing syndrome of inappropriate antidiuretic hormone secretion (SIADH) and cerebral/renal salt wasting syndrome (C/RSW) is difficult due to many similarities. SIADH is caused by excess of renal water reabsorption through inappropriate antidiuretic hormone secretion, and fluid restriction is the treatment of choice. On the other hand, C/RSW is caused by natriuresis, which is followed by volume depletion and negative sodium balance and replacement of water and sodium is the mainstay of treatment. Determinating volume status in hyponatremic patients is the key point in differential between SIADH and C/RSW. However, in most situations, differential diagnosis of these two diseases is difficult because they overlap in many clinical and laboratory aspects, especially to assess differences in volume status of these patients. Although distinction between the SIADH and C/RSW is difficult, improvement of hypouricemia and an increased fractional excretion of uric acid after the correction of hyponatremia in SIADH, not in C/RSW, may be one of the helpful points in discriminating the two diseases. In this review, we compare these two diseases regarding the pathophysiologic mechanisms, diagnosis, and therapeutic point of view. PMID:25657991 17. Cerebral salt wasting after traumatic brain injury: a review of the literature. PubMed Leonard, Jan; Garrett, Raymond E; Salottolo, Kristin; Slone, Denetta S; Mains, Charles W; Carrick, Matthew M; Bar-Or, David 2015-01-01 Electrolyte imbalances are common among patients with traumatic brain injury (TBI). Cerebral salt wasting (CSW) is an electrolyte imbalance characterized by hyponatremia and hypovolemia. Differentiating the syndrome of inappropriate antidiuretic hormone and CSW remains difficult and the pathophysiological mechanisms underlying CSW are unclear. Our intent was to review the literature on CSW within the TBI population, in order to report the incidence and timing of CSW after TBI, examine outcomes, and summarize the biochemical changes in patients who developed CSW. We searched MEDLINE through 2014, hand-reviewed citations, and searched abstracts from the American Association for the Surgery of Trauma (2003-2014). Publications were included if they were conducted within a TBI population, presented original data, and diagnosed CSW. Publications were excluded if they were review articles, discussed hyponatremia but did not differentiate the etiology causing hyponatremia, or presented cases with chronic disease. Fifteen of the 47 publications reviewed met the selection criteria; nine (60%) were case reports, five (33%) were prospective and 1 (7%) was a retrospective study. Incidence of CSW varied between 0.8 - 34.6%. The populations studied were heterogeneous and the criteria used to define hyponatremia and CSW varied. Though believed to play a role in the development of CSW, increased levels of natriuretic peptides in patients diagnosed with CSW were not consistently reported. These findings reinforce the elusiveness of the CSW diagnosis and the need for strict and consistent diagnostic criteria. PMID:26561391 18. Radioactive waste isolation in salt: peer review of Westinghouse Electric Corporation's report on reference conceptual designs for a repository waste package SciTech Connect Rote, D.M.; Hull, A.B.; Was, G.S.; Macdonald, D.D.; Wilde, B.E.; Russell, J.E.; Kruger, J.; Harrison, W.; Hambley, D.F. 1985-10-01 This report documents the findings of the peer panel constituted by Argonne National Laboratory to review Region A of Westinghouse Electric Corporation's report entitled Waste Package Reference Conceptual Designs for a Repository in Salt. The panel determined that the reviewed report does not provide reasonable assurance that US Nuclear Regulatory Commission (NRC) requirements for waste packages will be met by the proposed design. It also found that it is premature to call the design a ''reference design,'' or even a ''reference conceptual design.'' This review report provides guidance for the preparation of a more acceptable design document. 19. Mitigation of Salinity Buildup and Recovery of Wasted Salts in a Hybrid Osmotic Membrane Bioreactor-Electrodialysis System. PubMed Lu, Yaobin; He, Zhen 2015-09-01 The osmotic membrane bioreactor (OMBR) is an emerging technology that uses water osmosis to accomplish separation of biomass from the treated effluent; however, accumulation of salts in the wastewater due to water flux and loss of draw solute because of reverse salt flux seriously hinder OMBR development. In this study, a hybrid OMBR-electrodialysis (ED) system was proposed and investigated to alleviate the salinity buildup. The use of an ED (3 V applied) could maintain a relatively low conductivity of 8 mS cm(-1) in the feed solution, which allowed the OMBR to operate for 24 days, about 6 times longer than a conventional OMBR without a functional ED. It was found that the higher the voltage applied to the ED, the smaller area of ion-exchange membrane was needed for salt separation. The salts recovered by the ED were successfully reused as a draw solute in the OMBR. At an energy consumption of 1.88-4.01 kWh m(-3), the hybrid OMBR-ED system could achieve a stable water flux of about 6.23 L m(-2) h(-1) and an efficient waste salt recovery of 1.26 kg m(-3). The hybrid OMBR-ED system could be potentially more advantageous in terms of less waste saline water discharge and salt recovery compared with a combined OMBR and reverse osmosis system. It also offers potential advantages over the conventional OMBR+post ED treatment in higher water flux and less wastewater discharge. PMID:26238212 20. An optimal method for phosphorylation of rare earth chlorides in LiCl-KCl eutectic based waste salt NASA Astrophysics Data System (ADS) Eun, H. C.; Kim, J. H.; Cho, Y. Z.; Choi, J. H.; Lee, T. K.; Park, H. S.; Park, G. I. 2013-11-01 A study on an optimal method for the phosphorylation of rare earth chlorides in LiCl-KCl eutectic waste salt generated the pyrochemical process of spent nuclear fuel was performed. A reactor with a pitched four blade impeller was designed to create a homogeneous mixing zone in LiCl-KCl eutectic salt. A phosphorylation test of NdCl3 in the salt was carried out by changing the operation conditions (operation temperature, stirring rate, agent injection amount). Based on the results of the test, a proper operation condition (450 C, 300 rpm, 1 eq. of phosphorylation agent) for over a 0.99 conversion ratio of NdCl3 to NdPO4 was determined. Under this condition, multi-component rare earth (La, Ce, Pr, Nd, Sm, Eu, Gd, Y) chlorides were effectively converted into phosphate forms. It was confirmed that the existing regeneration process of LiCl-KCl eutectic waste salt can be greatly improved and simplified through these phosphorylation test results. 1. Quantum Theoretical Study of KCl and LiCl Clusters NASA Astrophysics Data System (ADS) Koetter, Ted; Hira, Ajit; Salazar, Justin; Jaramillo, Danelle 2014-03-01 This research focuses on the theoretical study of molecular clusters to examine the chemical properties of small KnClnandLinCln clusters (n = 2 - 20). The potentially important role of these molecular species in biochemical and medicinal processes is well known. This work applies the hybrid ab initio methods of quantum chemistry to derive the different alkali-halide (MnHn) geometries. Of particular interest is the competition between hexagonal ring geometries and rock salt structures. Electronic energies, rotational constants, dipole moments, and vibrational frequencies for these geometries are calculated. Magic numbers for cluster stability are identified and are related to the property of cluster compactness. Mapping of the singlet, triplet, and quintet, potential energy surfaces is performed. Calculations were performed to examine the interactions of these clusters with some atoms and molecules of biological interest, including O, O2, and Fe. Potential design of new medicinal drugs is explored. 2. Autosomal Dominant Pseudohypoaldosteronism Type 1 in an Infant with Salt Wasting Crisis Associated with Urinary Tract Infection and Obstructive Uropathy PubMed Central Bowden, Sasigarn A.; Cozzi, Corin; Hickey, Scott E.; Thrush, Devon Lamb; Astbury, Caroline; Nuthakki, Sushma 2013-01-01 Type 1 pseudohypoaldosteronism (PHA1) is a salt wasting syndrome caused by renal resistance to aldosterone. Primary renal PHA1 or autosomal dominant PHA1 is caused by mutations in mineralocorticoids receptor gene (NR3C2), while secondary PHA1 is frequently associated with urinary tract infection (UTI) and/or urinary tract malformations (UTM). We report a 14-day-old male infant presenting with severe hyperkalemia, hyponatremic dehydration, metabolic acidosis, and markedly elevated serum aldosterone level, initially thought to have secondary PHA1 due to the associated UTI and posterior urethral valves. His serum aldosterone remained elevated at 5 months of age, despite resolution of salt wasting symptoms. Chromosomal microarray analysis revealed a deletion of exons 35 in NR3C2 in the patient and his asymptomatic mother who also had elevated serum aldosterone level, confirming that he had primary or autosomal dominant PHA1. Our case raises the possibility that some patients with secondary PHA1 attributed to UTI and/or UTM may instead have primary autosomal dominant PHA1, for which genetic testing should be considered to identify the cause, determine future recurrence risk, and possibly prevent the life-threatening salt wasting in a subsequent family member. Future clinical research is needed to investigate the potential overlapping between secondary PHA1 and primary autosomal dominant PHA1. PMID:24455331 3. Alternative methods to manage waste salt from repository excavation in the Deaf Smith County and Swisher County locations, Texas: A scoping study: Technical report. [Salt and salt-laden material SciTech Connect Not Available 1987-01-01 This report describes and qualitatively evaluates eight options for managing the large volumes of salt and salt-laden rock that would result from the excavation of a high-level radioactive waste repository in Deaf Smith County or Swisher County, Texas. The options are: distribution for commercial use; ocean disposal; deep-well injection; disposal in multilevel mines on the site; disposal in abandoned salt mines off the site; disposal off the site in abandoned mines developed for minerals other than salt; disposal in excavated landfills; and surface disposal on alkali flats. The main features of each option are described, as well as the associated environmental and economic impacts, and regulatory constraints. The options are evaluated in terms of 11 factors that jointly constitute a test of relative suitability. The results of the evaluation and implications for further study are indicated. This document does not consider or include the actual numbers, findings, or conclusions contained in the final Deaf Smith County Environmental Assessment (DOE, 1986). 43 refs., 8 tabs. 4. Waste Stream Generated and Waste Disposal Plans for Molten Salt Reactor Experiment at Oak Ridge National Laboratory SciTech Connect Haghighi, M. H.; Szozda, R. M.; Jugan, M. R. 2002-02-26 The Molten Salt Reactor Experiment (MSRE) site is located in Tennessee, on the U.S. Department of Energy (DOE) Oak Ridge Reservation (ORR), south of the Oak Ridge National Laboratory (ORNL) main plant across Haw Ridge in Melton Valley. The MSRE was run by ORNL to demonstrate the desirable features of the molten-salt concept in a practical reactor that could be operated safely and reliably. It introduced the idea of a homogeneous reactor using fuel salt media and graphite moderation for power and breeder reactors. The MSRE reactor and associated components are located in cells beneath the floor in the high-bay area of Building 7503 (Figure 1). The reactor was operated from June 1965 to December 1969. When the reactor was shut down, fuel salt was drained from the reactor circuit to two drain tanks. A ''clean'' salt was then circulated through the reactor as a decontamination measure and drained to a third drain tank. When operations ceased, the fuel and flush salts were allowed t o cool and solidify in the drain tanks. At shutdown, the MSRE facility complex was placed in a surveillance and maintenance program. As a result of the S&M program, it was discovered in 1994 that gaseous uranium (233U/232U) hexafluoride (UF6) had moved throughout the MSRE process systems. The UF6 was generated when radiolysis of the fluorine salts caused the individual constituents to dissociate to their component atoms, including free fluorine.Some of the free fluorine combined with uranium fluorides (UF4) in the salt to form UF6. UF6 is gaseous at slightly above ambient temperatures; thus, periodic heating of the fuel salts (which was intended to remedy the radiolysis problems) and simple diffusion had allowed the UF6 to move out of the salt and into the process systems of MSRE. 5. Combined Utilization of Cation Exchanger and Neutral Receptor to Volume Reduction of Alkaline Tank Waste by Separation of Sodium Salts SciTech Connect Levitskaia, Tatiana G.; Lumetta, Gregg J.; Moyer, Bruce A. 2004-03-29 In this report, novel approaches to the selective liquid-liquid extraction separation of sodium hydroxide and sodium nitrate from high-level alkaline tank waste will be discussed. Sodium hydroxide can be successfully separated from alkaline tank-waste supernatants by weakly acidic lipophilic hydroxy compounds via a cation-exchange mechanism referred to as pseudo hydroxide extraction. In a multi-cycle process, as sodium hydroxide in the aqueous phase becomes depleted, it is helpful to have a neutral sodium receptor in the extraction system to exploit the high nitrate concentration in the waste solution to promote sodium removal by an ion-pair extraction process. Simultaneous utilization of an ionizable organic hydroxy compound and a neutral extractant (crown ether) in an organic phase results in the synergistic enhancement of ion exchange and improved separation selectivity due to the receptor's strong and selective sodium binding. Moreover, combination of the hydroxy compound and the crown ether provides for mutually increased solubility, even in a non-polar organic solvent. Accordingly, application of Isopar{reg_sign} L, a kerosene-like alkane solvent, becomes feasible. This investigation involves examination of such dual-mechanism extraction phases for sodium extraction from simulated and actual salt cake waste solutions. Sodium salts can be regenerated upon the contact of the loaded extraction phases with water. Finally, conditions of potential extraction/strip cycling will be discussed. 6. Renal salt wasting and chronic dehydration in claudin-7-deficient mice. PubMed Tatum, Rodney; Zhang, Yuguo; Salleng, Kenneth; Lu, Zhe; Lin, Jen-Jar; Lu, Qun; Jeansonne, Beverly G; Ding, Lei; Chen, Yan-Hua 2010-01-01 Claudin-7, a member of the claudin family, is highly expressed in distal nephrons of kidneys and has been reported to be involved in the regulation of paracellular Cl(-) permeability in cell cultures. To investigate the role of claudin-7 in vivo, we generated claudin-7 knockout mice (Cln7(-/-)) by the gene-targeting deletion method. Here we report that Cln7(-/-) mice were born viable, but died within 12 days after birth. Cln7(-/-) mice showed severe salt wasting, chronic dehydration, and growth retardation. We found that urine Na(+), Cl(-), and K(+) were significantly increased in Cln7(-/-) mice compared with that of Cln7(+/+) mice. Blood urea nitrogen and hematocrit were also significantly higher in Cln7(-/-) mice. The wrinkled skin was evident when Cln7(-/-) mice were approximately 1 wk old, indicating that they suffered from chronic fluid loss. Transepidermal water loss measurements showed no difference between Cln7(+/+) and Cln7(-/-) skin, suggesting that there was no transepidermal water barrier defect in Cln7(-/-) mice. Claudin-7 deletion resulted in the dramatic increase of aldosterone synthase mRNA level as early as 2 days after birth. The significant increases of epithelial Na(+) channel alpha, Na(+)-Cl(-) cotransporter, and aquaporin 2 mRNA levels revealed a compensatory response to the loss of electrolytes and fluid in Cln7(-/-) mice. Na(+)-K(+)-ATPase alpha(1) expression level was also greatly increased in distal convoluted tubules and collecting ducts where claudin-7 is normally expressed. Our study demonstrates that claudin-7 is essential for NaCl homeostasis in distal nephrons, and the paracellular ion transport pathway plays indispensable roles in keeping ionic balance in kidneys. PMID:19759267 7. Stabilization of 238Pu-contaminated combustible waste by molten salt oxidation NASA Astrophysics Data System (ADS) Stimmel, Jay J.; Remerowski, Mary Lynn; Ramsey, Kevin B.; Heslop, J. Mark 2000-07-01 Surrogate studies were conducted using the molten salt oxidation system at the Naval Surface Warfare Center-Indian Head Division. This system uses a rotary feed system and an alumina molten salt oxidation vessel. The combustible materials were tested individually and together in a homogenized mixture. A slurry containing pyrolyzed cheesecloth ash spiked with cerium oxide, which is used as a surrogate for plutonium, and ethylene glycol were also treated in the molten salt oxidation vessel. 8. Biochemical solubilization of toxic salts from residual geothermal brines and waste waters DOEpatents Premuzic, E.T.; Lin, M.S. 1994-11-22 A method of solubilizing metal salts such as metal sulfides in a geothermal sludge using mutant Thiobacilli selected for their ability to metabolize metal salts at high temperature is disclosed. The method includes the introduction of mutated Thiobacillus ferrooxidans and Thiobacillus thiooxidans to a geothermal sludge or brine. The microorganisms catalyze the solubilization of metal salts. For instance, in the case of metal sulfides, the microorganisms catalyze the solubilization to form soluble metal sulfates. 54 figs. 9. Biochemical solubilization of toxic salts from residual geothermal brines and waste waters DOEpatents Premuzic, Eugene T.; Lin, Mow S. 1994-11-22 A method of solubilizing metal salts such as metal sulfides in a geothermal sludge using mutant Thiobacilli selected for their ability to metabolize metal salts at high temperature is disclosed, The method includes the introduction of mutated Thiobacillus ferrooxidans and Thiobacillus thiooxidans to a geothermal sludge or brine. The microorganisms catalyze the solubilization of metal salts, For instance, in the case of metal sulfides, the microorganisms catalyze the solubilization to form soluble metal sulfates. 10. Evaluation of the role of threshold pressure in controlling flow of waste-generated gas into bedded salt at the Waste Isolation Pilot Plant SciTech Connect Davies, P.B. 1991-06-01 Anoxic corrosion and microbial degradation of contact-handled transuranic waste may produce sufficient quantities of gas over a long time period to generate high pressure in the disposal rooms at the Waste Isolation Pilot Plant (WIPP) repository. Dissipation of pressure by outward gas flow will be inhibited by the low permeability of the surrounding rock and by capillary forces that resist gas penetration into this water-saturated rock. Threshold pressure is the gas pressure required to overcome capillary resistance to initial gas penetration and to the development of interconnected gas pathways that would outward gas flow. The primary objectives of this study are to estimate the magnitude of threshold pressure in the bedded salt that surrounds the WIPP repository and to evaluate the role this parameter plays in controlling the outward flow of waste-generated gas. 54 refs., 9 figs., 4 tabs. 11. Emissions from energetic material waste during the Molten Salt Destruction process SciTech Connect Watkins, B.E.; Upadhye, R.S.; Pruneda, C.O.; Brummond, W.A. 1994-07-05 The Molten Salt Destruction (MSD) process is an alternative to open burn/open detonation for destroying energetic materials; MSD has inherently low gaseous emissions, and the salt bath can scrub both acidic gases and particulates. It was demonstrated that high explosives and a liquid propellant can be safely and completely destroyed using MSD. Gaseous emissions of NOx and CO are very low. Nitrate builds up in the salt bath when nitrate-rich materials are destroyed, but addition fuel reduces the nitrate to NO. A program has been begun to add catalytic materials to the bed to further reduce emissions; a small molten salt bath has been constructed for chemical kinetic studies. 12. Results of screening activities in salt states prior to the enactment of the Nationall Waste Policy Act SciTech Connect Carbiener, W.A. 1983-01-01 The identification of potential sites for a nuclear waste repository through screening procedures in the salt states is a well-established, deliberate process. This screening process has made it possible to carry out detailed studies of many of the most promising potential sites, and general studies of all the sites, in anticipation of the siting guidelines specified in the Nuclear Waste Policy Act. The screening work completed prior to the passage of the Act allowed the Secretary of Energy to identify seven salt sites as potentially acceptable under the provisions of Section 116(a) of the Act. These sites were formally identified by letters from Secretary Hodel to the states of Texas, Utah, Mississippi, and Louisiana on February 2, 1983. The potentially acceptable salt sites were in Deaf Smith and Swisher Counties in Texas; Davis and Lavender Canyons in the Gibson Dome location in Utah; Richton and Cypress Creek Domes in Mississippi; and Vacherie Dome in Louisiana. Further screening will include comparison of each potentially acceptable site against disqualification factors and selection of a preferred site in each of the three geohydrologic settings from those remaining, in accordance with the siting guidelines. These steps will be documented in statutory Environmental Assessments prepared for each site to be nominated for detailed characterization. 9 references. 13. Preservation of artifacts in salt mines as a natural analog for the storage of transuranic wastes at the WIPP repository SciTech Connect Martell, M.A.; Hansen, F.; Weiner, R. 1998-10-01 Use of natures laboratory for scientific analysis of complex systems is a largely untapped resource for understanding long-term disposal of hazardous materials. The Waste Isolation Pilot Plant (WIPP) in the US is a facility designed and approved for storage of transuranic waste in a salt medium. Isolation from the biosphere must be ensured for 10,000 years. Natural analogs provide a means to interpret the evolution of the underground disposal setting. Investigations of ancient sites where manmade materials have experienced mechanical and chemical processes over millennia provide scientific information unattainable by conventional laboratory methods. This paper presents examples of these pertinent natural analogs, provides examples of features relating to the WIPP application, and identifies potential avenues of future investigations. This paper cites examples of analogical information pertaining to the Hallstatt salt mine in Austria and Wieliczka salt mine in Poland. This paper intends to develop an appreciation for the applicability of natural analogs to the science and engineering of a long-term disposal facility in geomedia. 14. Electrochemistry and Spectroelectrochemistry of Europium(III) chloride in 3 LiCl – 2KCl from 643 to 1123 K SciTech Connect Schroll, Cynthia A.; Chatterjee, Sayandev; Levitskaia, Tatiana G.; Heineman, William R.; Bryan, Samuel A. 2013-09-09 The electrochemical and spectroelectrochemical behavior of Europium(III) chloride in a molten salt eutectic, 3 LiCl – 2 KCl, over a temperature range of 643 – 1123 K using differential pulse voltammetry, cyclic voltammetry, potential step chronoabsorptometry, and thin-layer spectroelectrochemistry is reported. The electrochemical reaction was determined to be the one electron reduction of Eu3+ to Eu2+ at all temperatures. The redox potential of Eu3+/2+ shifts to more positive potentials and the diffusion coefficient for Eu3+ increases as temperature increases. The results for the number of electrons transferred, redox potential and diffusion coefficient are in good agreement between the electrochemical and spectroelectrochemical techniques. 15. Use of zinc and copper (I) salts to reduce sulfur and nitrogen impurities during the pyrolysis of plastic and rubber waste to hydrocarbons DOEpatents Wingfield, Jr., Robert C.; Braslaw, Jacob; Gealer, Roy L. 1984-01-01 An improvement in a process for the pyrolytic conversion of rubber and plastic waste to hydrocarbon products which results in reduced levels of nitrogen and sulfur impurities in these products. The improvement comprises pyrolyzing the waste in the presence of at least about 1 weight percent of salts, based on the weight of the waste, preferably chloride or carbonate salts, of zinc or copper (I). This invention was made under contract with or subcontract thereunder of the Department of Energy Contract #DE-AC02-78-ER10049. 16. Zr electrorefining process for the treatment of cladding hull waste in LiCl-KCl molten salts SciTech Connect Lee, Chang Hwa; Lee, You Lee; Jeon, Min Ku; Kang, Kweon Ho; Choi, Yong Taek; Park, Geun Il 2013-07-01 Zr electrorefining for the treatment of Zircaloy-4 cladding hull waste is demonstrated in LiCl-KCl-ZrCl{sub 4} molten salts. Although a Zr oxide layer thicker than 5 ?m strongly inhibits the Zr dissolution process, pre-treatment processes increases the dissolution kinetics. For 10 g-scale experiments, the purities of the recovered Zr were 99.54 wt.% and 99.74 wt.% for fresh and oxidized cladding tubes, respectively, with no electrical contact issue. The optimal condition for Zr electrorefining has been found to improve the morphological feature of the recovered Zr, which reduces the salt incorporation by examining the effect of the process parameters such as the ZrCl{sub 4} concentration and the applied potential. 17. Radioactive waste isolation in salt: Peer review of the Fluor Technology, Inc. , report and position paper concerning waste emplacement mode and its effect on repository conceptual design SciTech Connect Hambley, D.F.; Russell, J.E.; Whitfield, R.G.; McGinnis, L.D.; Harrison, W.; Jacoby, C.H.; Bump, T.R.; Mraz, D.Z.; Busch, J.S.; Fischer, L.E. 1987-02-01 Recommendations for revising the Fluor Technology, Inc., draft position paper entitled Evaluation of Waste Emplacement Mode and the final report entitled Waste Package/Repository Impact Study include: reevaluate the relative rankings for the various emplacement modes; delete the following want objectives: maximize ability to locate the package horizon because sufficient flexibility exists to locate rooms in the relatively clean San Andres Unit 4 Salt and maximize far-field geologic integrity during retrieval because by definition the far field will be unaffected by thermal and stress perturbations caused by remining; give greater emphasis to want objectives regarding cost and use of present technology; delete the following statements from pages 1-1 and 1-2 of the draft position paper: ''No thought or study was given to the impacts of this configuration (vertical emplacement) on repository construction or short and long-term performance of the site'' and ''Subsequent salt repository designs adopted the vertical emplacement configuration as the accepted method without further evaluation.''; delete App. E and lines 8-17 of page 1-4 of the draft position paper because they are inappropriate; adopt a formal decision-analysis procedure for the 17 identified emplacement modes; revise App. F of the impact study to more accurately reflect current technology; consider designing the underground layout to take advantage of stress-relief techniques; consider eliminating reference to fuel assemblies <10 yr ''out-of-reactor''; model the temperature distribution, assuming that the repository is constructed in an infinitely large salt body; state that the results of creep analyses must be considered tentative until they can be validated by in situ measurements; and reevaluate the peak radial stresses on the waste package so that the calculated stress conditions more closely approximate expected in situ conditions. 18. Regulatory Framework for Salt Waste Disposal and Tank Closure at the Savannah River Site - 13663 SciTech Connect Thomas, Steve; Dickert, Ginger 2013-07-01 The end of the Cold War has left a legacy of approximately 37 million gallons of radioactive waste in the aging waste tanks at the Department of Energy's Savannah River Site (SRS). A robust program is in place to remove waste from these tanks, treat the waste to separate into a relatively small volume of high-level waste and a large volume of low-level waste, and to actively dispose of the low-level waste on-site and close the waste tanks and associated ancillary structures. To support performance-based, risk-informed decision making and to ensure compliance with all regulatory requirements, the U.S. Department of Energy (DOE) and its current and past contractors have worked closely with the South Carolina Department of Health and Environmental Control (SCDHEC), the U.S. Environmental Protection Agency (EPA) and the Nuclear Regulatory Commission (NRC) to develop and implement a framework for on-site low-level waste disposal and closure of the SRS waste tanks. The Atomic Energy Act of 1954, as amended, provides DOE the authority to manage defense-related radioactive waste. DOE Order 435.1 and its associated manual and guidance documents detail this radioactive waste management process. The DOE also has a requirement to consult with the NRC in determining that waste that formerly was classified as high-level waste can be safely managed as either low-level waste or transuranic waste. Once DOE makes a determination, NRC then has a responsibility to monitor DOE's actions in coordination with SCDHEC to ensure compliance with the Title 10 Code of Federal Regulations Part 61 (10CFR61), Subpart C performance objectives. The management of hazardous waste substances or components at SRS is regulated by SCDHEC and the EPA. The foundation for the interactions between DOE, SCDHEC and EPA is the SRS Federal Facility Agreement (FFA). Managing this array of requirements and successfully interacting with regulators, consultants and stakeholders is a challenging task but ensures thorough and thoughtful processes for disposing of the SRS low-level waste and the closure of the tank farm facilities. (authors) 19. Thermal Properties of LiCl-KCl Molten Salt for Nuclear Waste Separation SciTech Connect Sridharan, Kumar; Allen, Todd; Anderson, Mark; Simpson, Mike 2012-11-30 This project addresses both practical and fundamental scientific issues of direct relevance to operational challenges of the molten LiCl-KCl salt pyrochemical process, while providing avenues for improvements in the process. In order to understand the effects of the continually changing composition of the molten salt bath during the process, the project team will systematically vary the concentrations of rare earth surrogate elements, lanthanum, cerium, praseodymium, and neodymium, which will be added to the molten LiCl-KCl salt. They will also perform a limited number of focused experiments by the dissolution of depleted uranium. All experiments will be performed at 500 deg C. The project consists of the following tasks. Researchers will measure density of the molten salts using an instrument specifically designed for this purpose, and will determine the melting points with a differential scanning calorimeter. Knowledge of these properties is essential for salt mass accounting and taking the necessary steps to prevent melt freezing. The team will use cyclic voltammetry studies to determine redox potentials of the rare earth cations, as well as their diffusion coefficients and activities in the molten LiCl-KCl salt. In addition, the team will perform anodic stripping voltammetry to determine the concentration of the rare earth elements and their solubilities, and to develop the scientific basis for an on-line diagnostic system for in situ monitoring of the cation species concentration (rare earths in this case). Solubility and activity of the cation species are critically important for the prediction of the salt's useful lifetime and disposal. 20. Geohydrology of the northern Louisiana salt-dome basin pertinent to the storage of radioactive wastes; a progress report USGS Publications Warehouse Hosman, R.L. 1978-01-01 Salt domes in northern Louisiana are being considered as possible storage sites for nuclear wastes. The domes are in an area that received regional sedimentation through early Tertiary (Eocene) time with lesser amounts of Quaternary deposits. The Cretaceous-Tertiary accumulation is a few thousand feet thick; the major sands are regional aquifers that extend far beyond the boundaries of the salt-dome basin. Because of multiple aquifers, structural deformation, and variations in the hydraulic characteristics of cap rock, the ground-water hydrology around a salt dome may be highly complex. The Sparta Sand is the most productive and heavily used regional aquifer. It is either penetrated by or overlies most of the domes. A fluid entering the Sparta flow system would move toward one of the pumping centers, all at or near municipalities that pump from the Sparta. Movement could be toward surface drainage where local geologic and hydrologic conditions permit leakage to the surface or to a surficial aquifer. (Woodard-USGS) 1. Salton Sea Geothermal Field, California, as a near-field natural analog of a radioactive waste repository in salt SciTech Connect Elders, W.A.; Cohen, L.H. 1983-11-01 Since high concentrations of radionuclides and high temperatures are not normally encountered in salt domes or beds, finding an exact geologic analog of expected near-field conditions in a mined nuclear waste repository in salt will be difficult. The Salton Sea Geothermal Field, however, provides an opportunity to investigate the migration and retardation of naturally occurring U, Th, Ra, Cs, Sr and other elements in hot brines which have been moving through clay-rich sedimentary rocks for up to 100,000 years. The more than thirty deep wells drilled in this field to produce steam for electrical generation penetrate sedimentary rocks containing concentrated brines where temperatures reach 365/sup 0/C at only 2 km depth. The brines are primarily Na, K, Ca chlorides with up to 25% of total dissolved solids; they also contain high concentrations of metals such as Fe, Mn, Li, Zn, and Pb. This report describes the geology, geophysics and geochemistry of this system as a prelude to a study of the mobility of naturally occurring radionuclides and radionuclide analogs within it. The aim of this study is to provide data to assist in validating quantitative models of repository behavior and to use in designing and evaluating waste packages and engineered barriers. 128 references, 33 figures, 13 tables. 2. Separation of CsCl from a Ternary CsCl-LiCl-KCl Salt via a Melt Crystallization Technique for Pyroprocessing Waste Minimization SciTech Connect Ammon Williams; Supathorn Phongikaroon; Michael Simpson 2013-02-01 A parametric study has been conducted to identify the effects of several parameters on the separation of CsCl from molten LiCl-KCl salt via a melt crystallization process. A reverse vertical Bridgman technique was used to grow the salt crystals. The investigated parameters were: (1) the advancement rate, (2) the crucible lid configuration, (3) the amount of salt mixture, (4) the initial composition of CsCl, and (5) the temperature difference between the high and low furnace zones. From each grown crystal, samples were taken axially and analyzed using inductively coupled plasma mass spectrometry (ICP-MS). Results show that CsCl concentrations at the top of the crystals were low and increased to a maximum at the bottom of the salt. Salt (LiCl-KCl) recycle percentages for the experiments ranged from 50% to 75% and the CsCl composition in the waste salt was low. To increase the recycle percentage and the concentration of CsCl in the waste form, the possibility of using multiple crystallization stages was explored to further optimize the process. Results show that multiple crystallization stages are practical and the optimal experimental conditions should be operated at 5.0 mm/hr rate with a lid configuration and temperature difference of 200 C for a total of five crystallization stages. Under these conditions, up to 88% of the salt can be recycled. 3. Central neurogenic diabetes insipidus, syndrome of inappropriate secretion of antidiuretic hormone, and cerebral salt-wasting syndrome in traumatic brain injury. PubMed John, Cynthia A; Day, Michael W 2012-04-01 Central neurogenic diabetes insipidus, syndrome of inappropriate secretion of antidiuretic hormone, and cerebral salt-wasting syndrome are secondary events that affect patients with traumatic brain injury. All 3 syndromes affect both sodium and water balance; however, they have differences in pathophysiology, diagnosis, and treatment. Differentiating between hypernatremia (central neurogenic diabetes insipidus) and the 2 hyponatremia syndromes (syndrome of inappropriate secretion of antidiuretic hormone, and cerebral salt-wasting syndrome) is critical for preventing worsening neurological outcomes in patients with head injuries. PMID:22467619 4. Radioactive waste isolation in salt: peer review of the Office of Nuclear Waste Isolation's reports on preferred repository sites within the Palo Duro Basin, Texas SciTech Connect Fenster, D.; Edgar, D.; Gonzales, S.; Domenico, P.; Harrison, W.; Engelder, T.; Tisue, M. 1984-04-01 Documents are being submitted to the Salt Repository Project Office (SRPO) of the US Department of Energy (DOE) by Battelle Memorial Institute's Office of Nuclear Waste Isolation (ONWI) to satisfy milestones of the Salt Repository Project of the Civilian Radioactive Waste Management Program. Some of these documents are being reviewed by multidisciplinary groups of peers to ensure DOE of their adequacy and credibility. Adequacy of documents refers to their ability to meet the standards of the US Nuclear Regulatory Commission, as enunciated in 10 CFR 60, and the requirements of the National Environmental Policy Act and the Nuclear Waste Policy Act of 1982. Credibility of documents refers to the validity of the assumptions, methods, and conclusions, as well as to the completeness of coverage. This report summarizes Argonne's review of ONWI's two-volume draft report entitled Identification of Preferred Sites within the Palo Duro Basin: Vol. 1 - Palo Duro Location A, and Vol. 2 - Palo Duro Location B, dated January 1984. Argonne was requested by DOE to review these documents on January 17 and 24, 1984 (see App. A). The review procedure involved obtaining written comments on the reports from three members of Argonne's core peer review staff and three extramural experts in related research areas. The peer review panel met at Argonne on February 6, 1984, and reviewer comments were integrated into this report by the review session chairman, with the assistance of Argonne's core peer review staff. All of the peer review panelists concurred in the way in which their comments were represented in this report (see App. B). A letter report and a draft of this report were sent to SRPO on February 10, 1984, and April 17, 1984, respectively. 5 references. 5. Tank Waste Transport Stability: Summary of Slurry and Salt-Solution Studies for FY 2001 SciTech Connect Welch, T.D. 2002-06-07 Despite over 50 years of experience in transporting radioactive tank wastes to and from equipment and tanks at the Department of Energy's Hanford, Savannah River, and Oak Ridge sites, waste slurry transfer pipelines and process piping become plugged on occasion. At Hanford, several tank farm pipelines are no longer in service because of plugs. At Savannah River, solid deposits in the outlet line of the 2H evaporator have resulted in an unplanned extended downtime. Although waste transfer criteria and guidelines intended to prevent pipeline plugging are in place, they are not always adequate. To avoid pipeline plugging in the future, other factors that are not currently embodied in the transfer criteria may need to be considered. The work summarized here is being conducted to develop a better understanding of the chemical and waste flow dynamics during waste transfer. The goal is to eliminate pipeline plugs by improving analysis and engineering tools in the field that incorporate this understanding. 6. Technetium in alkaline, high-salt, radioactive tank waste supernate: Preliminary characterization and removal SciTech Connect Blanchard, D.L. Jr.; Brown, G.N.; Conradson, S.D. 1997-01-01 This report describes the initial work conducted at Pacific Northwest National Laboratory to study technetium (Tc) removal from Hanford tank waste supernates and Tc oxidation state in the supernates. Filtered supernate samples from four tanks were studied: a composite double shell slurry feed (DSSF) consisting of 70% from Tank AW-101, 20% from AP-106, and 10% from AP-102; and three complexant concentrate (CC) wastes (Tanks AN-107, SY-101, ANS SY-103) that are distinguished by having a high concentration of organic complexants. The work included batch contacts of these waste samples with Reillex{trademark}-HPQ (anion exchanger from Reilly Industries) and ABEC 5000 (a sorbent from Eichrom Industries), materials designed to effectively remove Tc as pertechnetate from tank wastes. A short study of Tc analysis methods was completed. A preliminary identification of the oxidation state of non-pertechnetate species in the supernates was made by analyzing the technetium x-ray absorption spectra of four CC waste samples. Molybdenum (Mo) and rhenium (Re) spiked test solutions and simulants were tested with electrospray ionization-mass spectrometry to evaluate the feasibility of the technique for identifying Tc species in waste samples. 7. Stepwise splitting of ribosomal proteins from yeast ribosomes by LiCl. PubMed Piir, Kerli; Tamm, Tiina; Kisly, Ivan; Tammsalu, Triin; Remme, Jaanus 2014-01-01 Structural studies have revealed that the core of the ribosome structure is conserved among ribosomes of all kingdoms. Kingdom-specific ribosomal proteins (r-proteins) are located in peripheral parts of the ribosome. In this work, the interactions between rRNA and r-proteins of eukaryote Saccharomyces cerevisiae ribosome were investigated applying LiCl induced splitting and quantitative mass spectrometry. R-proteins were divided into four groups according to their binding properties to the rRNA. Most yeast r-proteins are removed from rRNA by 0.5-1 M LiCl. Eukaryote-specific r-proteins are among the first to dissociate. The majority of the strong binders are known to be required for the early ribosome assembly events. As compared to the bacterial ribosome, yeast r-proteins are dissociated from rRNA at lower ionic strength. Our results demonstrate that the nature of protein-RNA interactions in the ribosome is not conserved between different kingdoms. PMID:24991888 8. Stepwise Splitting of Ribosomal Proteins from Yeast Ribosomes by LiCl PubMed Central Piir, Kerli; Tamm, Tiina; Kisly, Ivan; Tammsalu, Triin; Remme, Jaanus 2014-01-01 Structural studies have revealed that the core of the ribosome structure is conserved among ribosomes of all kingdoms. Kingdom-specific ribosomal proteins (r-proteins) are located in peripheral parts of the ribosome. In this work, the interactions between rRNA and r-proteins of eukaryote Saccharomyces cerevisiae ribosome were investigated applying LiCl induced splitting and quantitative mass spectrometry. R-proteins were divided into four groups according to their binding properties to the rRNA. Most yeast r-proteins are removed from rRNA by 0.5–1 M LiCl. Eukaryote-specific r-proteins are among the first to dissociate. The majority of the strong binders are known to be required for the early ribosome assembly events. As compared to the bacterial ribosome, yeast r-proteins are dissociated from rRNA at lower ionic strength. Our results demonstrate that the nature of protein-RNA interactions in the ribosome is not conserved between different kingdoms. PMID:24991888 9. [Salt-wasting nephropathy induced by foscarnet treatment for HHV-6 encephalitis in a hematopoietic stem cell transplant]. PubMed Najima, Yuho; Ohashi, Kazuteru; Ando, Minoru; Koshida, Akira; Yamashita, Takuya; Akiyama, Hideki; Sakamaki, Hisashi 2008-01-01 A 51-year-old woman with myelodysplastic syndrome developed HHV-6 encephalitis on day 34 after unrelated bone marrow transplantation. Although prompt treatment with foscarnet stabilized encephalitis and there were no serious neurological sequelae, the patient developed both hyponatremia and natriuresis 11 days after intravenous administration of foscarnet. Subsequent investigation demonstrated hyponatremia due to salt-wasting nephropathy induced by foscarnet. Discontinuation of foscarnet and fluid replacement therapy with adequate sodium chloride (NaCl) resulted in a gradual resolution of hyponatremia and natriuresis. Currently, 8 months after these clinical events, the patient is under outpatient treatment for persistent nephropathy that requires small amounts of oral NaCl supplement, but she is otherwise in good clinical condition. PMID:18277595 10. Radioactive waste isolation in salt: rationale and methodology for Argonne-conducted reviews of site characterization programs SciTech Connect Harrison, W.; Ditmars, J.D.; Tisue, M.W.; Hambley, D.F.; Fenster, D.F.; Rote, D.M. 1985-07-01 Both regulatory and technical concerns must be addressed in Argonne-conducted peer reviews of site characterization programs for individual sites for a high-level radioactive waste repository in salt. This report describes the regulatory framework within which reviews must be conducted and presents background information on the structure and purpose of site characterization programs as found in US Nuclear Regulatory Commission (NRC) Regulatory Guide 4.17 and Title 10, Part 60, of the Code of Federal Regulations. It also presents a methodology to assist reviewers in addressing technical concerns relating to their respective areas of expertise. The methodology concentrates on elements of prime importance to the US Department of Energy's advocacy of a given salt repository system during the NRC licensing process. Instructions are given for reviewing 12 site characterization program elements, starting with performance objectives, performance issues, and levels of performance of repository subsystem components; progressing through performance assessment; and ending with plans for data acquisition and evaluation. The success of a site characterization program in resolving repository performance issues will be determined by judging the likelihood that the proposed data acquisition activities will reduce uncertainties in the performance predictions. 8 refs., 3 figs., 5 tabs. 11. Measurement of unsaturated hydraulic properties of salt cake simulant relevant to hanford and SRS high-level waste tanks using a pilot-scale setup SciTech Connect Tachiev, G.; Yaari, G.; Long, S.; Srivastava, R.; Roelant, D. 2007-07-01 Closure of the remaining tanks and final disposition of the radioactive waste is a high priority task at both Savannah River Site (SRS) and Hanford. The radioactive waste in the tanks are generally found in layers: supernate (on top) containing soluble fission products, and salt-cake and sludge (on the bottom of the tank) containing insoluble actinides. One strategy for minimizing the waste volume is to segregate the low curie salt waste from the high curie salt supernate by draining the supernate and interstitial salt solution from the salt-cake. The retrieval of the interstitial fluid will require knowledge of relevant properties of salt-cake waste including drainage parameters, more specifically, its hydraulic properties. The hydraulic parameters of the salt-cake have significance with respect to: 1) Kinetics of the retrieval process; and 2) Equilibrium conditions of the drainage. While the saturated hydraulic properties of the salt waste (hydraulic conductivity in the vertical and horizontal direction) can be used to determine the kinetics of the flow through the salt waste, the unsaturated properties are needed in order to assess not only the time frame of tank drainage but also the equilibrium conditions. How much and how fast fluid can be drained at given initial and boundary conditions (atmospheric pressure and temperature) can be analyzed. A series of dissolution and drainage experiments was conducted using S-112, S-109 and Tank 41 simulants in a pilot-scale column (1' diameter, 10' high). The major goal of these experiments was to determine the hydraulic parameters of flow through the column and the dissolution patterns upon addition of fresh water. The hydraulic experiments were conducted using interstitial fluid as well as fresh water. A series of one-step outflow experiments were used to evaluate the drainage patterns for salt-bed heights ranging from 1' to 8'. Measured data include bulk densities and saturated hydraulic conductivities as a function of temperature, and water retention characteristics for ambient temperature. Experiments in the range of 22-43 deg. C confirmed that the hydraulic conductivity in the vertical direction is strongly dependent on the temperature with variations most likely caused by physical-chemical changes in the salt-cake structure and the viscosity of the interstitial fluid. Evidence for the changes in salt-cake structure include: increased effluent specific gravity to 1.47, appearance of voids in the column, and a 10% reduction of the column height. A two-orders-of-magnitude difference in viscosity was observed when the temperature increased from 22 deg. C to 43 deg. C. The liquid retention parameters were determined using static retention points (derived from the multi-step outflow time series). Inverse analyses of the outflow data yielded additional unsaturated hydraulic conductivity parameters. The inverse analysis was performed by fixing saturated and residual water contents and the pore connectivity factor. Numerical simulations showed that the amount of drained fluid is directly related to van Genuchten's shape parameters n and {alpha} . On the other hand, the rate of drainage was directly related to saturated hydraulic conductivity. The recommended values for {alpha} and n with uncertainties in the parenthesis are respectively: 3.2(0.32) 1/m and 2.6(0.4). The recommended value for saturated hydraulic conductivity of salt-cake is 6.0E-06(3E-06). This work provides critical data about unsaturated hydraulic properties which can be used in numerical models to better predict removal of interstitial liquid from tanks that directly impact tank and site cleanup schedule and costs. (authors) 12. Molten Salt Oxidation: A Thermal Technology for Waste Treatment and Demilitarization SciTech Connect Hsu, P C; Watkins, B; Pruneda, C; Kwak, S 2001-08-23 MSO is a good alternative to incineration for the treatment of a variety of organic wastes including obsolete explosives, low-level mixed waste streams, PCB contaminated oils, spent resins and carbon. The Lawrence Livermore National Laboratory (LLNL) has demonstrated the MSO process for the effective destruction of explosives, explosives-contaminated materials, and other wastes on a 1.5 kg/hr bench-scale unit and in an integrated MSO facility capable of treating 8 kg/hr of low-level radioactive mixed wastes. LLNL, under the direction and support of the Joint Demilitarization Technology (JDT) program, is currently building an integrated MSO plant for destroying explosives, explosives-contaminated sludge and explosives-contaminated activated charcoal. In a parallel effort, LLNL also provides technical support to DOE for the implementation of the MSO technology at industrial scale at Richland, Washington. Over 30 waste streams have been demonstrated with LLNL-built MSO systems. In this paper we will present our latest experimental data, our operational experience with MSO and also discuss its process capabilities. 13. Expedited demonstration of molten salt mixed waste treatment technology. Addendum 1 SciTech Connect Holtz, E.H. von; Hopper, R.W.; Adamson, M.G. 1995-04-27 The Final Forms portion (Section 4) of the TTP SF-2410-03 final report was incomplete. This was noted under the subsection Task Variances. The present report documents the work that was unfinished at that time, arranged in accord with the subsections of the Final Report. An assessment of the overall immobilization efficacy of polymer microencapsulation, as supported by this study, has been added. The study and demonstration of the polyethylene microencapsulation of salt residues is continuing under other auspices. A stand-alone report combining the results of the continuation with the contents of this memorandum and of Section 4 of the Final Report will be issued in later this year. 14. Multi physics modeling of a molten-salt electrolytic process for nuclear waste treatment NASA Astrophysics Data System (ADS) Kim, K. R.; Choi, S. Y.; Kim, J. G.; Paek, S.; Ahn, D. H.; Kwon, S. W.; Shim, J. B.; Kim, S. H.; Lee, H. S.; Park, B. G.; Yi, K. W.; Hwang, I. S. 2010-03-01 Multi physics electrochemical modeling in a framework of Computational Fluid Dynamics (CFD) code was proposed and dealt with in detail to simulate the electro-transport behaviour that appears in a molten-salt electrolytic system. The modeling approach in this study is focused on the mass transport and current arising due to the concentration and the surface overpotential based on a cell configuration and electrolyte turbulence. This comprehensive modelling approach was applied and compared to electroplating model in a prepared rotating cylinder Hull (RCH) cell system. 15. Radioactive waste isolation in salt: peer review of the Office of Nuclear Waste Isolation's Geochemical Program Plan SciTech Connect Harrison, W.; Seitz, M.; Fenster, D.; Lerman, A.; Brookins, D.; Tisue, M. 1984-02-01 Describe the management program for coordinating subcontractors and their work, and integrating research results. Appropriate flowcharts should be included. Provide more information on the overall scope of the program. For each subcontractor, provide specific workscopes that indicate whether analytical activities are developmental or routine, approximate number of analyses to be made, and something of the adequacy of the analyses to meet program goals. Indicate interfaces with other earth-science disciplines like hydrology and with other groups doing relevant geochemical research and engineering design. Address the priorities for each activity or group of activities. High priority should be given to early development of a geochemical statement of what constitutes suitable salt for a repository. Reference standard procedures for sampling, sample preservation, and sample analysis wherever appropriate or, if not appropriate, indicate that any deviations from standard procedures will be documented. Ensure that appropriate quality assurance procedures will be followed for the procedures listed above. Include specific procedures for the choice, verification, validation, and documentation of computer codes related to the geochemical aspects of repository performance assessment. Include activities addressing regional hydrochemistry and make clear that each principal hydrogeologic unit at each site will be studied geochemically. Indicate that proposed plans for obtaining hydrogeochemical data will be included in each site characterization plan. Describe how site geochemical stability will be handled, especially with respect to dissolution, postemplacement geochemistry, human influences, and climatic variations. Minor recommendations and suggested improvements in the text of the plan are given in Sec. 5. 16. Recommendation for Using Smaller (0.1 micro sign) Pore-Size Media for Filtration in Salt Waste Processing Project SciTech Connect Poirier, M.R. 2003-05-02 Based on experimental studies with simulated and actual wastes, we recommend adopting the use of 0.1-micron pore-size, sintered stainless-steel filter elements within the design of the Salt Waste Processing Facility. Furthermore, adopting the smaller pore size elements for the Actinide Removal Process would result in a significant risk to the start-up schedule due to delays for buying, installing, and testing new equipment. The existing 0.5-micron pore-size filters will provide nearly equivalent service with no additional capital investment. Unless the planned filter test at Building 512-S fails to meet specifications, the project should proceed with the existing equipment, including spares. When the existing equipment reaches the end of the service life, management can consider replacement with the smaller pore-size elements. The laboratory studies indicate that use of the smaller pore size equipment will result in greater protection against particulate fines passing to downstream facilities while giving equivalent or superior processing rates than provided by the 0.5-micron elements. 17. Properties of salt-saturated concrete and grout after six years in situ at the Waste Isolation Pilot Plant SciTech Connect Wakeley, L.D.; Harrington, P.T.; Weiss, C.A. Jr. 1993-06-01 Samples of concrete and grout were recovered from short boreholes in the repository floor at the Waste Isolation Pilot Plant more than six years after the concrete and grout were placed. Plugs from the Plug Test Matrix of the Plugging and Sealing Program of Sandia National Laboratories were overcored to include a shell of host rock. The cores were analyzed at the Waterways Experiment Station to assess their condition after six years of service, having potentially been exposed to those aspects of their service environment (salt, brine, fracturing, anhydrite, etc.) that could cause deterioration. Measured values of compressive strength and pulse velocity of both the grout and the concrete equaled or exceeded values from tests performed on laboratory-tested samples of the same mixtures at ages of one month to one year after casting. The phase assemblages had changed very little. Materials performed as intended and showed virtually no chemical or physical evidence of deterioration. The lowest values for strength and pulse velocity were measured for samples taken from the Disturbed Rock Zone, indicating the influence of cracking in this zone on the properties of enclosed seal materials. There was evidence of movement of brine in the system. Crystalline phases containing magnesium, potassium, sulfate, and other ions had been deposited on free surfaces in fractures and pilot holes. There was a reaction rim in the anhydrite immediately surrounding each recovered borehole plug, suggesting interaction between grout or concrete and host rock. However, the chemical changes apparent in this reaction rim were not reflected in the chemical composition of the adjacent concrete or grout. The grout and concrete studied here showed no signs of the deterioration found to have occurred in some parts of the concrete liner of the Waste Isolation Pilot Plant waste handling shaft. 18. On the importance of coupled THM processes to predict the long-term response of a generic salt repository for high-level nuclear waste NASA Astrophysics Data System (ADS) Blanco Martin, L.; Rutqvist, J.; Birkholzer, J. T. 2013-12-01 Salt is a potential medium for the underground disposal of nuclear waste because it has several assets, in particular its ability to creep and heal fractures generated by excavation and its water and gas tightness in the undisturbed state. In this research, we focus on disposal of heat-generating nuclear waste (such as spent fuel) and we consider a generic salt repository with in-drift emplacement of waste packages and subsequent backfill of the drifts with run-of-mine crushed salt. As the natural salt creeps, the crushed salt backfill gets progressively compacted and an engineered barrier system is subsequently created. In order to evaluate the integrity of the natural and engineered barriers over the long-term, it is important to consider the coupled effects of the thermal, hydraulic and mechanical processes that take place. In particular, the results obtained so far show how the porosity reduction of the crushed salt affects the saturation and pore pressure evolution throughout the repository, both in time and space. Such compaction is induced by the stress and temperature regime within the natural salt. Also, transport properties of the host rock are modified not only by thermo-mechanically and hydraulically-induced damaged processes, but also by healing/sealing of existing fractures. In addition, the THM properties of the backfill evolve towards those of the natural salt during the compaction process. All these changes are based on dedicated laboratory experiments and on theoretical considerations [1-3]. Different scenarios are modeled and compared to evaluate the relevance of different processes from the perspective of effective nuclear waste repositories. The sensitivity of the results to some parameters, such as capillarity, is also addressed. The simulations are conducted using an updated version of the TOUGH2-FLAC3D simulator, which is based on a sequential explicit method to couple flow and geomechanics [4]. A new capability for large strains and creep has been introduced and validated. The time-dependent geomechanical response of salt is determined using the Lux/Wolters constitutive model, developed at Clausthal University of Technology (Germany). References: [1] R. Wolters, and K.-H. Lux. Evaluation of Rock Salt Barriers with Respect to Tightness: Influence of Thermomechanical Damage, Fluid Infiltration and Sealing/Healing. Proceedings of the 7th International Conference on the Mechanical Behavior of Salt (SaltMech7). Paris: Balkema, Rotterdam (2012). [2] W. Bechthold et al., Backfilling and Sealing of Underground Repositories for Radioactive Waste in Salt (BAMBUS Project), European Atomic Energy Community, Report EUR19124 EN (1999). [3] J. Kim, E.L Sonnenthal and J. Rutqvist, 'Formulation and sequential numerical algorithms of coupled fluid/heat flow and geomechanics for multiple porosity materials', Int. J. Numer. Meth. Engng., 92, 425 (2012). [4] J. Rutqvist. Status of the TOUGH-FLAC simulator and recent applications related to coupled fluid flow and crustal deformations. Computational Geosciences, 37, 739-750 (2011). 19. Electronic excitations of bulk LiCl from many-body perturbation theory SciTech Connect Jiang, Yun-Feng; Wang, Neng-Ping; Rohlfing, Michael 2013-12-07 We present the quasiparticle band structure and the optical excitation spectrum of bulk LiCl, using many-body perturbation theory. Density-functional theory is used to calculate the ground-state geometry of the system. The quasiparticle band structure is calculated within the GW approximation. Taking the electron-hole interaction into consideration, electron-hole pair states and optical excitations are obtained by solving the Bethe-Salpeter equation for the electron-hole two-particle Green function. The calculated band gap is 9.5 eV, which is in good agreement with the experimental result of 9.4 eV. And the calculated optical absorption spectrum, which contains an exciton peak at 8.8 eV and a resonant-exciton peak at 9.8 eV, is also in good agreement with experimental data. 20. LiCl Dehumidifier LiBr absorption chiller hybrid air conditioning system with energy recovery DOEpatents Ko, Suk M. 1980-01-01 This invention relates to a hybrid air conditioning system that combines a solar powered LiCl dehumidifier with a LiBr absorption chiller. The desiccant dehumidifier removes the latent load by absorbing moisture from the air, and the sensible load is removed by the absorption chiller. The desiccant dehumidifier is coupled to a regenerator and the desiccant in the regenerator is heated by solar heated hot water to drive the moisture therefrom before being fed back to the dehumidifier. The heat of vaporization expended in the desiccant regenerator is recovered and used to partially preheat the driving fluid of the absorption chiller, thus substantially improving the overall COP of the hybrid system. 1. Selective Extraction and In Situ Reduction of Precious Metal Salts from Model Waste To Generate Hybrid Gels with Embedded Electrocatalytic Nanoparticles. PubMed Okesola, Babatunde O; Suravaram, Sindhu K; Parkin, Alison; Smith, David K 2016-01-01 A hydrogel based on 1,3:2,4-dibenzylidenesorbitol (DBS), modified with acyl hydrazides which extracts gold/silver salts from model waste is reported, with preferential uptake of precious heavy metals over other common metals. Reduction of gold/silver salts occurs spontaneously in the gel to yield metal nanoparticles located on the gel nanofibers. High nanoparticle loadings can be achieved, endowing the gel with electrochemical activity. These hybrid gels exhibit higher conductances than gels doped with carbon nanotubes, and can be used to modify electrode surfaces, enhancing electrocatalysis. We reason this simple, industrially and environmentally relevant approach to conducting materials is of considerable significance. PMID:26549625 2. Effects of resource activities upon repository siting and waste containment with reference to bedded salt SciTech Connect Ashby, J.; Rowe, J. 1980-02-01 The primary consideration for the suitability of a nuclear waste repository site is the overall ability of the repository to safely contain radioactive waste. This report is a discussion of the past, present, and future effects of resource activities on waste containment. Past and present resource activities which provide release pathways (i.e., leaky boreholes, adjacent mines) will receive initial evaluation during the early stages of any repository site study. However, other resource activities which may have subtle effects on containment (e.g., long-term pumping causing increased groundwater gradients, invasion of saline water causing lower retardation) and all potential future resource activities must also be considered during the site evaluation process. Resource activities will affect both the siting and the designing of repositories. Ideally, sites should be located in areas of low resource activity and low potential for future activity, and repository design should seek to eliminate or minimize the adverse effects of any resource activity. Buffer zones should be created to provide areas in which resource activities that might adversely affect containment can be restricted or curtailed. This could mean removing large areas of land from resource development. The impact of these frozen assets should be assessed in terms of their economic value and of their effect upon resource reserves. This step could require a major effort in data acquisition and analysis followed by extensive numerical modeling of regional fluid flow and mass transport. Numerical models should be used to assess the effects of resource activity upon containment and should include the cumulative effects of different resource activities. Analysis by other methods is probably not possible except for relatively simple cases. 3. Effect of biosolid waste compost on soil respiration in salt-affected soils NASA Astrophysics Data System (ADS) Raya, Silvia; Gómez, Ignacio; García, Fuensanta; Navarro, José; Jordán, Manuel Miguel; Belén Almendro, María; Martín Soriano, José 2013-04-01 A great part of mediterranean soils are affected by salinization. This is an important problem in semiarid areas increased by the use of low quality waters, the induced salinization due to high phreatic levels and adverse climatology. Salinization affects 25% of irrigated agriculture, producing important losses on the crops. In this situation, the application of organic matter to the soil is one of the possible solutions to improve their quality. The main objective of this research was to asses the relation between the salinity level (electrical conductivity, EC) in the soil and the response of microbial activity (soil respiration rate) after compost addition. The study was conducted for a year. Soil samples were collected near to an agricultural area in Crevillente and Elche, "El Hondo" Natural Park (Comunidad de Regantes from San Felipe Neri). The experiment was developed to determine and quantify the soil respiration rate in 8 different soils differing in salinity. The assay was done in close pots -in greenhouse conditions- containing soil mixed with different doses of sewage sludge compost (2, 4 and 6%) besides the control. They were maintained at 60% of water holding capacity (WHC). Soil samples were analyzed every four months for a year. The equipment used to estimate the soil respiration was a Bac-Trac and CO2 emitted by the soil biota was measured and quantified by electrical impedance changes. It was observed that the respiration rate increases as the proportion of compost added to each sample increases as well. The EC was incremented in each sampling period from the beginning of the experiment, probably due to the fact that soils were in pots and lixiviation was prevented, so the salts couldńt be lost from soil. Over time the compost has been degraded and, it was more susceptible to be mineralized. Salts were accumulated in the soil. Also it was observed a decrease of microbial activity with the increase of salinity in the soil. Keywords: soil respiration, compost, electrical conductivity, salinization, Bac-Trac References: Abdelbasset Lakhdar, Mokded Rabhi, Tahar Ghnaya, Francesco Montemurro, Naceur Jedidi , Chedly Abdelly. Effectiveness of compost use in salt-affected soil. Journal of Hazardous Materials 171 (2009) pp 29-37. M. Tejada, C. Garcia, J.L. Gonzalez , M.T. Hernandez . Use of organic amendment as a strategy for saline soil remediation:Influence on the physical, chemical and biological properties of soil. Soil Biology & Biochemistry 38 (2006) pp 1413-1421. I. Gomez; J.M. Disla Soriano; J. Navarro-Pedreño; F. García-Orenes; M.B. Almendro-Candel; M.M. Jordan. Quantification of soil respiration in different saline soil of Alicante (Spain). EGU General Assembly (2012). Viena. Ed. Geophysycal Research Abstracts. Vol 14 EGU2012-2399,(2012). (Acknowledgements: This work was supported by the Spanish MICINN. Project Ref.: CGL2009-11194) 4. Summary of four release consequence analyses for hypothetical nuclear waste repositories in salt and granite SciTech Connect Cole, C.R.; Bond, F.W. 1980-12-01 Release consequence methology developed under the Assessment of Effectiveness of Geologic Isolation Systems (AEGIS) program has now been applied to four hypothetical repository sites. This paper summarizes the results of these four studies in order to demonstrate that the far-field methodology developed under the AEGIS program offers a practical approach to the post-closure safety assessment of nuclear waste repositories sited in deep continental geologic formations. The four studies are briefly described and compared according to the following general categories: physical description of the repository (size, inventory, emplacement depth); geologic and hydrologic description of the site and the conceptual hydrologic model for the site; description of release scenario; hydrologic model implementation and results; engineered barriers and leach rate modeling; transport model implementation and results; and dose model implementation and results. These studies indicate the following: numerical modeling is a practical approach to post-closure safety assessment analysis for nuclear waste repositories; near-field modeling capability needs improvement to permit assessment of the consequences of human intrusion and pumping well scenarios; engineered barrier systems can be useful in mitigating consequences for postulated release scenarios that short-circuit the geohydrologic system; geohydrologic systems separating a repository from the natural biosphere discharge sites act to mitigate the consequences of postulated breaches in containment; and engineered barriers of types other than the containment or absorptive type may be useful. 5. Site characterization plan conceptual design report for a high-level nuclear waste repository in salt, vertical emplacement mode: Volume 1 SciTech Connect Not Available 1987-12-01 This Conceptual Design Report describes the conceptual design of a high-level nuclear waste repository in salt at a proposed site in Deaf Smith County, Texas. Waste receipt, processing, packing, and other surface facility operations are described. Operations in the shafts underground are described, including waste hoisting, transfer, and vertical emplacement. This report specifically addresses the vertical emplacement mode, the reference design for the repository. Waste retrieval capability is described. The report includes a description of the layout of the surface, shafts, and underground. Major equipment items are identified. The report includes plans for decommissioning and sealing of the facility. The report discusses how the repository will satisfy performance objectives. Chapters are included on basis for design, design analyses, and data requirements for completion of future design efforts. 105 figs., 52 tabs. 6. Site characterization plan conceptual design report for a high-level nuclear waste repository in salt, horizontal emplacment mode: Volume 1 SciTech Connect Not Available 1987-12-01 This Conceptual Design Report describes the conceptual design of a high-level nuclear waste repository in salt at a proposed site in Deaf Smith County, Texas. Waste receipt, processing, packaging, and other surface facility operations are described. Operations in the shafts and underground are described, including waste hoisting, transfer, and horizontal emplacement. This report specifically addresses the horizontal emplacement mode, the passive alternate design for the repository. Waste retrieval capability is described. The report includes a description of the layout of the surface, shafts, and underground. Major equipment items are identified. The report includes plans for decommissioning and sealing of the facility. The report discusses how the repository will satisfy performance objectives. Chapters are included on basis for design, design analyses, and data requirements for completion of future design efforts. 105 figs., 52 tabs. 7. Repository environmental parameters and models/methodologies relevant to assessing the performance of high-level waste packages in basalt, tuff, and salt SciTech Connect Claiborne, H.C.; Croff, A.G.; Griess, J.C.; Smith, F.J. 1987-09-01 This document provides specifications for models/methodologies that could be employed in determining postclosure repository environmental parameters relevant to the performance of high-level waste packages for the Basalt Waste Isolation Project (BWIP) at Richland, Washington, the tuff at Yucca Mountain by the Nevada Test Site, and the bedded salt in Deaf Smith County, Texas. Guidance is provided on the identify of the relevant repository environmental parameters; the models/methodologies employed to determine the parameters, and the input data base for the models/methodologies. Supporting studies included are an analysis of potential waste package failure modes leading to identification of the relevant repository environmental parameters, an evaluation of the credible range of the repository environmental parameters, and a summary of the review of existing models/methodologies currently employed in determining repository environmental parameters relevant to waste package performance. 327 refs., 26 figs., 19 tabs. 8. Effects of metal salt addition on odor and process stability during the anaerobic digestion of municipal waste sludge. PubMed Abbott, Timothy; Eskicioglu, Cigdem 2015-12-01 Anaerobic digestion (AD) is an effective way to recover energy and nutrients from organic waste; however, several issues including the solubilization of bound nutrients and the production of corrosive, highly odorous and toxic volatile sulfur compounds (VSCs) in AD biogas can limit its wider adoption. This study explored the effects of adding two different doses of ferric chloride, aluminum sulfate and magnesium hydroxide directly to the feed of complete mix semi-continuously fed mesophilic ADs on eight of the most odorous VSCs in AD biogas at three different organic loading rates (OLR). Ferric chloride was shown to be extremely effective in reducing VSCs by up to 87%, aluminum sulfate had the opposite effect and increased VSC levels by up to 920%, while magnesium hydroxide was not shown to have any significant impact. Ferric chloride, aluminum sulfate and magnesium hydroxide were effective in reducing the concentration of orthophosphate in AD effluent although both levels of alum addition caused digester failure at elevated OLRs. Extensive foaming was observed within the magnesium hydroxide dosed digesters, particularly at higher doses and high OLRs. Certain metal salt additions may be a valuable tool in overcoming barriers to AD and to meet regulatory targets. PMID:26260964 9. Radioactive waste isolation in salt: peer review of the Office of Nuclear Waste Isolation's plan to decommission and reclaim exploratory shafts and related facilities SciTech Connect Fenster, D.F.; Schubert, J.P.; Zellmer, S.D.; Harrison, W.; Simpson, D.G.; Busch, J.S. 1984-07-01 The following recommendations are made for improving the Office of Nuclear Waste Isolation's plan for decommissioning and reclaiming exploratory shafts and other facilities associated with site characterization: (1) Discuss more comprehensively the technical aspects of activities related to decommissioning and reclamation. More detailed information will help convince the staff of the US Nuclear Regulatory Commission and others that the activities as outlined in the plan are properly structured and that the stated goals can be achieved. (2) Address in considerably greater detail how the proposed activities will satisfy specific federal, state, and local laws and regulations. (3) State clearly the precise purpose of the plan, preferably at the beginning and under an appropriate heading. (4) Also under an appropriate heading and immediately after the section on purpose, describe the scope of the plan. The tasks covered by this plan and closely related tasks covered by other appropriate plans should be clearly differentiated. (5) Discuss the possible environmental effects of drilling the exploratory shaft, excavating drifts in salt, and drilling boreholes as part of site characterization. Mitigation activities should be designed to counter specific potential impacts. High priority should be given to minimizing groundwater contamination and restoring the surface to a condition consistent with the proposed land use following completion of characterization activities at sites not chosen for repository construction. (6) Define ambiguous technical terms, either in the text when first introduced or in an appended glossary. 10. Laboratory creep and mechanical tests on salt data report (1975-1996): Waste Isolation Pilot Plant (WIPP) thermal/structural interactions program SciTech Connect Mellegard, K.D. 1997-02-01 The Waste Isolation Pilot Plant (WIPP), a facility located in a bedded salt formation in Carlsbad, New Mexico, is being used by the U.S. Department of Energy to demonstrate the technology for safe handling and disposal of transuranic wastes produced by defense activities in the United States. In support of that demonstration, mechanical tests on salt were conducted in the laboratory to characterize material behavior at the stresses and temperatures expected for a nuclear waste repository. Many of those laboratory test programs have been carried out in the RE/SPEC Inc. rock mechanics laboratory in Rapid City, South Dakota; the first program being authorized in 1975 followed by additional testing programs that continue to the present. All of the WIPP laboratory data generated on salt at RE/SPEC Inc. over the last 20 years is presented in this data report. A variety of test procedures were used in performance of the work including quasi-static triaxial compression tests, constant stress (creep) tests, damage recovery tests, and multiaxial creep tests. The detailed data is presented in individual plots for each specimen tested. Typically, the controlled test conditions applied to each specimen are presented in a plot followed by additional plots of the measured specimen response. Extensive tables are included to summarize the tests that were performed. Both the tables and the plots contain cross-references to the technical reports where the data were originally reported. Also included are general descriptions of laboratory facilities, equipment, and procedures used to perform the work. 11. Protein refolding in predominantly organic media markedly enhanced by common salts SciTech Connect Rariy, R.V.; Klibanov, A.M. 1999-03-20 The refolding/reoxidation of unfolded/reduced hen egg-white lysozyme was investigated in a variety of predominantly nonaqueous media consisting of protein-dissolving organic solvents and water. It was discovered that LiCl and other common salts dramatically increased the refolding yield of lysozyme in such nonaqueous systems, while reducing it in water. The mechanism of this surprising phenomenon appears to involve salt-induced suppression of nonspecific lysozyme aggregation during refolding due to an enhanced protein solubility. 12. Worth its salt? NASA Astrophysics Data System (ADS) The idea that all underground salt deposits can serve as storage sites for toxic and nuclear waste does not always hold waterliterally. According to Daniel Ronen and Brian Berkowitz of Israel's Weizmann Institute of Science and Yoseph Yechieli of the Geological Survey of Israel, some buried salt layers are in fact highly conductive of liquids, suggesting that wastes buried in their confines could easily leech into groundwater and nearby soil.When drilling three wells into a 10,000-year-old salt layer near the Dead Sea, the researchers found that groundwater had seeped into the layer and had absorbed some of its salt. 13. Ceramicrete stabilization of radioactive-salt-containing liquid waste and sludge water. Final CRADA report. SciTech Connect Ehst, D.; Nuclear Engineering Division 2010-08-04 It was found that the Ceramicrete Specimens incorporated the Streams 1 and 2 sludges with the adjusted loading about 41.6 and 31.6%, respectively, have a high solidity. The visible cracks in the matrix materials and around the anionite AV-17 granules included could not obtain. The granules mentioned above fixed by Ceramicrete matrix very strongly. Consequently, we can conclude that irradiation of Ceramecrete matrix, goes from the high radioactive elements, not result the structural degradation. Based on the chemical analysis of specimens No.462 and No.461 used it was shown that these matrix included the formation elements (P, K, Mg, O), but in the different samples their correlations are different. These ratios of the content of elements included are about {+-} 10%. This information shows a great homogeneity of matrix prepared. In the list of the elements founded, expect the matrix formation elements, we detected also Ca and Si (from the wollastonite - the necessary for Ceramicrete compound); Na, Al, S, O, Cl, Fe, Ni also have been detected in the Specimen No.642 from the waste forms: NaCl, Al(OH){sub 3}, Na{sub 2}SO{sub 4}. Fe(OH){sub 3}, nickel ferrocyanide and Ni(NO{sub 3})2. The unintelligible results also were found from analysis of an AV-17 granules, in which we obtain the great amount of K. The X-ray radiographs of the Ceramicrete specimens with loading 41.4 % of Stream 1 and 31.6% of Stream 2, respectively showed that the realization of the advance technology, created at GEOHKI, leads to formation of excellent ceramic matrix with high amount of radioactive streams up to 40% and more. Really, during the interaction with start compounds MgO and KH{sub 2}PO{sub 4} with the present of H{sub 3}BO{sub 3} and Wollastonite this process run with high speed under the controlled regimes. That fact that the Ceramicrete matrix with 30-40% of Streams 1 and 2 have a crystalline form, not amorphous matter, allows to permit that these matrix should be very stable, reliable for incorporation of a radionuclides. 14. Recommended new criteria for the selection of nuclear waste repository sites in Columbia River basalt and US Gulf Coast domed salt SciTech Connect Steinborn, T.L.; Wagoner, J.L.; Qualheim, B.; Fitts, C.R.; Stetkar, R.E.; Turnbull, R.W. 1980-06-16 Screening criteria and specifications are recommended to aid in the evaluation of sites proposed for nuclear waste disposal in basalt and domed salt. The recommended new criteria proposed in this report are intended to supplement existing repository-related criteria for nuclear waste disposal. The existing criteria are contained in 10 CFR 60 sections which define siting criteria of the Nuclear Regulatory Commission (NRC), and ONWI 33(2) which defines siting criteria of the Office of Nuclear Waste Isolation (ONWI) for the Department of Energy. The specifications are conditions or parameter values that the authors recommend be applied in site acceptance evaluations. The siting concerns covered in this report include repository depth, host rock extent, seismic setting, structural and tectonic conditions, groundwater and rock geochemistry, volcanism, surface and subsurface hydrology, and socioeconomic issues, such as natural resources, land use, and population distribution. 15. Effect of LiCl pretreatment on cholinomimetic-induced seizures and seizure-induced brain edema in rats. PubMed Terry, J B; Padzernik, T L; Nelson, S R 1990-06-22 Male Sprague-Dawley rats received LiCl (5 mEq/kg; sc) or saline 24 h prior to injection of cholinomimetics. Physostigmine (PHY, 0.54-0.80 mg/kg), diisopropylfluorophosphate (DFP, 1.3-2.5 mg/kg), pilocarpine (PIL, 23-30 mg/kg), or saline was injected subcutaneously at time 0. Rats were observed for seizure activity for 2 h, survivors were killed 24 h later and edema was measured in samples from parietal and piriform cortices, dorsal thalmus, and hippocampus. None of the rats pretreated with saline had seizures when given doses of cholinomimetics alone. However, rats pretreated with LiCl had the following incidence of seizures: PHY 68%, DFP 71% and PIL 100%. Rats given cholinomimetic agents alone did not have brain edema. In contrast, all LiCl-pretreated rats that seized had pronounced brain edema which was greatest in the piriform cortex. Thus, these studies demonstrate that LiCl pretreatment potentiates cholinomimetic-induced seizures. Further, cholinomimetic-induced seizures produce brain changes resulting in edema. PMID:2166259 16. Results of brine flow testing and disassembly of a crushed salt/bentonite block seal at the Waste Isolation Pilot Plant SciTech Connect Finley, R.E.; Jones, R.L. 1994-03-01 The Small-Scale Seal Performance Tests, Series C, a set of in situ experiments conducted at the Waste Isolation Pilot Plant, are designed to evaluate the performance of various seal materials emplaced in large (0.9-m-diameter) boreholes. This report documents the results of fluid (brine) flow testing and water and clay content analyses performed on one emplaced seal comprised of 100% salt blocks and 50%/50% crushed salt/bentonite blocks and disassembled after nearly three years of brine injection testing. Results from the water content analyses of 212 samples taken from within this seal show uniform water content throughout the 50%/50% salt/bentonite blocks with saturations about 100%. Clay content analyses from the 100% salt endcaps of the seal show a background clay content of about 1% by weight uniformly distributed, with the exception of samples taken at the base of the seal at the borehole wall interface. These samples show clay contents up to 3% by weight, which suggests some bentonite may have migrated under pressure to that interface. Results of the brine-flow testing show that the permeability to brine for this seal was about 2 to 3 {times} 10{sup {minus}4} darcy (2 to 3 {times} 10{sup {minus}16} m{sup 2}). 17. Performance of the gas bubble column in molten salt systems SciTech Connect Sada, E.; Katoh, S.; Nakanishi, A.; Yamanishi, T.; Yoshii, H. 1984-01-01 Experimental data on the gas holdup and the mean bubble size in a bubble column with a single nozzle was obtained for gas-molten salt systems of a eutectic mixture of LiCl (58 mol %)-KCl (42 mol %) and molten NaNO/sub 3/. The liquid-phase mass transfer coefficient K /SUB L/ was evaluated from the specific surface area a and the volumetric coefficient K /SUB L/ a data for oxygen and carbon dioxide absorption into molten NaNO/sub 3/. The dimensionless correlations of the performance of bubble columns for aqueous solutions can be extended to the gas-molten salt systems. 18. Expected near-field thermal environments in a sequentially loaded spent-fuel or high-level waste repository in salt SciTech Connect Rickertsen, L.D.; Arbital, J.G.; Claiborne, H.C. 1982-01-01 This report describes the effect of realistic waste emplacement schedules on repository thermal environments. Virtually all estimates to date have been based on instantaneous loading of wastes having uniform properties throughout the repository. However, more realistic scenarios involving sequential emplacement of wastes reflect the gradual filling of the repository over its lifetime. These cases provide temperatures that can be less extreme than with the simple approximation. At isolated locations in the repository, the temperatures approach the instantaneous-loading limit. However, for most of the repository, temperature rises in the near-field are 10 to 40 years behind the conservative estimates depending on the waste type and the location in the repository. Results are presented for both spent-fuel and high-level reprocessing waste repositories in salt, for a regional repository concept, and for a single national repository concept. The national repository is filled sooner and therefore more closely approximates the instantaneously loaded repository. However, temperatures in the near-field are still 20/sup 0/C or more below the values in the simple model for 40 years after startup of repository emplacement operations. The results suggest that current repository design concepts based on the instantaneous-loading predictions are very conservative. Therefore, experiments to monitor temperatures in a test and evaluation facility, for example, will need to take into account the reduced temperatures in order to provide data used in predicting repository performance. 19. Rheology Of MonoSodium Titanate (MST) And Modified Mst (mMST) Mixtures Relevant To The Salt Waste Processing Facility SciTech Connect Koopman, D. C.; Martino, C. J.; Shehee, T. C.; Poirier, M. R. 2013-07-31 The Savannah River National Laboratory performed measurements of the rheology of suspensions and settled layers of treated material applicable to the Savannah River Site Salt Waste Processing Facility. Suspended solids mixtures included monosodium titanate (MST) or modified MST (mMST) at various solid concentrations and soluble ion concentrations with and without the inclusion of kaolin clay or simulated sludge. Layers of settled solids were MST/sludge or mMST/sludge mixtures, either with or without sorbed strontium, over a range of initial solids concentrations, soluble ion concentrations, and settling times. 20. LiCl solvation in N-methyl-acetamide (NMA) as a model for understanding Li(+) binding to an amide plane. PubMed Manin, Nikolai; da Silva, Mauricio C; Zdravkovic, Igor; Eliseeva, Olga; Dyshin, Alexey; Ya?ar, Orhan; Salahub, Dennis R; Kolker, Arkadiy M; Kiselev, Michael G; Noskov, Sergei Yu 2016-01-27 The thermodynamics of ion solvation in non-aqueous solvents remains of great significance for understanding cellular transport and ion homeostasis for the design of novel ion-selective materials and applications in molecular pharmacology. Molecular simulations play pivotal roles in connecting experimental measurements to the microscopic structures of liquids. One of the most useful and versatile mimetic systems for understanding biological ion transport is N-methyl-acetamide (NMA). A plethora of theoretical studies for ion solvation in NMA have appeared recently, but further progress is limited by two factors. One is an apparent lack of experimental data on solubility and thermodynamics of solvation for a broad panel of 1?:?1 salts over an appropriate temperature and concentration range. The second concern is more substantial and has to do with the limitations hardwired in the additive (fixed charge) approximations used for most of the existing force-fields. In this submission, we report on the experimental evaluation of LiCl solvation in NMA over a broad range of concentrations and temperatures and compare the results with those of MD simulations with several additive and one polarizable force-field (Drude). By comparing our simulations and experimental results to density functional theory computations, we discuss the limiting factors in existing potential functions. To evaluate the possible implications of explicit and implicit polarizability treatments on ion permeation across biological channels, we performed potential of mean force (PMF) computations for Li(+) transport through a model narrow ion channel with additive and polarizable force-fields. PMID:26784370 1. Enhanced tolerance to NaCl and LiCl stresses by over-expressing Caragana korshinskii sodium/proton exchanger 1 (CkNHX1) and the hydrophilic C terminus is required for the activity of CkNHX1 in Atsos3-1 mutant and yeast SciTech Connect Yang, Da-Hai; Department of Plant Physiology, Institute of General Botany and Plant Physiology, Friedrich-Schiller-University, Dornburger Strasse 159, 07743 Jena ; Song, Li-Ying; Hu, Jun; Yin, Wei-Bo; Li, Zhi-Guo; Chen, Yu-Hong; Su, Xiao-Hua; Wang, Richard R.-C.; Hu, Zan-Min 2012-01-13 Highlights: Black-Right-Pointing-Pointer CkNHX1 was isolated from Caragana korshinskii. Black-Right-Pointing-Pointer CkNHX1 was expressed mainly in roots, and significantly induced by NaCl in stems. Black-Right-Pointing-Pointer Expression of CkNHX1 enhanced the resistance to NaCl and LiCl in yeast and Atsos3-1. Black-Right-Pointing-Pointer Expression of CkNHX1-{Delta}C had little effect on NaCl/LiCl tolerance in Atsos3-1. Black-Right-Pointing-Pointer C-terminal region of CkNHX1 is required for its Na{sup +} and Li{sup +} transporting activity. -- Abstract: Sodium/proton exchangers (NHX antiporters) play important roles in plant responses to salt stress. Previous research showed that hydrophilic C-terminal region of Arabidopsis AtNHX1 negatively regulates the Na{sup +}/H{sup +} transporting activity. In this study, CkNHX1 were isolated from Caragana korshinskii, a pea shrub with high tolerance to salt, drought, and cold stresses. Transcripts of CkNHX1 were detected predominantly in roots, and were significantly induced by NaCl stress in stems. Transgenic yeast and Arabidopsisthalianasos3-1 (Atsos3-1) mutant over-expressing CkNHX1 and its hydrophilic C terminus-truncated derivative, CkNHX1-{Delta}C, were generated and subjected to NaCl and LiCl stresses. Expression of CkNHX1 significantly enhanced the resistance to NaCl and LiCl stresses in yeast and Atsos3-1 mutant. Whereas, compared with expression of CkNHX1, the expression of CkNHX1-{Delta}C had much less effect on NaCl tolerance in Atsos3-1 and LiCl tolerance in yeast and Atsos3-1. All together, these results suggest that the predominant expression of CkNHX1 in roots might contribute to keep C. korshinskii adapting to the high salt condition in this plant's living environment; CkNHX1 could recover the phenotype of Atsos3-1 mutant; and the hydrophilic C-terminal region of CkNHX1 should be required for Na{sup +}/H{sup +} and Li{sup +}/H{sup +} exchanging activity of CkNHX1. 2. Electrodialysis technology for salt recovery from aluminum salt cake SciTech Connect Hryn, J. N.; Krumdick, G.; Graziano, D.; Sreenivasarao, K. 2000-02-02 Electrodialysis technology for recovering salt from aluminum salt cake is being developed at Argonne National Laboratory. Salt cake, a slag-like aluminum-industry waste stream, contains aluminum metal, salt (NaCl and KCl), and nonmetallics (primarily aluminum oxide). Salt cake can be recycled by digesting with water and filtering to recover the metal and oxide values. A major obstacle to widespread salt cake recycling is the cost of recovering salt from the process brine. Electrodialysis technology developed at Argonne appears to be a cost-effective approach to handling the salt brines, compared to evaporation or disposal. In Argonne's technology, the salt brine is concentrated until salt crystals are precipitated in the electrodialysis stack; the crystals are recovered downstream. The technology is being evaluated on the pilot scale using Eurodia's EUR 40-76-5 stack. 3. Radioactive waste isolation in salt: peer review of the D'Appolonia report on Schematic Designs for Penetration Seals for a Repository in the Permian Basin, Texas SciTech Connect Hambley, D.F.; Stormont, J.C.; Russell, J.E.; Edgar, D.E.; Fenster, D.F.; Harrison, W.; Tisue, M.W. 1984-09-01 Argonne made the following recommedations for improving the reviewed reports. The authors of the report should: state the major assumptions of the study in Sec. 1.1 rather than later in the report; consider using salt for the shaft seals in salt horizons; reconsider whether keys are needed for the bulkheads; provide for interface grouting because use of expansive cement will not guarantee that interfaces will be impermeable; discuss the sealing schedule and, where appropriate, consider what needs to be done to ensure that emplaced radioactive waste could be retrieved if necessary; describe in more detail the sealing of the Dockum and Ogallala aquifers; consider an as low as reasonably achievable approach to performance requirements for the initial design phase; address the concerns in the 1983 US Nuclear Regulatory Commission document entitled Draft Technical Position: Borehole and Shaft Sealing of High-Level Nuclear Waste Repositories; cite the requirements for release of radioactivity by referring to specific clauses in the regulations of the US Environmental Protection Agency; and provide further explanation in the outline of future activities about materials development and verification testing. More emphasis on development of accelerated testing programs is also required. 4. Radiation damage measurements on rock salt and other minerals for waste disposal applications. Quarterly report, January 1, 1980-March 31, 1980 SciTech Connect Swyler, K J; Loman, J M; Teutonico, L J; Elgort, G E; Levy, P W 1980-04-10 Different aspects of radiation damage in both synthetic NaCl crystals and various natural rock salt samples as well as granite, basalt and other minerals which will be important for radioactive waste disposal applications are being investigated. The principal means of measuring radiation damage is the determination of F-center concentrations, and the concentration and size of sodium metal colloid particles. Formation of these and other defects during irradiation and the annealing of defects and characterization of other processes occurring after irradiation are being studied as a function of dose rate, total dose, sample temperature during irradiation, strain applied prior to and during irradiation, etc. Measurements are being made on synthetic NaCl and natural rock salt samples from different geological locations, including some potential repository sites. It will be necessary to determine if radiation damage in the minerals from different localities is similar. If non-negligible differences are observed a detailed study must be made for each locality under consideration. Almost all current studies are being made on rock salt but other minerals particularly granite and basalt are being phased into the program. It is now established that radiation damage formation in both natural and synthetic rock salt is strongly dependent on strain. The strain related effects strongly indicate that the damage formation processes and in particular the colloid nucleation processes are related to the strain induced disolcations. A temporary theoretical effort has been started to determine which dislocation related effects are important for radiation damage processes and, most importantly, what dislocation interactions are most likely to create nucleation sites for colloid particles. If these preliminary studies indicate that additional theoretical studies will be useful an effort will be made to have them extended. 5. Purification of used eutectic (LiCl-KCl) salt electrolyte from pyroprocessing NASA Astrophysics Data System (ADS) Cho, Yung-Zun; Lee, Tae-Kyo; Eun, Hee-Chul; Choi, Jung-Hoon; Kim, In-Tae; Park, Geun-Il 2013-06-01 The separation characteristics of surrogate rare-earth fission products in a eutectic (LiCl-KCl) molten salt were investigated. This system is based on the eutectic salt used for the pyroprocessing treatment of used nuclear fuel (UNF). The investigation was performed using an integrated rare-earth separation apparatus comprising a precipitation reactor, a solid detachment device, and a layer separation device. To separate rare-earth fission products, a phosphate precipitation method using both Li3PO4 and K3PO4 as a precipitant was performed. The use of an equivalent phosphate precipitant composed of 0.408 molar ratio-K3PO4 and 0.592 molar ratio-Li3PO4 can preserve the original eutectic ratio, LiCl-0.592 molar ratio (or 45.2 wt%), as well as provide a high separation efficiency of over 99.5% under conditions of 550 C and Ar sparging when using La, Nd, Ce, and Pr chlorides. The mixture of La, Nd, Ce, and Pr phosphate had a typical monoclinic (or monazite) structure, which has been proposed as a reliable host matrix for the permanent disposal of a high-level waste form. To maximize the reusability of purified eutectic waste salt after rare-earth separation, the successive rare-earth separation process, which uses both phosphate precipitation and an oxygen sparging method, were introduced and tested with eight rare-earth (Y, La, Ce, Pr, Nd, Sm, Eu and Gd) chlorides. In the successive rare-earth separation process, the phosphate reaction was terminated within 1 h at 550 C, and a 4-8 h oxygen sparging time were required to obtain over a 99% separation efficiency at 700-750 C. The mixture of rare-earth precipitates separated by the successive rare-earth separation process was found to be phosphate, oxychloride, and oxide. Through the successive rare-earth separation process, the eutectic ratio of purified salt maintained its original value, and impurity content including the residual precipitant of purified salt can be minimized. 6. Influence of soil properties on trace element availability and plant accumulation in a Mediterranean salt marsh polluted by mining wastes: implications for phytomanagement. PubMed Conesa, H M; María-Cervantes, A; Alvarez-Rogel, J; González-Alcaraz, M N 2011-09-15 The aims of this study were to determine the factors which control metal and As phytoavailability in the different microenvironments (Sand Dunes, Salt Flat, Dry River and Shrubs) present at a Mediterranean salt marsh polluted by mining wastes. We performed a field study following a plot sampling survey. The analyses of soil parameters (pH, electrical conductivity (EC), organic carbon contents, etc.), total metal and As concentrations and their phytoavailability assessed with EDTA were related to each microenvironment and the corresponding plant species uptake. The averages of pH and EC were slightly alkaline (pH ≈ 7.5) and saline (≈ 2.2 to 17.1 dS m(-1)) respectively. The soil samples from the Salt Flat subzone showed the highest metal concentrations (e.g. 51 mg kg(-1) Cd, 11,600 mg kg(-1) Pb) while for As, the highest concentrations occurred in the Dry River (380 mg kg(-1) As). The total metal and EDTA-extractable concentrations occurred as it follows: Salt Flat>Dry River>Degraded Dunes>Shrubs. In relation to plant metal and As accumulation, the highest root concentrations were obtained in the species from the Salt Flat subzone: ~17 mg kg(-1) As, ~620 mg kg(-1) Pb, for both, Juncus maritimus and Arthrocnemum macrostachyum. However the highest metal and As shoot concentrations occurred in species from the Sand Dunes: ~23 mg kg(-1) As ~270 mg kg(-1) Pb for Dittrichia viscosa; ~23 mg kg(-1) As, ~390 mg kg(-1) Zn for Crucianella maritima. The occurrence of edaphic gradients including salinity and texture determined the vegetation distribution. However, it cannot be concluded that there was a disturbance due to metal(loid)s soil concentrations in terms of vegetation composition except in the Degraded Dunes and Dry River. The higher EDTA-extractable concentrations were coincidental with the most saline soils but this did not result in higher metal(loid)s plant accumulation. PMID:21851964 7. Summary of computational support and general documentation for computer code (GENTREE) used in Office of Nuclear Waste Isolation Pilot Salt Site Selection Project SciTech Connect Beatty, J.A.; Younker, J.L.; Rousseau, W.F.; Elayat, H.A. 1983-12-06 A Decision Tree Computer Model was adapted for the purposes of a Pilot Salt Site Selection Project conducted by the Office of Nuclear Waste Isolation (ONWI). A deterministic computer model was developed to structure the site selection problem with submodels reflecting the five major outcome categories (Cost, Safety, Delay, Environment, Community Impact) to be evaluated in the decision process. Time-saving modifications were made in the tree code as part of the effort. In addition, format changes allowed retention of information items which are valuable in directing future research and in isolation of key variabilities in the Site Selection Decision Model. The deterministic code was linked to the modified tree code and the entire program was transferred to the ONWI-VAX computer for future use by the ONWI project. 8. Role of cationic size in the optical properties of the LiCl crystal surface: theoretical study. PubMed Abdel Halim, Wael Salah; Abdullah, Noha; Abdel-Aal, Safaa; Shalabi, A S 2012-06-01 The size of the cations (either Ca(2+), Sr(2+), Ga(+), or Au(+)) at the F(A1)-type color centers on the (100) surface of LiCl crystal plays an important role in the optical properties of this surface. In this work, double-well potentials at this surface were investigated using ab initio quantum mechanical methods. Quantum clusters were embedded in simulated Coulomb fields that closely approximate the Madelung fields of the host surface, and the ions that were the nearest neighbors to the F(A1) site were allowed to relax to equilibrium. The calculated Stokes-shifted optical transition bands, optical-optical conversion efficiency, and relaxed excited states of the defect-containing surface, as well as the orientational destruction of the color centers, recording sensitivity, exciton (energy) transfer, and the Glasner-Tompkins empirical relation were all found to be sensitive to the size of the dopant cation. PMID:22033757 9. Ab initio MRSDCI study on the low-lying electronic states of the lithium chloride molecule (LiCl). PubMed Kurosaki, Yuzuru; Yokoyama, Keiichi 2012-08-14 Potential energy curves (PECs) for the low-lying states of the lithium chloride molecule (LiCl) have been calculated using the internally contracted multireference single- and double-excitation configuration interaction (MRSDCI) method with the aug-cc-PVnZ (AVnZ) and aug-cc-PCVnZ (ACVnZ) basis sets, where n = T, Q, and 5. First, we calculate PECs for 7 spin-orbit (SO)-free ?-S states, X(1)?(+), A(1)?(+), (3)?(+), (1)?, and (3)?, and then obtain PECs for 13 SO ? states, X0(+), A0(+), B0(+), 0(-)(I), 0(-)(II), 1(I), 1(II), 1(III), and 2, by diagonalizing the matrix of the electronic Hamiltonian plus the Breit-Pauli SO Hamiltonian. The MRSDCI calculations not including core orbital correlation through the single and double excitations are also performed with the AV5Z and ACV5Z basis sets. The Davidson corrections (Q0) are added to both the ?-S and ? state energies. Vibrational eigenstates for the obtained X(1)?(+) and X0(+) PECs are calculated by solving the time-independent Schro?dinger equation with the grid method. Thus, the effects of basis set, core orbital correlation, and the Davidson correction on the X(1)?(+) and X0(+) PECs of LiCl are investigated by comparing the spectroscopic constants calculated from the PECs with one another and with experiment. It is confirmed that to accurately predict the spectroscopic constants we need to include core-electron correlation in the CI expansion and use the basis sets designed to describe core-valence correlation, i.e., ACVnZ. The SO PECs presented in this paper will be of help in the future study of diatomic alkali halide dynamics. PMID:22897271 10. Actinide removal from spent salts DOEpatents Hsu, Peter C. (Pleasanton, CA); von Holtz, Erica H. (Livermore, CA); Hipple, David L. (Livermore, CA); Summers, Leslie J. (Livermore, CA); Adamson, Martyn G. (Danville, CA) 2002-01-01 A method for removing actinide contaminants (uranium and thorium) from the spent salt of a molten salt oxidation (MSO) reactor is described. Spent salt is removed from the reactor and analyzed to determine the contaminants present and the carbonate concentration. The salt is dissolved in water, and one or more reagents are added to precipitate the thorium as thorium oxide and/or the uranium as either uranium oxide or as a diuranate salt. The precipitated materials are filtered, dried and packaged for disposal as radioactive waste. About 90% of the thorium and/or uranium present is removed by filtration. After filtration, salt solutions having a carbonate concentration >20% can be dried and returned to the reactor for re-use. Salt solutions containing a carbonate concentration <20% require further clean-up using an ion exchange column, which yields salt solutions that contain less than 0.1 ppm of thorium or uranium. 11. Metals removal from spent salts DOEpatents Hsu, Peter C.; Von Holtz, Erica H.; Hipple, David L.; Summers, Leslie J.; Brummond, William A.; Adamson, Martyn G. 2002-01-01 A method and apparatus for removing metal contaminants from the spent salt of a molten salt oxidation (MSO) reactor is described. Spent salt is removed from the reactor and analyzed to determine the contaminants present and the carbonate concentration. The salt is dissolved in water, and one or more reagents may be added to precipitate the metal oxide and/or the metal as either metal oxide, metal hydroxide, or as a salt. The precipitated materials are filtered, dried and packaged for disposal as waste or can be immobilized as ceramic pellets. More than about 90% of the metals and mineral residues (ashes) present are removed by filtration. After filtration, salt solutions having a carbonate concentration >20% can be spray-dried and returned to the reactor for re-use. Salt solutions containing a carbonate concentration <20% require further clean-up using an ion exchange column, which yields salt solutions that contain less than 1.0 ppm of contaminants. 12. Consolidation, permeability, and strength of crushed salt/bentonite mixtures with application to the WIPP (Waste Isolation Pilot Plant) SciTech Connect Pfeifle, T.W. ) 1991-01-01 Three tests were performed to measure the consolidation, permeability, and compressive strength of specimens prepared from bentonite/crushed salt mixtures. Each mixture comprised 30% bentonite and 70% crushed salt based on total dry weight. Brine was added to each mixture to adjust its water content to either 5 or 10% (nominal) of the total dry weight of the mixture. In the consolidation tests, each specimen was subjected to multiple stages of successively higher hydrostatic stress (pressure). During each stage, the pressure was maintained at a constant level and volumetric strain data were continuously logged. By using multiple stages, consolidation data were obtained at several pressures and the time required to consolidate the specimens to full saturation was reduced. Once full saturation was achieved, each specimen was subjected to a final test stage in which the hydrostatic stress was reduced and a permeability test performed. Permeability was measured using the steady flow of brine and was found to range between 1 {times} 10{sup {minus}17} and 5 {times} 10{sup {minus}17} m{sup 2}. After the final test stage, unconfined compressive strength was determined for each specimen and was found to range between 0.5 and 8.1 MPa. Two constitutive models were fitted to the consolidation data. One relatively simple model related volumetric strain to time while the other related instantaneous density to time, pressure, and initial density. 8 refs., 9 figs., 8 tabs. 13. Radioactive waste isolation in salt: peer review of the Office of Nuclear Waste Isolation's report on the Organic Geochemistry of Deep Groundwaters from the Palo Duro Basin, Texas SciTech Connect Fenster, D.F.; Brookins, D.G.; Harrison, W.; Seitz, M.G.; Lerman, A.; Stamoudis, V.C. 1984-08-01 This report summarizes Argonne's review of the Office of Nuclear Waste Isolation's (ONWI's) final report entitled The Organic Geochemistry of Deep Ground Waters from the Palo Duro Basin, Texas, dated September 1983. Recommendations are made for improving the ONWI report. The main recommendation is to make the text consistent with the title and with the objective of the project as stated in the introduction. Three alternatives are suggested to accomplish this. 14. Investigation of residual anode material after electrorefining uranium in molten chloride salt NASA Astrophysics Data System (ADS) Rose, M. A.; Williamson, M. A.; Willit, J. 2015-12-01 A buildup of material at uranium anodes during uranium electrorefining in molten chloride salts has been observed. Potentiodynamic testing has been conducted using a three electrode cell, with a uranium working electrode in both LiCl/KCl eutectic and LiCl each containing ∼5 mol% UCl3. The anodic current response was observed at 50° intervals between 450 °C and 650 °C in the eutectic salt. These tests revealed a buildup of material at the anode in LiCl/KCl salt, which was sampled at room temperature, and analyzed using ICP-MS, XRD and SEM techniques. Examination of the analytical data, current response curves and published phase diagrams has established that as the uranium anode dissolves, the U3+ ion concentration in the diffusion layer surrounding the electrode rises precipitously to levels, which may at low temperatures exceed the solubility limit for UCl3 or in the case of the eutectic salt for K2UCl5. The reduction in current response observed at low temperature in eutectic salt is eliminated at 650 °C, where K2UCl5 is absent due to its congruent melting and only simple concentration polarization effects are seen. In LiCl similar concentration effects are seen though significantly longer time at applied potential is required to effect a reduction in the current response as compared to the eutectic salt. 15. Salt splitting using ceramic membranes SciTech Connect Kurath, D.E.; Hollenberg, G.W.; Jue, Jan-Fong 1997-01-01 Inorganic ceramic membranes for salt splitting of radioactively contaminated sodium salt solutions are being developed for treating US Department of Energy tank wastes. The process consists of electrochemical separation of sodium ions from the salt solution using sodium (Na) Super Ion Conductors (NaSICON) membranes. In contrast to conventional organic-based bipolar or ion exchange membranes used in salt splitting, NaSICON membranes are resistant to gamma/beta radiation and are highly selective for sodium ions. Potential applications include (1) caustic recycle for sludge leaching, regeneration of ion exchange resins, inhibition of corrosion in carbon steel tanks, or retrieval of tank wastes; (2) pH adjustment and reduction of competing cations to enhance cesium ion exchange processes; (3) sodium reduction in high-level waste sludges; and (4) sodium removal from acidic wastes to facilitate calcining. Initial experiments with dysprosium-based NaSICON membranes have demonstrated the feasibility of the process. 16. Recycling of aluminum salt cake SciTech Connect Jody, B.J.; Daniels, E.J.; Bonsignore, P.V.; Karvelas, D.E. 1991-12-01 The secondary aluminum industry generates more than 110 {times} 10{sup 3} tons of salt-cake waste every year. This waste stream contains about 3--5% aluminum, 15--30% aluminum oxide, 30--40% sodium chloride, and 20--30% potassium chloride. As much as 50% of the content of this waste is combined salt (sodium and potassium chlorides). Salt-cake waste is currently disposed of in conventional landfills. In addition, over 50 {times} 10{sup 3} tons of black dross that is not economical to reprocess a rotary furnace for aluminum recovery ends up in landfills. The composition of the dross is similar to that of salt cake, except that it contains higher concentrations of aluminum (up to 20%) and correspondingly lower amounts of salts. Because of the high solubility of the salts in water, these residues, when put in landfills, represent a potential source of pollution to surface-water and groundwater supplies. The increasing number of environmental regulations on the generation and disposal of industrial wastes are likely to restrict the disposal of these salt-containing wastes in conventional landfills. Processes exist that employ the dissolution and recovery of the salts from the waste stream. These wet-processing methods are economical only when the aluminum concentration in that waste exceeds about 10%. Argonne National Laboratory (ANL) conducted a study in which existing technologies were reviewed and new concepts that are potentially more cost-effective than existing processes were developed and evaluated. These include freeze crystallization, solvent/antisolvent extraction, common-ion effect, high-pressure/high-temperature process, and capillary-effect systems. This paper presents some of the technical and economic results of the aforementioned ANL study. 17. Heavy metals, salts and organic residues in old solid urban waste landfills and surface waters in their discharge areas: determinants for restoring their impact. PubMed Pastor, J; Hernández, A J 2012-03-01 This study was designed to determine the state of polluted soils in the main landfills of the Community of Madrid (central Spain), as part of a continuous assessment of the impacts of urban solid waste (USW) landfills that were capped with a layer of soil 20 years ago. Our analysis of this problem has been highly conditioned by the constant re-use of many of the USW landfills, since they have never been the target of any specific restoration plan. Our periodical analysis of cover soils and soils from discharge areas of the landfills indicates soil pollution has worsened over the years. Here, we examined heavy metal, salts, and organic compounds in soil and surface water samples taken from 15 landfills in the Madrid region. Impacts of the landfill soil covers on nematode and plant diversity were also evaluated. These analyses continue to reveal the presence of heavy metals (Zn, Cu, Cr, Ni, Pb, Cd) in soils, and salts (sulphates, chlorides and nitrates) in soils and surface waters. In addition, non-agricultural organic compounds, mainly aromatic and aliphatic hydrocarbons, often appeared in very high concentrations, and high levels of insecticides such as gamma-HCH (lindane) were also detected in soils. Around 50% of the water samples collected showed chemical demand of oxygen (CDO) values in excess of 150 mg/l. Traces of phenolic compounds were detected in some landfills, some of which exhibited high levels of 2-chlorophenol and pentachlorophenol. All these factors are conditioning both the revegetation of the landfill systems and the remediation of their slopes and terrestrial ecosystems arising in their discharge areas. This work updates the current situation and discusses risks for the health of the ecosystems, humans, domestic animals and wildlife living close to these landfills. PMID:21764209 18. [Bio-oil production from biomass pyrolysis in molten salt]. PubMed Ji, Dengxiang; Cai, Tengyue; Ai, Ning; Yu, Fengwen; Jiang, Hongtao; Ji, Jianbing 2011-03-01 In order to investigate the effects of pyrolysis conditions on bio-oil production from biomass in molten salt, experiments of biomass pyrolysis were carried out in a self-designed reactor in which the molten salt ZnCl2-KCl (with mole ratio 7/6) was selected as heat carrier, catalyst and dispersion agent. The effects of metal salt added into ZnCl2-KCl and biomass material on biomass pyrolysis were discussed, and the main compositions of bio-oil were determined by GC-MS. Metal salt added into molten salt could affect pyrolysis production yields remarkably. Lanthanon salt could enhance bio-oil yield and decrease water content in bio-oil, when mole fraction of 5.0% LaCl3 was added, bio-oil yield could reach up to 32.0%, and water content of bio-oil could reduce to 61.5%. The bio-oil and char yields were higher when rice straw was pyrolysed, while gas yield was higher when rice husk was used. Metal salts showed great selectivity on compositions of bio-oil. LiCl and FeCl2 promoted biomass to pyrolyse into smaller molecular weight compounds. CrCl3, CaCl2 and LaCl3 could restrain second pyrolysis of bio-oil. The research provided a scientific reference for production of bio-oil from biomass pyrolysis in molten salt. PMID:21650030 19. Preliminary petrological and geochemical results from the Salton Sea Geothermal Field, California: A near-field natural analog of a radioactive waste repository in salt: Topical report No. 2 SciTech Connect Elders, W.A.; Cohen, L.H.; Williams, A.E.; Neville, S.; Collier, P.; Oakes, C. 1986-03-01 High concentrations of radionuclides and high temperatures are not naturally encountered in salt beds. For this reason, the Salton Sea Geothermal Field (SSGF) may be the best available geologic analog of some of the processes expected to occur in high level nuclear waste repositories in salt. Subsurface temperatures and brine concentrations in the SSGF span most of the temperature range and fluid inclusion brine range expected in a salt repository, and the clay-rich sedimentary rocks are similar to those which host bedded or domal salts. As many of the chemical processes observed in the SSGF are similar to those expected to occur in or near a salt repository, data derived from it can be used in the validation of geochemical models of the near-field of a repository in salt. This report describes preliminary data on petrology and geochemistry, emphasizing the distribution of rare earth elements and U and Th, of cores and cuttings from several deep wells chosen to span a range of temperature gradients and salinities. Subsurface temperature logs have been augmented by fluid inclusion studies, to reveal the effects of brines of varying temperature and salinity. The presence of brines with different oxygen isotopic signatures also indicate lack of mixing. Whole rock major, minor and trace element analyses and data on brine compositions are being used to study chemical migration in these sediments. 65 refs., 20 figs., 3 tabs. 20. Modeling and Field Test Planning Activities in Support of Disposal of Heat-Generating Waste in Salt SciTech Connect Rutqvist, Jonny; Blanco Martin, Laura; Mukhopadhyay, Sumit; Houseworth, Jim; Birkholzer, Jens 2014-09-26 The modeling efforts in support of the field test planning conducted at LBNL leverage on recent developments of tools for modeling coupled thermal-hydrological-mechanical-chemical (THMC) processes in salt and their effect on brine migration at high temperatures. This work includes development related to, and implementation of, essential capabilities, as well as testing the model against relevant information and published experimental data related to the fate and transport of water. These are modeling capabilities that will be suitable for assisting in the design of field experiment, especially related to multiphase flow processes coupled with mechanical deformations, at high temperature. In this report, we first examine previous generic repository modeling results, focusing on the first 20 years to investigate the expected evolution of the different processes that could be monitored in a full-scale heater experiment, and then present new results from ongoing modeling of the Thermal Simulation for Drift Emplacement (TSDE) experiment, a heater experiment on the in-drift emplacement concept at the Asse Mine, Germany, and provide an update on the ongoing model developments for modeling brine migration. LBNL also supported field test planning activities via contributions to and technical review of framework documents and test plans, as well as participation in workshops associated with field test planning. 1. Crushed Salt Constitutive Model SciTech Connect Callahan, G.D. 1999-02-01 The constitutive model used to describe the deformation of crushed salt is presented in this report. Two mechanisms -- dislocation creep and grain boundary diffusional pressure solution -- are combined to form the basis for the constitutive model governing the deformation of crushed salt. The constitutive model is generalized to represent three-dimensional states of stress. Upon complete consolidation, the crushed-salt model reproduces the Multimechanism Deformation (M-D) model typically used for the Waste Isolation Pilot Plant (WIPP) host geological formation salt. New shear consolidation tests are combined with an existing database that includes hydrostatic consolidation and shear consolidation tests conducted on WIPP and southeastern New Mexico salt. Nonlinear least-squares model fitting to the database produced two sets of material parameter values for the model -- one for the shear consolidation tests and one for a combination of the shear and hydrostatic consolidation tests. Using the parameter values determined from the fitted database, the constitutive model is validated against constant strain-rate tests. Shaft seal problems are analyzed to demonstrate model-predicted consolidation of the shaft seal crushed-salt component. Based on the fitting statistics, the ability of the model to predict the test data, and the ability of the model to predict load paths and test data outside of the fitted database, the model appears to capture the creep consolidation behavior of crushed salt reasonably well. 2. Effects of high concentration of salts on the esterase activity and structure of a kiwifruit peptidase, actinidain. PubMed Morimoto, Koichi; Furuta, Erino; Hashimoto, Hiroshi; Inouye, Kuniyo 2006-06-01 Effects of salts on the activity and stability of actinidain were examined. With increasing salt concentration up to 0.5 M, the activity (kcat/Km) for N-alpha-Cbz-L-lysine p-nitrophenyl ester decreased to 40% of that in the absence of salt. The inhibitor constant Ki of LiCl, NaCl, and KCl was 0.16-0.43 M. With 3 M KCl and NaCl, the specificity constant kcat/Km recovered to 110 and 75%, respectively. No re-activation was observed with LiCl. The inhibition and re-activation were dependent on the changes in both Km and kcat, whereas no CD change was observed. The tryptophan fluorescence of actinidain was not affected by 0-0.5 M salt, but a considerable decrease in its intensity was observed with increasing salt concentration from 0.5 to 3.0 M. These results suggest that the inhibition observed with the lower salt concentration (<0.5 M) is due to attenuation of the electrostatic interaction between the enzyme and substrate, and the higher concentration (0.5-3.0 M) induces structural change in the states of tryptophan residues, which is associated with the re-activation. Actinidain keeps considerably high activity and stability even in the presence of 3 M salts. PMID:16788057 3. Studies of the plasticizing effect of different hydrophilic inorganic salts on starch/poly (vinyl alcohol) films. PubMed Jiang, Xiancai; Li, Hongmei; Luo, Yong; Zhao, Yulai; Hou, Linxi 2016-01-01 The effects of different inorganic salts LiCl, MgCl26H2O, CaCl2, and AlCl36H2O on the crystalline, thermal, water vapor barrier, and tensile properties of starch/PVA films were studied. The high plasticizing efficiency of all these four inorganic salts for starch/PVA film was confirmed by the obtained results. These four salts all had a good compatibility with starch/PVA within the content of 15 wt% and starch/PVA became completely miscible with the addition of 15 wt% inorganic salts. All these four salts had a strong destroying effect on the crystals of starch and PVA. Among these four salts, AlCl36H2O had the largest negative effect on the thermal stability of starch/PVA and LiCl had the largest improving effect on the water sorption rate of starch/PVA film. On the whole MgCl26H2O and CaCl2 were the more suitable plasticizer for starch/PVA film among these four inorganic salts. With the addition of 15 wt% MgCl26H2O and CaCl2, the elongation at break of starch/PVA film could reach to 418.83% and 434.80%, respectively. PMID:26608006 4. Variations of organic matters and microbial community in thermophilic anaerobic digestion of waste activated sludge with the addition of ferric salts. PubMed Yu, Bao; Lou, Ziyang; Zhang, Dongling; Shan, Aidang; Yuan, Haiping; Zhu, Nanwen; Zhang, Kanghan 2015-03-01 Ferric salts will influence the thermophilic anaerobic digestion of waste activated sludge (WAS). FeCl3 was found to contribute to the anaerobic digestion process with a cumulative biogas production of 357 mL/gVS, 79.6% higher than that in the control group, and Fe2(SO4)3 had no distinct impact, while Fe(NO3)3 inhibited the methanogenesis process. A favorable balance between the release of organic matters from WAS and consumption rate was established after dosing FeCl3 from the perspective of variations of soluble COD, volatile fatty acids (VFAs) and the dissolved organic matters (DOM) assessed by EEM fluorescence spectroscopy and fluorescence regional integration (FRI) technique. Conversely, the system with Fe(NO3)3 achieved an unsuitable substrates environment. Pyrosequencing revealed that the anaerobic digestion system with FeCl3 enriched Coprothermobacter for proteins fermentation and Methanosarcina for methanogenesis with the values of 18.7% and 63.2%, respectively, while that with the supplementation of Fe(NO3)3 obtained the lowest relative abundance. PMID:25545098 5. Mitigation of the inhibitory effect of soap by magnesium salt treatment of crude glycerol--a novel approach for enhanced biohydrogen production from the biodiesel industry waste. PubMed Sarma, Saurabh Jyoti; Brar, Satinder Kaur; Le Bihan, Yann; Buelna, Gerardo; Soccol, Carlos Ricardo 2014-01-01 Owing to its inhibitory effect on microbial growth, soap present in crude glycerol (CG) is a concern in biological valorization of the biodiesel manufacturing waste. By salting out strategy, up to 42% of the soap has been removed and the approach has beneficial effect on H2 production; however, removal of more than 7% of the soap was found to be inhibitory. Actually, soap is utilized as a co-substrate and due to removal; the carbon-nitrogen ratio of the medium might have decreased to reduce the production. Alternatively, without changing the carbon-nitrogen ratio of CG, MgSO4 treatment can convert the soap to its inactive form (scum). The approach was found to increase the H2 production rate (33.82%), cumulative H2 production (34.70%) as well as glycerol utilization (nearly 2.5-folds). Additionally, the treatment can increase the Mg (a nutrient) content of the medium from 0.57 ppm to 201.92 ppm. PMID:24189384 6. Increased Lifetime for Biomass and Waste to Energy Power Plant Boilers with HVOF Coatings: High Temperature Corrosion Testing Under Chlorine-Containing Molten Salt NASA Astrophysics Data System (ADS) Oksa, Maria; Tuurna, Satu; Varis, Tommi 2013-06-01 Heat exchanger surfaces of waste to energy and biomass power plant boilers experience often severe corrosion due to very aggressive components in the used fuels. High velocity oxy-fuel (HVOF) coatings offer excellent protection for boiler tubes against high temperature corrosion due to their high density and good adherence to the substrate material. Several thermal spray coatings with high chromium content were sprayed with HVOF technique. Their mechanical properties and high temperature corrosion resistance were tested and analyzed. The coating materials included NiCr, IN625, Ni-21Cr-10W-9Mo-4Cu, and iron-based partly amorphous alloy SHS9172 (Fe-25Cr-15W-12Nb-6Mo). High temperature corrosion testing was performed in NaCl-KCl-Na2SO4 salt with controlled H2O atmosphere at 575 and 625 °C. The corrosion test results of the coatings were compared to corrosion resistance of tube materials (X20, Alloy 263 and Sanicro 25). 7. Small-Column Ion-Exchange Alternative to Remove 137Cs from Low-Curie Salt Waste: Summary of Phase 1 SciTech Connect Walker, JR.,J.F. 2004-05-12 A Small-Column Ion-Exchange (SCIX) system is being evaluated for removing cesium from the Type 2 and/or Type 3 dissolved saltcake wastes at the Savannah River Site (SRS) to ensure that the dissolved saltcake meets the waste acceptance criteria at the Saltstone Facility. Both crystalline silicotitanate (CST) and IONSIV{trademark} IE-96 zeolite were evaluated as the ion-exchange media. The accelerated alternative, using CST in the SCIX, could save as much as$3 billion in operating and storage costs and {approx}20 years in processing time compared to the current baseline. With its proven high cesium-loading capacity for the expected dissolved saltcake compositions and temperatures, CST is the preferred sorbent for SCIX. The low-cost alternative sorbent, zeolite, greatly increases the volume of sorbent required because of its much lower cesium-loading capacity. Thus, zeolite greatly increases the cost for the alternative, mainly because of the increased number of Defense Waste Processing Facility canisters required to dispose of the loaded sorbent (potentially over 7000 for zeolite, compared with <500 for CST). The models previously developed for predicting cesium loading on CST compared favorably with laboratory measurements of equilibrium distribution ratios and column loading performance using dissolved saltcake simulants. These models predict that a column of 432 gal of CST can operate at 25 gal/min and treat 100,000 to 900,000 gal of dissolved saltcake, depending on the solution composition. An average value of 300,000 gal per column was used for the cost benefit analysis. Predicted cesium loading on the CST is normally below 300 Ci/L; however, solutions with low salt concentrations could potentially load the CST to 630 Ci/L. Heat transfer calculations predict nonboiling temperatures for the small columns with loadings <100 Ci/L with only natural convection cooling. For the loadings up to the maximum calculated for the tank farm (630 Ci/L), a water cooling system is required to ensure that no boiling occurs in the column if the process flow is stopped. Dose rate calculations indicate that the maximum dose rate above the tank riser is expected to be {approx} 10{sup -2} mrem/h for a column loaded at 300 Ci/L in the riser. The risk analysis indicates a net beneficial impact with no major problems likely to prevent implementation or completion of saltcake treatment. 8. Pseudo-binary electrolyte, LiBH4-LiCl, for bulk-type all-solid-state lithium-sulfur battery. PubMed Unemoto, Atsushi; Chen, ChunLin; Wang, Zhongchang; Matsuo, Motoaki; Ikeshoji, Tamio; Orimo, Shin-Ichi 2015-01-26 The ionic conduction and electrochemical and thermal stabilities of the LiBH4-LiCl solid-state electrolyte were investigated for use in bulk-type all-solid-state lithium-sulfur batteries. The LiBH4-LiCl solid-state electrolyte exhibiting a lithium ionic conductivity of [Formula: see text] at 373 K, forms a reversible interface with a lithium metal electrode and has a wide electrochemical potential window up to 5 V. By means of the high-energy mechanical ball-milling technique, we prepared a composite powder consisting of elemental sulfur and mixed conductive additive, i.e., Ketjen black and Maxsorb. In that composite powder, homogeneous dispersion of the materials is achieved on a nanometer scale, and thereby a high concentration of the interface among them is induced. Such nanometer-scale dispersals of both elemental sulfur and carbon materials play an important role in enhancing the electrochemical reaction of elemental sulfur. The highly deformable LiBH4-LiCl electrolyte assists in the formation of a high concentration of tight interfaces with the sulfur-carbon composite powder. The LiBH4-LiCl electrolyte also allows the formation of the interface between the positive electrode and the electrolyte layers, and thus the Li-ion transport paths are established at that interface. As a result, our battery exhibits high discharge capacities of 1377, 856, and 636 mAh g(-1) for the 1st, 2nd, and 5th discharges, respectively, at 373 K. These results imply that complex hydride-based solid-state electrolytes that contain Cl-ions in the crystal would be integrated into rechargeable batteries. PMID:26041380 9. Pseudo-binary electrolyte, LiBH4-LiCl, for bulk-type all-solid-state lithium-sulfur battery NASA Astrophysics Data System (ADS) Unemoto, Atsushi; Chen, ChunLin; Wang, Zhongchang; Matsuo, Motoaki; Ikeshoji, Tamio; Orimo, Shin-ichi 2015-06-01 The ionic conduction and electrochemical and thermal stabilities of the LiBH4-LiCl solid-state electrolyte were investigated for use in bulk-type all-solid-state lithium-sulfur batteries. The LiBH4-LiCl solid-state electrolyte exhibiting a lithium ionic conductivity of log ?ft( ? /S c{{m}-1} \\right)=-3.3 at 373 K, forms a reversible interface with a lithium metal electrode and has a wide electrochemical potential window up to 5 V. By means of the high-energy mechanical ball-milling technique, we prepared a composite powder consisting of elemental sulfur and mixed conductive additive, i.e., Ketjen black and Maxsorb. In that composite powder, homogeneous dispersion of the materials is achieved on a nanometer scale, and thereby a high concentration of the interface among them is induced. Such nanometer-scale dispersals of both elemental sulfur and carbon materials play an important role in enhancing the electrochemical reaction of elemental sulfur. The highly deformable LiBH4-LiCl electrolyte assists in the formation of a high concentration of tight interfaces with the sulfur-carbon composite powder. The LiBH4-LiCl electrolyte also allows the formation of the interface between the positive electrode and the electrolyte layers, and thus the Li-ion transport paths are established at that interface. As a result, our battery exhibits high discharge capacities of 1377, 856, and 636 mAh g-1 for the 1st, 2nd, and 5th discharges, respectively, at 373 K. These results imply that complex hydride-based solid-state electrolytes that contain Cl-ions in the crystal would be integrated into rechargeable batteries. 10. Application of lithium in molten-salt reduction processes. SciTech Connect Gourishankar, K. V. 1998-11-11 Metallothermic reductions have been extensively studied in the field of extractive metallurgy. At Argonne National Laboratory (ANL), we have developed a molten-salt based reduction process using lithium. This process was originally developed to reduce actinide oxides present in spent nuclear fuel. Preliminary thermodynamic considerations indicate that this process has the potential to be adapted for the extraction of other metals. The reduction is carried out at 650 C in a molten-salt (LiCl) medium. Lithium oxide (Li{sub 2}O), produced during the reduction of the actinide oxides, dissolves in the molten salt. At the end of the reduction step, the lithium is regenerated from the salt by an electrowinning process. The lithium and the salt from the electrowinning are then reused for reduction of the next batch of oxide fuel. The process cycle has been successfully demonstrated on an engineering scale in a specially designed pyroprocessing facility. This paper discusses the applicability of lithium in molten-salt reduction processes with specific reference to our process. Results are presented from our work on actinide oxides to highlight the role of lithium and its effect on process variables in these molten-salt based reduction processes. 11. Combination of PPT with LiCl Treatment Prevented Bilateral Ovariectomy-Induced Hippocampal-Dependent Cognition Deficit in Rats. PubMed Qu, Na; Zhou, Xiang-Yu; Han, Li; Wang, Lei; Xu, Jia-Xin; Zhang, Teng; Chu, Jiang; Chen, Qiao; Wang, Jian-Zhi; Zhang, Qi; Tian, Qing 2016-03-01 Estrogen deprivation is a high risk of cognitive dysfunction in neurodegenerative diseases, and the early used estrogen replacement has been proved effective in many studies. Because of the adverse actions, selective estrogen receptor modulating has been raised to substitute for estrogen replacement. In this study, we observed in hippocampus of bilaterally ovariectomized rats that the level of estrogen receptor ? (ER?) was decreased in nuclei with activated glycogen synthase kinase-3? (GSK-3?) in cytoplasm at 8weeks after operation. The level of nuclear ER? is important for its transcriptional property, and the inhibition of GSK-3? benefits to ER? nuclear translocation. Then, we used 4,4k,4a-(4-propyl-[1H]-pyrazole-1, 3, 5-triyl) trisphenol (PPT) (1mg/kg/day), an agonist of ER?, combined with LiCl (40mg/kg/day), an inhibitor of GSK-3?, to treat the ovariectomized rats. After the combination treatment of these two drugs (PPT?+?LiCl), the improved learning and memory abilities of ovariectomized rats in Morris water maze, increased dendritic spines in CA1 region, and decreased tau phosphorylation at Ser-396 in hippocampus were observed. Furthermore, PPT?+?LiCl treatment significantly increased ER? level in the nuclear fraction of hippocampus, and in the cytoplasmic fraction, the total level of GSK-3? was declined after treatment with its increased phosphorylation at Ser-9 (inactivation form). This study suggested that PPT?+?LiCl treatment could inhibit the activation of cytoplasmic GSK-3? and promote the nuclear translocation of ER?, and ER? together with GSK-3? maybe the targets to preserve hippocampus-dependent cognitive ability after long-term ovariectomy. PMID:25535091 12. Oxygen isotope activities and concentrations in aqueous salt solutions at elevated temperatures: Consequences for isotope geochemistry USGS Publications Warehouse Truesdell, A.H. 1974-01-01 Studies of the effect of dissolved salts on the oxygen isotope activity ratio of water have been extended to 275??C. Dehydrated salts were added to water of known isotope composition and the solutions were equilibrated with CO2 which was sampled for analysis. For comparison similar studies were made using pure water. Results on water nearly coincide with earlier calculations. Salt effects diminish with increasing temperature only for solutions of MgCl2 and LiCl. Other salt solutions show complex behavior due to the temperature-dependent formation of ion pairs of changing character. Equilibrium fractionations (103 ln ??) between 1 molal solutions and pure water at 25, 100, and 275??C are: NaCl 0.0, -1.5, +1.0; KCl 0.0, -1.0, +2.0; LiCl -1.0, -0.6, -0.5; CaCl2 -0.4, -1.8, +0.8; MgCl2 -1.1, -0.7, -0.3; MgSO4 -1.1, +0.1, -; NaF (0.8 m) 0.0, -1.5, -0.3; and NH4Cl (0.55 m) 0.0, -1.2, -1.3. These effects are significant in the isotope study of hot saline fluids responsible for ore deposition and of fluids found in certain geothermal systems. Minor modification of published isotope geothermometers may be required. ?? 1974. 13. Effects of salts and ethanol on the population and morphology of triblock copolymer micelles in solution. PubMed Denkova, A G; Mendes, E; Coppens, M-O 2008-01-24 The morphological changes of micelles composed of triblock copolymer of ethylene oxide and propylene oxide (EO20PO70EO20) in the presence of different inorganic salts and ethanol have been investigated using dynamic light scattering (DLS), rheometry, and cryogenic transmission electron microscopy (cryo-EM). The following salts were studied: KF, KCl, KI, LiCl, and CsCl. In the presence of KF, KCl, and CsCl, spherical and wormlike micelles coexist. LiCl and KI have little influence on the morphology of the micelles, whereas KF has the most pronounced effect. In agreement with the well-known Hoffmeister anion salt series, F- has the strongest effect of the three anions studied (F-, Cl-, I-). In contrast, the effectiveness of the cation type does not follow the original Hoffmeister cation series. The addition of ethanol to the KCl micellar solutions leads to the formation of more or longer wormlike micelles, which start to interact at certain copolymer concentrations depending on the volume fraction of ethanol added. Both the dilute and the semidilute regimes of the wormlike micelles were studied. The length of the micelles reaches a maximum value at around 8-10 vol % ethanol, after which it decreases again. At higher ethanol concentrations (18 vol %), spherical micelles are formed. Conclusions from this study enhance our understanding of the role played by ethanol and salts in the formation of micelle-templated mesoporous materials, such as SBA-15. PMID:18166031 14. Salt Tolerance PubMed Central Xiong, Liming; Zhu, Jian-Kang 2002-01-01 Studying salt stress is an important means to the understanding of plant ion homeostasis and osmo-balance. Salt stress research also benefits agriculture because soil salinity significantly limits plant productivity on agricultural lands. Decades of physiological and molecular studies have generated a large body of literature regarding potential salt tolerance determinants. Recent advances in applying molecular genetic analysis and genomics tools in the model plant Arabidopsis thaliana are shading light on the molecular nature of salt tolerance effectors and regulatory pathways. PMID:22303210 15. Vitrification in the presence of salts SciTech Connect Marra, J.C.; Andrews, M.K.; Schumacher, R.F. 1994-07-01 Glass is an advantageous material for the immobilization of nuclear wastes because of the simplicity of processing and its unique ability to accept a wide variety of waste elements into its network structure. Unfortunately, some anionic species which are present in the nuclear waste streams have only limited solubility in oxide glasses. This can result in either vitrification concerns or it can affect the integrity, of the final vitrified waste form. The presence of immiscible salts can also corrode metals and refractories in the vitrification unit as well as degrade components in the off-gas system. The presence of a molten salt layer on the melt may alter the batch melting rate and increase operational safety concerns. These safety concerns relate to the interaction of the molten salt and the melter cooling fluids. Some preliminary data from ongoing experimental efforts examining the solubility of molten salts in glasses and the interaction of salts with melter component materials is included. 16. Mixing of zeolite powders and molten salt SciTech Connect Pereira, C.; Zyryanov, V.N.; Lewis, M.A.; Ackerman, J.P. 1996-05-01 Transuranics and fission products in a molten salt can be incorporated into zeolite A by an ion exchange process and by a batch mixing or blending process. The zeolite is then mixed with glass and consolidated into a monolithic waste form for geologic disposal. Both processes require mixing of zeolite powders with molten salt at elevated temperatures (>700 K). Complete occlusion of salt and a uniform distribution of chloride and fission products are desired for incorporation of the powders into the final waste form. The relative effectiveness of the blending process was studied over a series of temperature, time, and composition profiles. The major criteria for determining the effectiveness of the mixing operations were the level and uniformity of residual free salt in the mixtures. High operating temperatures (>775 K) improved salt occlusion. Reducing the chloride levels in the mixture to below 80% of the full salt capacity of the zeolite significantly reduced the free salt level in the final product. 17. Temporal and qualitative dynamics of conditioned taste aversions in C57BL/6J and DBA/2J mice self-administering LiCl. PubMed Rebecca Glatt, A; St John, Steven J; Lu, Lianyi; Boughter, John D 2016-01-01 Self-administration of LiCl solution has been shown to result in the formation of a conditioned taste aversion (CTA) that generalizes to NaCl in rats. This paradigm may have considerable ecological validity as it models CTA learning in natural settings, and also allows for the investigation of drinking microstructure as an assay of potential shifts in stimulus palatability. We used this paradigm to examine possible mouse strain differences in CTA acquisition, generalization, and extinction. In the first experiment, C57BL/6J (B6) and DBA/2J (D2) mice self-administered LiCl (or control NaCl) over a 20-minute free access acquisition period and were tested on the following day with a panel of taste solutions available in brief (5-s) trials delivered in random order. In the second experiment, mice again self-administered LiCl or NaCl (at low, 0.12 M, or high, 0.24 M concentrations) in a 20-minute session, and on the following day received a 20-minute free access period to equimolar NaCl. Strain differences were found for aspects of ingestive behavior, with B6 mice showing greater consumption of all stimuli, including water, while D2 mice lick faster, in less frequent but longer bursts. We did not, however, find evidence of a robust strain difference in taste aversion learning. Both strains demonstrated profound alterations in licking microstructure in the generalization session relative to controls. We suggest that a decrease in "lick efficiency" (the percentage of inter-lick intervals within a burst of short duration vs. longer duration) reflects avoidance behavior, and signals a shift in palatability of a stimulus following CTA. PMID:26524511 18. Radioactive waste isolation in salt: SciTech Connect Ditmars, J.D.; Baecher, G.B.; Edgar, D.E.; Dowding, C.H. 1988-03-01 The approach presumes that measurements are undertaken to support performance predictions. A quantitative performance objective like groundwater travel time is compared with performance predictions. The approach recognizes that such predictions are uncertain because the measurements upon which they are based are uncertain. The effectiveness of measurement activities is quantified by an index, ..beta.., that reflects the number of standard deviations separating the best estimate of performance from the performance objective. Measurements that reduce the uncertainty in predictions lead to increased values of ..beta... Evaluating ..beta.. for a particular measurement scheme requires identifying the measured quantities that significantly affect prediction uncertainty. Sources of uncertainty are spatial variation, noise, estimation error, and measurement bias. Changing the measurement scheme to increase ..beta.. increases the likelihood of a performance objective being achieved or exceeded. The application of the ..beta..-index method to the Richton dome site in Mississippi focuses on uncertainties in hydraulic conductivity data in relation to groundwater travel time predictions. The ..beta.. values for four different measurement schemes for hydraulic conductivity are determined. 44 refs., 14 figs., 15 tabs. 19. Salt-specific effects observed in calorimetric studies of alkali and tetraalkylammonium salt solutions of poly(thiophen-3-ylacetic acid). PubMed Hostnik, Gregor; Vlachy, Vojko; Bondarev, Dmitrij; Vohldal, Jir Combining Breve ; Cerar, Janez 2015-01-28 The enthalpies of dilution ?Hdil of aqueous solutions of a conjugated polyelectrolyte, poly(thiophen-3-ylacetic acid), neutralized by lithium, sodium, cesium, tetramethyl-, tetraethyl-, tetrapropyl-, and tetrabutylammonium hydroxides, were determined in the concentration range from cp = 2 10(-3) to 1 10(-1) monomol dm(-3) and for T = 278.15, 298.15, and 318.15 K. At low concentrations the dilution of the alkali PTAA salts yields an endothermic effect, which is in part a consequence of the hydrolysis. An exception is PTALi at 278.15 K, where ?Hdil < 0. In the case of tetraalkylammonium salts the enthalpies of dilution increase in the order TBA < TPA < TEA < TMA. Only the TBA salt of PTAA yields an exothermic effect upon dilution in the whole temperature range. In the second part of the study we measured the enthalpies of mixing, ?Hmix, of various salts of poly(thiophen-3-ylacetic acid) with LiCl, NaCl, KCl, and CsCl solutions in water. When lithium salt of PTAA is mixed with LiCl ?Hmix is positive. For mixing experiments with other alkali chlorides the effect is exothermic. In addition, the enthalpies of mixing of PTALi with tetramethyl-, tetraethyl-, tetrapropyl-, and tetrabutylammonium chloride were measured at T = 278.15 K, 298.15 K, and 318.15 K. Popular polyelectrolyte theories, such as Manning's limiting law, predict for the heat to be released upon dilution, and consumed upon mixing; the agreement between this purely electrostatic theory and experiments is at best qualitative. The ?Hmix values are correlated with the enthalpies of hydration of the cations of the low molecular mass salts added to the solution. PMID:25491322 20. Salt assault SciTech Connect Duey, R. 1999-12-01 If the offshore arena is the last great frontier for oil and gas exploration and development, then salt may very well be the last great barrier to exploring that frontier. In the game of seismic acquisition, salt refuses to follow the rules. Its relatively low density coupled with high velocity and structural complexity plays havoc with seismic imaging, rendering data acquired from beneath the salt virtually useless without massive processing and interpretation efforts. Additionally, salt's amazing mobility means surrounding sediments also are mobile. This makes exploration in salt provinces particularly challenging. So understanding where the salt is, and how and why it got there, is becoming increasingly important, and research is under way around the world to get a better handle on this tricky substance. Hart's E&P went to the experts--academics who work at universities or in consulting capacities to study salt deposition and tectonics--and asked them about their research and its impact on oil and gas exploration. The paper discusses the latest findings, modeling, practical applications, and questions that remain. 1. Clean salt process final report SciTech Connect Herting, D.L. 1996-09-30 A process has been demonstrated in the laboratory for separating clean, virtually non-radioactive sodium nitrate from Hanford tank waste using fractional crystallization. The name of the process is the Clean Salt Process. Flowsheet modeling has shown that the process is capable of reducing the volume of vitrified low activity waste (LAW) by 80 to 90 %. Construction of the Clean Salt processing plant would cost less than $1 10 million, and would eliminate the need for building a$2.2 billion large scale vitrification plant planned for Privatization Phase 11. Disposal costs for the vitrified LAW would also be reduced by an estimated $240 million. This report provides a summary of five years of laboratory and engineering development activities, beginning in fiscal year 1992. Topics covered include laboratory testing of a variety of processing options; proof-of-principle demonstrations with actual waste samples from Hanford tanks 241-U-110 (U-110), 241-SY-101 (101-SY), and 241-AN-102 (102-AN); descriptions of the primary solubility phase diagrams that govem the process; a review of environmental regulations governing disposition of the reclaimed salt and an assessment of the potential beneficial uses of the reclaimed salt; preliminary plant design and construction cost estimates. A detailed description is given for the large scale laboratory demonstration of the process using waste from tank 241-AW-101 (101-AW), a candidate waste for 0044vitrification during Phase I Privatization. 2. Bath Salts MedlinePLUS ... feel paranoid, and develop a psychosis that resembles schizophrenia. Other long-term effects include: physical exhaustion insomnia ... in them. The mix of chemicals and these unknown ingredients can cause dangerous side effects. Bath salts ... 3. Bath Salts MedlinePLUS ... decreased ability to think clearly mood disturbances and psychosis Like other stimulants, bath salts also can affect ... hallucinations, hear voices, feel paranoid, and develop a psychosis that resembles schizophrenia. Other long-term effects include: ... 4. Thermophysical properties of reconsolidating crushed salt. SciTech Connect Bauer, Stephen J.; Urquhart, Alexander 2014-03-01 Reconsolidated crushed salt is being considered as a backfilling material placed upon nuclear waste within a salt repository environment. In-depth knowledge of thermal and mechanical properties of the crushed salt as it reconsolidates is critical to thermal/mechanical modeling of the reconsolidation process. An experimental study was completed to quantitatively evaluate the thermal conductivity of reconsolidated crushed salt as a function of porosity and temperature. The crushed salt for this study came from the Waste Isolation Pilot Plant (WIPP). In this work the thermal conductivity of crushed salt with porosity ranging from 1% to 40% was determined from room temperature up to 300oC, using two different experimental methods. Thermal properties (including thermal conductivity, thermal diffusivity and specific heat) of single-crystal salt were determined for the same temperature range. The salt was observed to dewater during heating; weight loss from the dewatering was quantified. The thermal conductivity of reconsolidated crushed salt decreases with increasing porosity; conversely, thermal conductivity increases as the salt consolidates. The thermal conductivity of reconsolidated crushed salt for a given porosity decreases with increasing temperature. A simple mixture theory model is presented to predict and compare to the data developed in this study. 5. Nitrate Salt Surrogate Blending Scoping Test Plan SciTech Connect Anast, Kurt Roy 2015-11-13 Test blending equipment identified in the “Engineering Options Assessment Report: Nitrate Salt Waste Stream Processing”. Determine if the equipment will provide adequate mixing of zeolite and surrogate salt/Swheat stream; optimize equipment type and operational sequencing; impact of baffles and inserts on mixing performance; and means of validating mixing performance 6. Anomalous ion effects on rupture and lifetime of aqueous foam films from monovalent salt solutions up to saturation concentration NASA Astrophysics Data System (ADS) Karakashev, S. I.; Nguyen, P. T.; Tsekov, R.; Hampton, M. A.; Nguyen, A. V. 2008-09-01 We report the effects of ions on rupture and lifetime of aqueous foam films formed from sodium chloride (NaCl), lithium chloride (LiCl), sodium acetate (NaAc), and sodium chlorate (NaClO 3) using microinterferometry. In the case of NaCl and LiCl, the foam films prepared from the salt solutions below 0.1 M were unstable they thinned until rupturing. The film lifetime measured from the first interferogram (appearing at a film thickness on the order of 500 nm) until the film rupture was only a second or so. However, relatively long lasting and nondraining films prepared from salt solutions above 0.1 M were observed. The film lifetime was significantly longer by 1 to 2 orders of magnitude, i.e., from 10 to 100 s. Importantly, both the film lifetime and the (average) thickness of the nondraining films increased with increasing salt concentration. This effect has not been observed with foam films stabilized by surfactants. The film lifetime and thickness also increased with increasing film radius. The films exhibited significant surface corrugations. The films with large radii often contained standing dimples. There was a critical film radius below which the films thinned until rupturing. In the cases of NaAc and NaClO 3, the films were unstable at all radii and salt concentrations they thinned until rupturing, ruling out the effect of solution viscosity on stabilizing the films. 7. Salt splitting with ceramic membranes SciTech Connect Kurath, D. 1996-10-01 The purpose of this task is to develop ceramic membrane technologies for salt splitting of radioactively contaminated sodium salt solutions. This technology has the potential to reduce the low-level waste (LLW) disposal volume, the pH and sodium hydroxide content for subsequent processing steps, the sodium content of interstitial liquid in high-level waste (HLW) sludges, and provide sodium hydroxide free of aluminum for recycle within processing plants at the DOE complex. Potential deployment sites include Hanford, Savannah River, and Idaho National Engineering Laboratory (INEL). The technical approach consists of electrochemical separation of sodium ions from the salt solution using sodium (Na) Super Ion Conductors (NaSICON). As the name implies, sodium ions are transported rapidly through these ceramic crystals even at room temperatures. 8. Radioactive waste isolation in salt: peer review of the Texas Bureau of Economic Geology's report on the Petrographic, Stratigraphic, and Structural Evidence for Dissolution of Upper Permian Bedded Salt, Texas Panhandle SciTech Connect Fenster, D.F.; Anderson, R.Y.; Gonzales, S.; Baker, V.R.; Edgar, D.E.; Harrison, W. 1984-08-01 The following recommendations for improving the Texas Bureau of Economic Geology (TBEG) report entitled Petrographic, Stratigraphic, and Structural Evidence for Dissolution of Upper Permian Bedded Salt, Texas Panhandle have been abstracted from the body of this review report. The TBEG report should be resided to conform to one of the following alternatives: (1) If the report is intended to be a review or summary of previous work, it should contain more raw data, be edited to give equal treatment to all types of data, and include summary tables and additional figures. (2) If the report is intended to be a description and interpretation of petrographic evidence for salt dissolution, supported by collateral stratigraphic and structural evidence, the relevant indirect and direct data should become the focal point of the report. The following recommendations apply to one or both of the options listed above. (1) The text should differentiate more carefully between the data and inferences based on those data. (2) The authors should retain the qualifiers present in cited works. Statements in the report that are based on earlier papers are sometimes stronger than those in the papers themselves. (3) The next revision should present more complete data. (4) The authors should achieve a more balanced presentation of alternative hypotheses and interpretations. They could then discuss the relative merits of the alternative interpretations. (5) More attention should be given to clear exposition. 9. Conceptual designs for waste packages for horizontal or vertical emplacement in a repository in salt for reference in the site characterization plan SciTech Connect Not Available 1987-06-01 This report includes the options of horizontal and vertical emplacement, the addition of a phased repository, an additional waste form (intact spent fuel), revised geotechnical data appropriate for the Deaf Smith County site, new corrosion data for the container, and new repository design data. The waste package consists of waste form and canister within a thick-walled, low-carbon steel container surrounded by packing. The container is a hollow cylinder with a flat head welded to each end. The design concepts for the waste container or vertical and horizontal emplacement are identical. This report discusses the results of analyses of aspects of the reference waste package concept needing changes because of new data and information believed applicable to the Deaf Smith County site. Included are waste package conceptual designs or (1) the reference defense high-level waste form from the Savannah River Plant; (2) intact spent fuel with our pressurized-water-reactor or nine boiling-water-reactor assemblies per package for emplacement during Phase 1 of repository operation; and (3) spent fuel which has been disassembled and consolidated into a segmented cylindrical canister with rods from either 12 pressurized-water-reactor or 30 boiling-water-reactor assemblies per package for emplacement during Phase 2. 30 refs., 61 figs., 30 tabs. 10. SOLUTION MINING IN SALT DOMES OF THE GULF COAST EMBAYMENT SciTech Connect Griswold, G. B. 1981-02-01 Following a description of salt resources in the salt domes of the gulf coast embayment, mining, particularly solution mining, is described. A scenario is constructed which could lead to release of radioactive waste stored in a salt dome via inadvertent solution mining and the consequences of this scenario are analyzed. 11. High-Temperature Continuous-Flow Zincations of Functionalized Arenes and Heteroarenes Using (Cy2N)2Zn·2LiCl. PubMed Becker, Matthias R; Knochel, Paul 2016-03-18 The treatment of sensitive arenes and heteroarenes with the zinc bis-amide (Cy2N)2Zn·2LiCl (0.55 equiv), prepared in quantitative yield by the reaction of Cy2NLi with ZnCl2, leads under flow conditions to a fast zincation within 10 min at temperatures between 25 and 100 °C. The resulting organozinc reagents can be trapped with various organic halides (allylic bromides, aryl iodides) in high yields. Moreover, complementary metalation regioselectivities can be obtained for several substituted pyridines compared to commonly used LiCl-activated TMP-zinc (TMP = 2,2,6,6-tetramethylpiperidyl) and -magnesium bases. PMID:26942552 12. Insertion/Rearrangement Reactivity of a Lutetacyclopentadiene towards N,N'-Diphenylcarbodiimide: Cooperative Effect of the Metal Center, Concentration of LiCl, and Solvent. PubMed Xu, Ling; Wei, Junnian; Zhang, Wen-Xiong; Xi, Zhenfeng 2015-10-26 The reactivity of lutetacyclopentadiene towards N,N'-diphenylcarbodiimide (DPC) was systematically investigated to efficiently construct three types of new N-containing fused cyclic complexes. The outcome of these reactions significantly depended on the the metal center, the concentration of LiCl, the number of equivalents of DPC, and the solvent. Thus, two unexpected reaction modes of amidinate were discovered for the first time, which were ascribed to an unusual ?(1) coordination mode of amidinate driven by a rigid seven-membered ring. These results are in striking contrast with the previously well-investigated zirconacyclopentadiene, which reacts with DPC to give azazirconacyclopentene and alkyne through ?,?' C?C bond cleavage. The difference in reactivity between lutetacyclopentadiene and zirconacyclopentadiene can be attributed to the highly ionic character of the Lu?C(sp(2) ) bonds. DFT calculations agreed well with the experimental results. PMID:26359652 13. Listeria phage and phage tail induction triggered by components of bacterial growth media (phosphate, LiCl, nalidixic acid, and acriflavine). PubMed Lematre, Jean-Paul; Duroux, Amandine; Pimpie, Romain; Duez, Jean-Marie; Milat, Marie-Louise 2015-03-01 The detection of Listeria monocytogenes from food is currently carried out using a double enrichment. For the ISO methodology, this double enrichment is performed using half-Fraser and Fraser broths, in which the overgrowth of L. innocua can occur in samples where both species are present. In this study, we analyzed the induction of phages and phage tails of Listeria spp. in these media and in two brain heart infusion (BHI) broths (BHIM [bioMrieux] and BHIK [Biokar]) to identify putative effectors. It appears that Na2HPO4 at concentrations ranging from 1 to 40 g/liter with an initial pH of 7.5 can induce phage or phage tail production of Listeria spp., especially with 10 g/liter of Na2HPO4 and a pH of 7.5, conditions present in half-Fraser and Fraser broths. Exposure to LiCl in BHIM (18 to 21 g/liter) can also induce phage and phage tail release, but in half-Fraser and Fraser broths, the concentration of LiCl is much lower (3 g/liter). Low phage titers were induced by acriflavine and/or nalidixic acid. We also show that the production of phages and phage tails can occur in half-Fraser and Fraser broths. This study points out that induction of phages and phage tails could be triggered by compounds present in enrichment media. This could lead to a false-negative result for the detection of L. monocytogenes in food products. PMID:25595760 14. Listeria Phage and Phage Tail Induction Triggered by Components of Bacterial Growth Media (Phosphate, LiCl, Nalidixic Acid, and Acriflavine) PubMed Central Duroux, Amandine; Pimpie, Romain; Duez, Jean-Marie; Milat, Marie-Louise 2015-01-01 The detection of Listeria monocytogenes from food is currently carried out using a double enrichment. For the ISO methodology, this double enrichment is performed using half-Fraser and Fraser broths, in which the overgrowth of L. innocua can occur in samples where both species are present. In this study, we analyzed the induction of phages and phage tails of Listeria spp. in these media and in two brain heart infusion (BHI) broths (BHIM [bioMérieux] and BHIK [Biokar]) to identify putative effectors. It appears that Na2HPO4 at concentrations ranging from 1 to 40 g/liter with an initial pH of 7.5 can induce phage or phage tail production of Listeria spp., especially with 10 g/liter of Na2HPO4 and a pH of 7.5, conditions present in half-Fraser and Fraser broths. Exposure to LiCl in BHIM (18 to 21 g/liter) can also induce phage and phage tail release, but in half-Fraser and Fraser broths, the concentration of LiCl is much lower (3 g/liter). Low phage titers were induced by acriflavine and/or nalidixic acid. We also show that the production of phages and phage tails can occur in half-Fraser and Fraser broths. This study points out that induction of phages and phage tails could be triggered by compounds present in enrichment media. This could lead to a false-negative result for the detection of L. monocytogenes in food products. PMID:25595760 15. 76 FR 47613 - Board Meeting: September 13-14, 2011-Salt Lake City, UT; the U.S. Nuclear Waste Technical Review... Federal Register 2010, 2011, 2012, 2013, 2014 2011-08-05 ... transportation and long-term storage of spent nuclear fuel (SNF), studies of various geologic media that might be considered for disposing of SNF and high-level radioactive waste (HLW), and DOE's Used Fuel Disposition... 16. Process to separate transuranic elements from nuclear waste DOEpatents Johnson, Terry R. (Wheaton, IL); Ackerman, John P. (Downers Grove, IL); Tomczuk, Zygmunt (Orland Park, IL); Fischer, Donald F. (Glen Ellyn, IL) 1989-01-01 A process for removing transuranic elements from a waste chloride electrolytic salt containing transuranic elements in addition to rare earth and other fission product elements so the salt waste may be disposed of more easily and the valuable transuranic elements may be recovered for reuse. The salt is contacted with a cadmium-uranium alloy which selectively extracts the transuranic elements from the salt. The waste salt is generated during the reprocessing of nuclear fuel associated with the Integral Fast Reactor (IFR). 17. A new compound heterozygous frameshift mutation in the type II 3{beta}-hydroxysteroid dehydrogenase 3{beta}-HSD gene causes salt-wasting 3{beta}-HSD deficiency congenital adrenal hyperplasia SciTech Connect Zhang, L.; Sakkal-Alkaddour, S.; Chang, Ying T.; Yang, Xiaojiang; Songya Pang 1996-01-01 We report a new compound heterozygous frameshift mutation in the type II 3{Beta}-hydroxysteroid dehydrogenase (3{beta}-HSD) gene in a Pakistanian female child with the salt-wasting form of 3{Beta}-HSD deficiency congenital adrenal hyperplasia. The etiology for her congenital adrenal hyperplasia was not defined. Although the family history suggested possible 3{beta}-HSd deficiency disorder, suppressed adrenal function caused by excess glucocorticoid therapy in this child at 7 yr of age did not allow hormonal diagnosis. To confirm 3{beta}-HSD deficiency, we sequenced the type II 3{beta}-HSD gene in the patient, her family, and the parents of her deceased paternal cousins. The type II 3{beta}-HSD gene region of a putative promotor, exons I, II, III, and IV, and exon-intron boundaries were amplified by PCR and sequenced in all subjects. The DNA sequence of the child revealed a single nucleotide deletion at codon 318 [ACA(Thr){r_arrow}AA] in exon IV in one allele, and two nucleotide deletions at codon 273 [AAA(Lys){r_arrow}A] in exon IV in the other allele. The remaining gene sequences were normal. The codon 318 mutation was found in one allele from the father, brother, and parents of the deceased paternal cousins. The codon 273 mutation was found in one allele of the mother and a sister. These findings confirmed inherited 3{beta}-HSD deficiency in the child caused by the compound heterozygous type II 3{beta}-HSD gene mutation. Both codons at codons 279 and 367, respectively, are predicted to result in an altered and truncated type II 3{beta}-HSD protein, thereby causing salt-wasting 3{beta}-HSD deficiency in the patient. 21 refs., 2 figs., 1 tab. 18. Selective Separation of Cs and Sr from LiCl-Based Salt for Electrochemical Processing of Oxide Spent Nuclear Fuel SciTech Connect P Sachdev 2008-07-01 Electrochemical processing technology is currently being used for the treatment of metallic spent fuel from the Experimental Breeder Reactor-II at Idaho National Laboratory. The treatment of oxide-based spent nuclear fuel via electrochemical processing is possible provided there is a front-end oxide reduction step. During this reduction process, certain fission products, including Cs and Sr, partition into the salt phase and form chlorides. Both solid state and molten LiCl-zeolite-A ion exchange tests were conducted for selectively removing Cs and Sr from LiCl-based salt. The solid-state tests produced in excess of 99% removal of Cs and Sr. The molten state tests failed due to phase transformation of the zeolite structure when in contact with the molten LiCl salt. 19. Bases, Assumptions, and Results of the Flowsheet Calculations for the Decision Phase Salt Disposition Alternatives SciTech Connect Elder, H.H. 2001-07-11 The HLW salt waste (salt cake and supernate) now stored at the SRS must be treated to remove insoluble sludge solids and reduce the soluble concentration of radioactive cesium radioactive strontium and transuranic contaminants (principally Pu and Np). These treatments will enable the salt solution to be processed for disposal as saltstone, a solid low-level waste. 20. Efficacy of a solution-based approach for making sodalite waste forms for an oxide reduction salt utilized in the reprocessing of used uranium oxide fuel DOE PAGESBeta Riley, Brian J.; Pierce, David A.; Frank, Steven M.; Matyáš, Josef; Burns, Carolyne A. 2015-04-01 This paper describes the various approaches attempted to make solution-derived sodalite with a LiCl-Li2O oxide reduction salt used to dissolve used uranium oxide fuel so the uranium can be recovered and recycled. The approaches include modified sol-gel and solutionbased synthesis processes. As-made products were mixed with 5 and 10 mass% of a Na2O-B2O3- SiO2 glass binder and these, along with product without a binder, were heated using either a cold-press-and-sinter method or hot uniaxial pressing. The results demonstrate the limitation of sodalite yield due to the fast intermediate reactions between Na+ and Cl- to form halite in solution and Li2Omore » and SiO2 to form lithium silicates (e.g., Li2SiO3 or Li2Si2O5) in the calcined and sintered pellets. The results show that pellets can be made with high sodalite fractions in the crystalline product (~92 mass%) and low porosities using a solution-based approach and this LiCl-Li2O salt but that the incorporation of Li into the sodalite is low.« less 1. Efficacy of a Solution-Based Approach for Making Sodalite Waste Forms for an Oxide Reduction Salt Utilized in the Reprocessing of Used Uranium Oxide Fuel SciTech Connect Riley, Brian J.; Pierce, David A.; Frank, Steven M.; Matyas, Josef; Burns, Carolyn A. 2015-04-01 This paper describes various approaches for making sodalite with a LiCl-Li2O oxide reduction salt used to recover uranium from used oxide fuel. The approaches include sol-gel and solution-based synthesis processes. As-made products were mixed with 5 and 10 mass% of a Na2O-B2O3-SiO2 glass binder and these, along with product without a binder, were heated using either a cold-press-and-sinter method or hot uniaxial pressing. The results demonstrate the limitation of sodalite yield due to the fast intermediate reactions between Na+ and Cl- to form halite in solution and Li2O and SiO2 to form lithium silicates (e.g., Li2SiO3 or Li2Si2O5) in the calcined and sintered pellets. The results show that pellets can be made with high sodalite fractions (~92 mass%) and low porosities using a solution-based approach and this LiCl-Li2O salt. 2. Efficacy of a solution-based approach for making sodalite waste forms for an oxide reduction salt utilized in the reprocessing of used uranium oxide fuel NASA Astrophysics Data System (ADS) Riley, Brian J.; Pierce, David A.; Frank, Steven M.; Maty, Josef; Burns, Carolyne A. 2015-04-01 This paper describes the various approaches evaluated for making solution-derived sodalite with a LiCl-Li2O oxide reduction salt selected to dissolve used uranium oxide fuel so the uranium can be recovered and recycled. The approaches include modified sol-gel and solution-based synthesis processes. As-made products were mixed with 5 and 10 mass% of a Na2O-B2O3-SiO2 glass binder and these, along with product without a binder, were heated using either a cold-press-and-sinter method or hot uniaxial pressing. The results demonstrate the limitation of sodalite yield due to the fast intermediate reactions between Na+ and Cl- to form halite in solution and Li2O and SiO2 to form lithium silicates (e.g., Li2SiO3 or Li2Si2O5) in the calcined and sintered pellets. The results show that pellets can be made with high sodalite fractions in the crystalline product (?92 mass%) and low porosities using a solution-based approach and this LiCl-Li2O salt but that the incorporation of Li into the sodalite is low. 3. Efficacy of a solution-based approach for making sodalite waste forms for an oxide reduction salt utilized in the reprocessing of used uranium oxide fuel SciTech Connect Riley, Brian J.; Pierce, David A.; Frank, Steven M.; Matyáš, Josef; Burns, Carolyne A. 2015-04-01 This paper describes the various approaches attempted to make solution-derived sodalite with a LiCl-Li2O oxide reduction salt used to dissolve used uranium oxide fuel so the uranium can be recovered and recycled. The approaches include modified sol-gel and solutionbased synthesis processes. As-made products were mixed with 5 and 10 mass% of a Na2O-B2O3- SiO2 glass binder and these, along with product without a binder, were heated using either a cold-press-and-sinter method or hot uniaxial pressing. The results demonstrate the limitation of sodalite yield due to the fast intermediate reactions between Na+ and Cl- to form halite in solution and Li2O and SiO2 to form lithium silicates (e.g., Li2SiO3 or Li2Si2O5) in the calcined and sintered pellets. The results show that pellets can be made with high sodalite fractions in the crystalline product (~92 mass%) and low porosities using a solution-based approach and this LiCl-Li2O salt but that the incorporation of Li into the sodalite is low. 4. Salt splitting using ceramic membranes SciTech Connect Kurath, D.E. 1997-10-01 Many radioactive aqueous wastes in the DOE complex have high concentrations of sodium that can negatively affect waste treatment and disposal operations. Sodium can decrease the durability of waste forms such as glass and is the primary contributor to large disposal volumes. Waste treatment processes such as cesium ion exchange, sludge washing, and calcination are made less efficient and more expensive because of the high sodium concentrations. Pacific Northwest National Laboratory (PNNL) and Ceramatec Inc. (Salt Lake City UT) are developing an electrochemical salt splitting process based on inorganic ceramic sodium (Na), super-ionic conductor (NaSICON) membranes that shows promise for mitigating the impact of sodium. In this process, the waste is added to the anode compartment, and an electrical potential is applied to the cell. This drives sodium ions through the membrane, but the membrane rejects most other cations (e.g., Sr{sup +2}, Cs{sup +}). The charge balance in the anode compartment is maintained by generating H{sup +} from the electrolysis of water. The charge balance in the cathode is maintained by generating OH{sup {minus}}, either from the electrolysis of water or from oxygen and water using an oxygen cathode. The normal gaseous products of the electrolysis of water are oxygen at the anode and hydrogen at the cathode. Potentially flammable gas mixtures can be prevented by providing adequate volumes of a sweep gas, using an alternative reductant or destruction of the hydrogen as it is generated. As H{sup +} is generated in the anode compartment, the pH drops. The process may be operated with either an alkaline (pH>12) or an acidic anolyte (pH <1). The benefits of salt splitting using ceramic membranes are (1) waste volume reduction and reduced chemical procurement costs by recycling of NaOH; and (2) direct reduction of sodium in process streams, which enhances subsequent operations such as cesium ion exchange, calcination, and vitrification. 5. Subsidence and collapse at Texas Salt Domes SciTech Connect Mullican, W.F. 1989-01-01 This book provides a description of the mechanisms and extent of natural and man-induced subsidence and collapse at Texas salt domes. In the Houston diapir province, Frasch mining has caused subsidence bowls and collapse sinkholes at 12 of the 14 sulfur-productive domes. Understanding the structural and hydrologic instability that results at the surface and subsurface is crucial in evaluating the suitability of salt domes as repositories for waste disposal. Part of the Bureau's Coastal Salt Dome Program, this study used aerial photographs, remote-sensing methods, historical and modern topographic maps, and field checks to detect subsidence and collapse associated with natural salt diapiric processes and commercial resource recovery and to determine which processes are likely to reduce the stability and integrity of hydrologic and structural barriers around salt diapirs. Figures and tables illustrating the extent and evolution of subsidence and collapse, along with photographs showing their effects, highlight the text discussion of the salt domes detailed in this study-Boling, Orchard, Moss Bluff, Spindletop, Hoskins Mound, Fannett, Long Point, Nash, High Island, Bryan Mound, Clemens, and Gulf. The author concludes that Frasch sulfur mining from cap rocks causes the most catastrophic subsidence and collapse and that subsidence over salt domes includes processes ranging from trough subsidence to various types of subsurface caving. He concludes that salt domes characterized by subsidence and collapse are unfavorable sites for storage/disposal of hazardous wastes. 6. Thermodynamic Properties of Magnesium Chloride Hydroxide Hydrate (Mg3Cl(OH)5:4H2O, Phase 5), and Its importance to Nuclear Waste Isolation in Geological Repositories in Salt Formations NASA Astrophysics Data System (ADS) Xiong, Y.; Deng, H.; Nemer, M. B.; Johnsen, S. 2009-12-01 MgO (bulk, pure MgO corresponding to the mineral periclase) is the only engineered barrier certified by the Environmental Protection Agency for emplacement in the Waste Isolation Pilot Plant (WIPP) in the US, and an Mg(OH)2-based engineered barrier (bulk, pure Mg(OH)2 corresponding to brucite) is to be employed in the Asse repository in Germany. Both the WIPP and the Asse are located in salt formations. The WIPP is a U.S. Department of Energy geological repository being used for the permanent disposal of defense-related transuranic waste (TRU waste). The repository is 655 m below the surface, and is situated in the Salado Formation, a Permian salt bed mainly composed of halite, and of lesser amounts of polyhalite, anhydrite, gypsum, magnesite, clays and quartz. The WIPP Generic Weep Brine (GWB), a Na-Mg-Cl dominated brine, is associated with the Salado Formation. The previous vendor for MgO for the WIPP was Premier Chemicals and the current vendor is Martin Marietta Materials. Experimental studies of both Premier MgO and Martin Marietta MgO with the GWB at SNL indicate the formation of magnesium chloride hydroxide hydrate, Mg3Cl(OH)5:4H2O, termed as phase 5. However, this important phase is lacking in the existing thermodynamic database. In this study, the solubility constant of phase 5 is determined from a series of solubility experiments in MgCl2-NaCl solutions. The solubility constant at 25 oC for the following reaction, Mg3Cl(OH)5:4H2O + 5H+ = 3Mg2+ + 9H2O(l) + Cl- is recommended as 43.210.33 (2?) based on the Specific Interaction Theory (SIT) model for extrapolation to infinite dilution. The log K obtained via the Pitzer equations is identical to the above value within the quoted uncertainty. The Gibbs free energy and enthalpy of formation for phase 5 at 25 oC are derived as -33842 (2?) kJ mol-1 and -38966 (2?) kJ mol-1, respectively. The standard entropy and heat capacity of phase 5 at 25 oC are estimated as 39320 J mol-1 K-1 and 37419 J mol-1 K-1, respectively. Phase 5, and its similar phase, phase 3 (Mg2Cl(OH)3:4H2O), could have a significant role in influencing the geochemical conditions in geological repositories for nuclear waste in salt formations where MgO or brucite is employed as engineered barriers, when Na-Mg-Cl dominated brines react with MgO or brucite. Based on our solubility constant for phase 5 in combination with the literature value for phase 3, we predict that the composition for the invariant point of phase 5 and phase 3 would be mMg = 1.70 and pmH = 8.93 in the Mg-Cl binary system. The recent WIPP Compliance Recertification Application PA Baseline Calculations indicate that phase 5 instead of phase 3 is indeed a stable phase when GWB equilibrates with actinide-source-term phases, brucite, magnesium carbonates, halite and anhydrite. 1. This research is funded by WIPP programs administered by the U.S. Department of Energy. 2. Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energys National Nuclear Security Administration under contract DE-AC04-94AL85000. 7. Treatment of organic waste DOEpatents Grantham, LeRoy F. 1979-01-01 An organic waste containing at least one element selected from the group consisting of strontium, cesium, iodine and ruthenium is treated to achieve a substantial reduction in the volume of the waste and provide for fixation of the selected element in an inert salt. The method of treatment comprises introducing the organic waste and a source of oxygen into a molten salt bath maintained at an elevated temperature to produce solid and gaseous reaction products. The gaseous reaction products comprise carbon dioxide and water vapor, and the solid reaction products comprise the inorganic ash constituents of the organic waste and the selected element which is retained in the molten salt. The molten salt bath comprises one or more alkali metal carbonates, and may optionally include from 1 to about 25 wt.% of an alkali metal sulfate. 8. Electrodialysis-ion exchange for the separation of dissolved salts SciTech Connect Baroch, C.J.; Grant, P.J. 1995-12-31 The Department of Energy generates and stores a significant quantity of low level, high level, and mixed wastes. As some of the DOE facilities are decontaminated and decommissioned, additional and possibly different forms of wastes will be generated. A significant portion of these wastes are aqueous streams containing acids, bases, and salts, or are wet solids containing inorganic salts. Some of these wastes are quite dilute solutions, whereas others contain large quantities of nitrates either in the form of dissolved salts or acids. Many of the wastes are also contaminated with heavy metals, radioactive products, or organics. Some of these wastes are in storage because a satisfactory treatment and disposal processes have not been developed. This report describes the process of electrodialysis-ion exchange (EDIX) for treating aqueous wastes streams consisting of nitrates, sodium, organics, heavy metals, and radioactive species. 9. Low-salt diet MedlinePLUS ... you cook, replace salt with other seasonings. Pepper, garlic, herbs, and lemon are good choices. Avoid packaged spice blends. They often contain salt. Use garlic and onion powder, not garlic and onion salt. ... 10. A Low-Temperature Crossover in Water Dynamics in an Aqueous LiCl Solution: Diffusion Probed by Neutron Spin-Echo and NMR SciTech Connect Mamontov, Eugene; Faraone, Antonio; Hagaman, Edward {Ed} W; Han, Kee Sung; Fratini, E 2010-01-01 Aqueous solutions of lithium chloride are an excellent model system for studying the dynamics of water molecules down to low temperatures without freezing. The apparent dynamic crossover observed in an aqueous solution of LiCl at about 220 to 225 K [Mamontov, JPCB 2009, 113, 14073] is located practically at the same temperature as the crossover found for pure water confined in small hydrophilic pores. This finding suggests a strong similarity of water behavior in these two types of systems. At the same time, studies of solutions allow more effective explorations of the long-range diffusion dynamics, because the water molecules are not confined inside an impenetrable matrix. In contrast to the earlier incoherent quasielastic neutron scattering results obtained for the scattering momentum transfers of 0.3 {angstrom}{sup -1} {le} Q {le} 0.9 {angstrom}{sup -1}, our present incoherent neutron spin-echo measurements at a lower Q of 0.1 {angstrom}{sup -1} exhibit no apparent crossover in the relaxation times down to 200 K. At the same time, our present nuclear magnetic resonance measurements of the diffusion coefficients clearly show a deviation at the lower temperatures from the non-Arrhenius law obtained at the higher temperatures. Our results are consistent with a scenario in which more than one relaxational component may exist below the temperature of the dynamic crossover in water. 11. A low-temperature crossover in water dynamics in an aqueous LiCl solution: diffusion probed by neutron spin-echo and nuclear magnetic resonance. PubMed Mamontov, E; Faraone, A; Hagaman, E W; Han, K S; Fratini, E 2010-12-23 Aqueous solutions of lithium chloride are an excellent model system for studying the dynamics of water molecules down to low temperatures without freezing. The apparent dynamic crossover observed in an aqueous solution of LiCl at about 220 to 225 K [Mamontov, JPCB 2009, 113, 14073] is located practically at the same temperature as the crossover found for pure water confined in small hydrophilic pores. This finding suggests a strong similarity of water behavior in these two types of systems. At the same time, studies of solutions allow more effective explorations of the long-range diffusion dynamics, because the water molecules are not confined inside an impenetrable matrix. In contrast to the earlier incoherent quasielastic neutron scattering results obtained for the scattering momentum transfers of 0.3 (-1) ? Q ? 0.9 (-1), our present incoherent neutron spin-echo measurements at a lower Q of 0.1 (-1) exhibit no apparent crossover in the relaxation times down to 200 K. At the same time, our present nuclear magnetic resonance measurements of the diffusion coefficients clearly show a deviation at the lower temperatures from the non-Arrhenius law obtained at the higher temperatures. Our results are consistent with a scenario in which more than one relaxational component may exist below the temperature of the dynamic crossover in water. PMID:21117619 12. Polyethylene encapsulation of molten salt oxidation mixed low-level radioactive salt residues SciTech Connect Lageraaen, P.R.; Kalb, P.D.; Grimmett, D.L.; Gay, R.L.; Newman, C.D. 1995-10-01 A limited scope treatability study was conducted for polyethylene encapsulation of salt residues generated by a Molten Salt Oxidation (MSO) technology demonstration at the Energy Technology Engineering Center (ETEC), operated by Rockwell International for the US Department of Energy (DOE). During 1992 and 1993, ETEC performed a demonstration with a prototype MSO unit and treated approximately 50 gallons of mixed waste comprised of radioactively contaminated oils produced by hot cell operations. A sample of the mixed waste contaminated spent salt was used during the BNL polyethylene encapsulation treatability study. A nominal waste loading of 50 wt % was successfully processed and waste form test specimens were made for Toxicity Characteristic Leaching Procedure (TCLP) testing. The encapsulated product was compared with base-line TCLP results for total chromium and was found to be well within allowable EPA guidelines. 13. Deuterium isotope effects in the solvolysis of benzal chlorides. 4. Salt effects and. cap alpha. -deuterium isotope effects SciTech Connect Vitullo, V.P.; Wilgis, F.P. 1981-02-25 The effect of LiClO/sub 4/, Bu/sub 4/NClO/sub 4/, LiCl, Bu/sub 4/NCl, and mixtures of these salts on the rate and ..cap alpha..-deuterium isotope effects are reported for the solvolysis of p-methoxybenzal chloride in 75% and 85% (v/v) dioxane-water mixtures at 25/sup 0/C. In the less polar of these solvents (85D), k/sub ext//sup 0/ values from the Winstein equation for the two salts LiClO/sub 4/ and Bu/sub 4/NClO/sub 4/ are the same, and very little common-ion rate depression is observed although induced common-ion rate depression is observed. In the more polar solvent (75D), extensive common-ion rate depression is observed for both LiCl and Bu/sub 4/NCl. k/sub ext//sup 0/ values for Bu/sub 4/NClO/sub 4/ and LiClO/sub 4/ are not the same in this more polar solvent. These results are interpreted by using Winstein's ion-pair mechanism for this solvolysis. 14. Experimental determination of the solubility constant for magnesium chloride hydroxide hydrate (Mg 3Cl(OH) 54H 2O, phase 5) at room temperature, and its importance to nuclear waste isolation in geological repositories in salt formations NASA Astrophysics Data System (ADS) Xiong, Yongliang; Deng, Haoran; Nemer, Martin; Johnsen, Shelly 2010-08-01 In this study, the solubility constant of magnesium chloride hydroxide hydrate, Mg 3Cl(OH) 54H 2O, termed as phase 5, is determined from a series of solubility experiments in MgCl 2-NaCl solutions. The solubility constant in logarithmic units at 25 C for the following reaction, MgCl(OH)4HO+5H=3Mg+9HO(l)+Cl is calculated as 43.21 0.33 (2 ?) based on the specific interaction theory (SIT) model for extrapolation to infinite dilution. The Gibbs free energy and enthalpy of formation for phase 5 at 25 C are derived as -3384 2 (2 ?) kJ mol -1 and -3896 6 (2 ?) kJ mol -1, respectively. MgO (bulk, pure MgO corresponding to the mineral periclase) is the only engineered barrier certified by the Environmental Protection Agency (EPA) for emplacement in the Waste Isolation Pilot Plant (WIPP) in the US, and an Mg(OH) 2-based engineered barrier (bulk, pure Mg(OH) 2 corresponding to brucite) is to be employed in the Asse repository in Germany. Phase 5, and its similar phase, phase 3 (Mg 2Cl(OH) 34H 2O), could have a significant role in influencing the geochemical conditions in geological repositories for nuclear waste in salt formations where MgO or brucite is employed as engineered barriers. Based on our solubility constant for phase 5 in combination with the literature value for phase 3, we predict that the composition for the invariant point of phase 5 and phase 3 would be mMg = 1.70 and pmH = 8.94 in the Mg-Cl binary system. The recent WIPP Compliance Recertification Application Performance Assessment Baseline Calculations indicate that phase 5, instead of phase 3, is indeed a stable phase when the WIPP Generic Weep Brine (GWB), a Na-Mg-Cl-dominated brine associated with the Salado Formation, equilibrates with actinide-source-term phases, brucite, magnesium carbonates, halite and anhydrite. Therefore, phase 5 is important to the WIPP, and potentially important to other repositories in salt formations. 15. Stability of YBa sub 2 Cu sub 3 O sub 7 minus x in molten chloride salts SciTech Connect Raeder, C.H.; Knorr, D.B. . Dept. of Materials Engineering) 1990-08-01 The decomposition products of YBa{sub 2}Cu{sub 3}O{sub 7{minus}x} depend on the composition of the molten chloride salt for exposure at 1173 K in air. The presence of dichloride salts such as CuCl{sub 2}, CaCl{sub 2}, or MgCl{sub 2} promote formation of CuO, Cu{sub 2}Y{sub 2}O{sub 59} and loss of barium to the chloride salt as BaCl{sub 2}. Salts based on BaCl{sub 2} or containing LiCl results in YBa{sub 2}Cu{sub 3}O{sub 7{minus}x} decomposition products of Y{sub 2}BaCuP{sub 59} CuO, and BaCl{sub 2}. High barium activity in the salt supports formation of the Y{sub 2}BaCuO{sub 5} phase and reaction of CO{sub 2} with the salt producing BaCO{sub 3}. Decomposition is most sluggish in binary NaCl-KCl salts where minimal amounts of reaction or decomposition products are observed. 16. Crack closure and healing studies in WIPP (Waste Isolation Pilot Plant) salt using compressional wave velocity and attenuation measurements: Test methods and results SciTech Connect Brodsky, N.S. ) 1990-11-01 Compressional wave ultrasonic data were used to qualitatively assess the extent of crack closure during hydrostatic compression of damaged specimens of WIPP salt. Cracks were introduced during constant strain-rate triaxial tests at low confining pressure (0.5 MPa) as specimens were taken to either 0.5, 1.0, or 1.5 percent axial strain. For three specimens taken to 1.0 percent axial strain, the pressure was increased to 5, 10 or 15 MPa. For the remaining specimens, pressure was raised to 15 MPa. Waveforms for compressional waves traveling both parallel and perpendicular to the direction of maximum principal stress were measured in the undamaged state, during constant strain-rate tests, and then monitored as functions of time while the specimens were held at pressure. Both wave velocities and amplitudes increased over time at pressure, indicating that cracks closed and perhaps healed. The recovery of ultrasonic wave characteristics depended upon both pressure and damage level. The higher the pressure, the greater the velocity recovery; however, amplitude recovery showed no clear correlation with pressure. For both amplitudes and velocities, recoveries were greatest in the specimens with the least damage. 13 refs., 15 figs., 1 tab. 17. Effect of Salt Ions on the Proton-Transfer Rate in 3-Hydroxyflavone NASA Astrophysics Data System (ADS) Tomin, V. I.; Ushakou, D. V. 2015-05-01 The effects of lithium and chloride ions on the spectral characteristics of the two-band fluorescence of 3-hydroxyflavone solutions were studied. The obtained absorption, emission, and fluorescence excitation spectra showed that the salt ions effectively modulated intramolecular proton transfer in the excited state, which led to increases in the shortwavelength band intensity and the fluorescence quantum yield. Thus, the quantum yield for 3-hydroxyflavone in acetonitrile with salt concentrations 0.005, 0.01, and 0.02 M increased by 1.5, 2.0, and 3.0 times, respectively, with standard excitation in the principal absorption band. The intensity ratio of the short-wavelength and longwavelength bands also increased as the salt concentration increased. The increase was so significant that the color change of the fluorescence was easily noticed visually. The good sensitivity of the fluorescence-band intensity ratio to the salt concentration could provide a basis for solving the inverse problem, i.e., determining the salt concentration in solutions and biological objects. An explanation allowing the obtained changes in the fluorescence spectral properties after adding LiCl to the solution to be interpreted was proposed. 18. Understanding radioactive waste SciTech Connect Murray, R.L. 1981-12-01 This document contains information on all aspects of radioactive wastes. Facts are presented about radioactive wastes simply, clearly and in an unbiased manner which makes the information readily accessible to the interested public. The contents are as follows: questions and concerns about wastes; atoms and chemistry; radioactivity; kinds of radiation; biological effects of radiation; radiation standards and protection; fission and fission products; the Manhattan Project; defense and development; uses of isotopes and radiation; classification of wastes; spent fuels from nuclear reactors; storage of spent fuel; reprocessing, recycling, and resources; uranium mill tailings; low-level wastes; transportation; methods of handling high-level nuclear wastes; project salt vault; multiple barrier approach; research on waste isolation; legal requiremnts; the national waste management program; societal aspects of radioactive wastes; perspectives; glossary; appendix A (scientific American articles); appendix B (reference material on wastes). (ATT) 19. Geochemistry and petrology of surface samples, six boreholes and brines from the Salton Sea geothermal field: A natural analog of a nuclear waste repository in salt: Report No. 3 SciTech Connect Not Available 1987-05-01 Cuttings from six wells in the Salton Sea geothermal field, and rocks at outcrop that are correlative in age with those encountered at depth in the wells were analyzed in detail. Mineralogy, petrography, x-ray diffraction, electron microprobe, instrumental neutron activation analysis, fission track radiography, oxygen and stable carbon isotopic, uranium-thorium series disequilibrium, and fluid inclusion analyses are reported. Where fluids were being produced from wells, brine chemistry as well as stable isotope and uranium-thorium series analyses are reported. Particular attention has been paid to defining zones of fluid-rock interaction in which analyses of coexisting geothermal reservoir brine and hydrothermally altered sediments could be acquired. A wide span of temperatures, from surficial to greater than 300/degree/C, and salinities ranging from relatively dilute ground waters up to brines of 25 wt% total dissolved solids, span a range of environments that might be encountered in a waste repository in salt. Progressive hydrothermal alteration, mineral formation and element mobility are documented in the data presented. 52 refs., 25 figs., 49 tabs. 20. Military nuclear waste disposal NASA Astrophysics Data System (ADS) Robb, David W. 1984-04-01 A National Research Council (NRC) panel has endorsed a plan for a proposed underground military nuclear waste disposal facility located on a site near Carlsbad, N.M. The Department of Energy (DOE) asked NRC to evaluate the geologic suitability of the site.The NRC panel, chaired by Frank L. Parker of Vanderbilt University, concluded in its final report that the important issues about the geology of the site have been resolved Those issues include the purity and volume of salt, the absence of brine pockets at the repository horizon in the areas excavated, the absence of breccia pipes and of toxic gases, and the nearly horizontal bedding of the salt. Thick underground salt beds have long been considered prime candidates for nuclear waste repositories. The existence of salt beds is believed to indicate long-term stability. In addition, the salt is flexible and will seal cracks and discontinuities over time. 1. Electrolyte salts for power sources DOEpatents Doddapaneni, N.; Ingersoll, D. 1995-11-28 Electrolyte salts are disclosed for power sources comprising salts of phenyl polysulfonic acids and phenyl polyphosphonic acids. The preferred salts are alkali and alkaline earth metal salts, most preferably lithium salts. 2 figs. 2. Electrolyte salts for power sources DOEpatents Doddapaneni, Narayan (10516 Royal Birkdale, NE., Albuquerque, NM 87111); Ingersoll, David (5824 Mimosa Pl., NE., Albuquerque, NM 87111) 1995-01-01 Electrolyte salts for power sources comprising salts of phenyl polysulfonic acids and phenyl polyphosphonic acids. The preferred salts are alkali and alkaline earth metal salts, most preferably lithium salts. 3. Nuclear waste solutions DOEpatents Walker, Darrel D. (1684 Partridge Dr., Aiken, SC 29801); Ebra, Martha A. (129 Hasty Rd., Aiken, SC 29801) 1987-01-01 High efficiency removal of technetium values from a nuclear waste stream is achieved by addition to the waste stream of a precipitant contributing tetraphenylphosphonium cation, such that a substantial portion of the technetium values are precipitated as an insoluble pertechnetate salt. 4. Deformation-assisted fluid percolation in rock salt NASA Astrophysics Data System (ADS) Ghanbarzadeh, Soheil; Hesse, Marc A.; Prodanović, Maša; Gardner, James E. 2015-11-01 Deep geological storage sites for nuclear waste are commonly located in rock salt to ensure hydrological isolation from groundwater. The low permeability of static rock salt is due to a percolation threshold. However, deformation may be able to overcome this threshold and allow fluid flow. We confirm the percolation threshold in static experiments on synthetic salt samples with x-ray microtomography. We then analyze wells penetrating salt deposits in the Gulf of Mexico. The observed hydrocarbon distributions in rock salt require that percolation occurred at porosities considerably below the static threshold due to deformation-assisted percolation. Therefore, the design of nuclear waste repositories in salt should guard against deformation-driven fluid percolation. In general, static percolation thresholds may not always limit fluid flow in deforming environments. 5. Deformation-assisted fluid percolation in rock salt. PubMed Ghanbarzadeh, Soheil; Hesse, Marc A; Prodanovi?, Maa; Gardner, James E 2015-11-27 Deep geological storage sites for nuclear waste are commonly located in rock salt to ensure hydrological isolation from groundwater. The low permeability of static rock salt is due to a percolation threshold. However, deformation may be able to overcome this threshold and allow fluid flow. We confirm the percolation threshold in static experiments on synthetic salt samples with x-ray microtomography. We then analyze wells penetrating salt deposits in the Gulf of Mexico. The observed hydrocarbon distributions in rock salt require that percolation occurred at porosities considerably below the static threshold due to deformation-assisted percolation. Therefore, the design of nuclear waste repositories in salt should guard against deformation-driven fluid percolation. In general, static percolation thresholds may not always limit fluid flow in deforming environments. PMID:26612949 6. Application to ship nonmixed transuranic waste to the Nevada Test Site for interim storage. Waste Cerification Program SciTech Connect Not Available 1993-12-01 This report documents various regulations on radioactive waste processing and discusses how the Waste Isolation Pilot Plant will comply with and meet these requirements. Specific procedures are discussed concerning transuranic, metal scrap, salt block, solid, and glove box wastes. 7. Relation between the diffusivity, viscosity, and ionic radius of LiCl in water, methanol, and ethylene glycol: a molecular dynamics simulation. PubMed Kumar, Parveen; Varanasi, Srinivasa R; Yashonath, S 2013-07-11 A molecular dynamics (MD) investigation of LiCl in water, methanol, and ethylene glycol (EG) at 298 K is reported. Several structural and dynamical properties of the ions as well as the solvent such as self-diffusivity, radial distribution functions, void and neck distributions, velocity autocorrelation functions, and mean residence times of solvent in the first solvation shell have been computed. The results show that the reciprocal relationship between the self-diffusivity of the ions and the viscosity is valid in almost all solvents with the exception of water. From an analysis of radial distribution functions and coordination numbers the nature of hydrogen bonding within the solvent and its influence on the void and neck distribution becomes evident. It is seen that the solvent?solvent interaction is important in EG while solute?solvent interactions dominate in water and methanol. From Voronoi tessellation, it is seen that the voids and necks within methanol are larger as compared to those within water or EG. On the basis of the void and neck distributions obtained from MD simulations and literature experimental data of limiting ion conductivity for various ions of different sizes, we show that there is a relation between the void and neck radius on the one hand and dependence of conductivity on the ionic radius on the other. It is shown that the presence of large diameter voids and necks in methanol is responsible for maximum in limiting ion conductivity (?0) of TMA+, while in water and EG, the maximum is seen for Rb+. In the case of monovalent anions, maximum in ?0 as a function ionic radius is seen for Br? in water and EG but for the larger ClO4 ? ion in methanol. The relation between the void and neck distribution and the variation in ?0 with ionic radius arises via the Levitation effect which is discussed. These studies show the importance of the solvent structure and the associated void structure. PMID:23800019 8. Fracture and Healing of Rock Salt Related to Salt Caverns SciTech Connect Chan, K.S.; Fossum, A.F.; Munson, D.E. 1999-03-01 In recent years, serious investigations of potential extension of the useful life of older caverns or of the use of abandoned caverns for waste disposal have been of interest to the technical community. All of the potential applications depend upon understanding the reamer in which older caverns and sealing systems can fail. Such an understanding will require a more detailed knowledge of the fracture of salt than has been necessary to date. Fortunately, the knowledge of the fracture and healing of salt has made significant advances in the last decade, and is in a position to yield meaningful insights to older cavern behavior. In particular, micromechanical mechanisms of fracture and the concept of a fracture mechanism map have been essential guides, as has the utilization of continuum damage mechanics. The Multimechanism Deformation Coupled Fracture (MDCF) model, which is summarized extensively in this work was developed specifically to treat both the creep and fracture of salt, and was later extended to incorporate the fracture healing process known to occur in rock salt. Fracture in salt is based on the formation and evolution of microfractures, which may take the form of wing tip cracks, either in the body or the boundary of the grain. This type of crack deforms under shear to produce a strain, and furthermore, the opening of the wing cracks produce volume strain or dilatancy. In the presence of a confining pressure, microcrack formation may be suppressed, as is often the case for triaxial compression tests or natural underground stress situations. However, if the confining pressure is insufficient to suppress fracture, then the fractures will evolve with time to give the characteristic tertiary creep response. Two first order kinetics processes, closure of cracks and healing of cracks, control the healing process. Significantly, volume strain produced by microfractures may lead to changes in the permeability of the salt, which can become a major concern in cavern sealing and operation. The MDCF model is used in three simulations of field experiments in which indirect measures were obtained of the generation of damage. The results of the simulations help to verify the model and suggest that the model captures the correct fracture behavior of rock salt. The model is used in this work to estimate the generation and location of damage around a cylindrical storage cavern. The results are interesting because stress conditions around the cylindrical cavern do not lead to large amounts of damage. Moreover, the damage is such that general failure can not readily occur, nor does the extent of the damage suggest possible increased permeation when the surrounding salt is impermeable. 9. Pyrochemical investigations into recovering plutonium from americium extraction salt residues SciTech Connect Fife, K.W.; West, M.H. 1987-05-01 Progress into developing a pyrochemical technique for separating and recovering plutonium from spent americium extraction waste salts has concentrated on selective chemical reduction with lanthanum metal and calcium metal and on the solvent extraction of americium with calcium metal. Both techniques are effective for recovering plutonium from the waste salt, although neither appears suitable as a separation technique for recycling a plutonium stream back to mainline purification processes. 17 refs., 13 figs., 2 tabs. 10. Technical review of Molten Salt Oxidation SciTech Connect Not Available 1993-12-01 The process was reviewed for destruction of mixed low-level radioactive waste. Results: extensive development work and scaleup has been documented on coal gasification and hazardous waste which forms a strong experience base for this MSO process; it is clearly applicable to DOE wastes such as organic liquids and low-ash wastes. It also has potential for processing difficult-to-treat wastes such as nuclear grade graphite and TBP, and it may be suitable for other problem waste streams such as sodium metal. MSO operating systems may be constructed in relatively small units for small quantity generators. Public perceptions could be favorable if acceptable performance data are presented fairly; MSO will likely require compliance with regulations for incineration. Use of MSO for offgas treatment may be complicated by salt carryover. Figs, tabs, refs. 11. Ammonia Solubility in High Concentration Salt Solutions SciTech Connect HEDENGREN, D.C. 2000-02-01 Solubility data for ammonia in water and various dilute solutions are abundant in the literature. However, there is a noticeable lack of ammonia solubility data for high salt, basic solutions of various mixtures of salts including those found in many of the Hanford Washington underground waste tanks. As a result, models based on solubility data for dilute salt solutions have been used to extrapolate to high salt solutions. These significant extrapolations need to be checked against actual laboratory data. Some indirect vapor measurements have been made. A more direct approach is to determine the ratio of solubility of ammonia in water to its solubility in high salt solutions. In various experiments, pairs of solutions, one of which is water and the other a high salt solution, are allowed to come to equilibrium with a common ammonia vapor pressure. The ratio of concentrations of ammonia in the two solutions is equal to the ratio of the respective ammonia solubilities (Henry's Law constants) at a given temperature. This information can then be used to refine the models that predict vapor space compositions of ammonia. Ammonia at Hanford is of concern because of its toxicity in the environment and its contribution to the flammability of vapor space gas mixtures in waste tanks. 12. Heterogeneity of Equilibrium Molten Globule State of Cytochrome c Induced by Weak Salt Denaturants under Physiological Condition PubMed Central Rahaman, Hamidur; Alam Khan, Md. Khurshid; Hassan, Md. Imtaiyaz; Islam, Asimul; Moosavi-Movahedi, Ali Akbar; Ahmad, Faizan 2015-01-01 While many proteins are recognized to undergo folding via intermediate(s), the heterogeneity of equilibrium folding intermediate(s) along the folding pathway is less understood. In our present study, FTIR spectroscopy, far- and near-UV circular dichroism (CD), ANS and tryptophan fluorescence, near IR absorbance spectroscopy and dynamic light scattering (DLS) were used to study the structural and thermodynamic characteristics of the native (N), denatured (D) and intermediate state (X) of goat cytochorme c (cyt-c) induced by weak salt denaturants (LiBr, LiCl and LiClO4) at pH 6.0 and 25C. The LiBr-induced denaturation of cyt-c measured by Soret absorption (??400) and CD ([?]409), is a three-step process, N ? X ? D. It is observed that the X state obtained along the denaturation pathway of cyt-c possesses common structural and thermodynamic characteristics of the molten globule (MG) state. The MG state of cyt-c induced by LiBr is compared for its structural and thermodynamic parameters with those found in other solvent conditions such as LiCl, LiClO4 and acidic pH. Our observations suggest: (1) that the LiBr-induced MG state of cyt-c retains the native Met80-Fe(III) axial bond and Trp59-propionate interactions; (2) that LiBr-induced MG state of cyt-c is more compact retaining the hydrophobic interactions in comparison to the MG states induced by LiCl, LiClO4 and 0.5 M NaCl at pH 2.0; and (3) that there exists heterogeneity of equilibrium intermediates along the unfolding pathway of cyt-c as highly ordered (X1), classical (X2) and disordered (X3), i.e., D ? X3 ? X2 ? X1 ? N. PMID:25849212 13. Heterogeneity of equilibrium molten globule state of cytochrome c induced by weak salt denaturants under physiological condition. PubMed Rahaman, Hamidur; Alam Khan, Md Khurshid; Hassan, Md Imtaiyaz; Islam, Asimul; Moosavi-Movahedi, Ali Akbar; Ahmad, Faizan 2015-01-01 While many proteins are recognized to undergo folding via intermediate(s), the heterogeneity of equilibrium folding intermediate(s) along the folding pathway is less understood. In our present study, FTIR spectroscopy, far- and near-UV circular dichroism (CD), ANS and tryptophan fluorescence, near IR absorbance spectroscopy and dynamic light scattering (DLS) were used to study the structural and thermodynamic characteristics of the native (N), denatured (D) and intermediate state (X) of goat cytochorme c (cyt-c) induced by weak salt denaturants (LiBr, LiCl and LiClO4) at pH 6.0 and 25C. The LiBr-induced denaturation of cyt-c measured by Soret absorption (??400) and CD ([?]409), is a three-step process, N ? X ? D. It is observed that the X state obtained along the denaturation pathway of cyt-c possesses common structural and thermodynamic characteristics of the molten globule (MG) state. The MG state of cyt-c induced by LiBr is compared for its structural and thermodynamic parameters with those found in other solvent conditions such as LiCl, LiClO4 and acidic pH. Our observations suggest: (1) that the LiBr-induced MG state of cyt-c retains the native Met80-Fe(III) axial bond and Trp59-propionate interactions; (2) that LiBr-induced MG state of cyt-c is more compact retaining the hydrophobic interactions in comparison to the MG states induced by LiCl, LiClO4 and 0.5 M NaCl at pH 2.0; and (3) that there exists heterogeneity of equilibrium intermediates along the unfolding pathway of cyt-c as highly ordered (X1), classical (X2) and disordered (X3), i.e., D ? X3 ? X2 ? X1 ? N. PMID:25849212 14. Salt repository project closeout status report SciTech Connect 1988-06-01 This report provides an overview of the scope and status of the US Department of Energy (DOEs) Salt Repository Project (SRP) at the time when the project was terminated by the Nuclear Waste Policy Amendments Act of 1987. The report reviews the 10-year program of siting a geologic repository for high-level nuclear waste in rock salt formations. Its purpose is to aid persons interested in the information developed during the course of this effort. Each area is briefly described and the major items of information are noted. This report, the three salt Environmental Assessments, and the Site Characterization Plan are the suggested starting points for any search of the literature and information developed by the program participants. Prior to termination, DOE was preparing to characterize three candidate sites for the first mined geologic repository for the permanent disposal of high-level nuclear waste. The sites were in Nevada, a site in volcanic tuff; Texas, a site in bedded salt (halite); and Washington, a site in basalt. These sites, identified by the screening process described in Chapter 3, were selected from the nine potentially acceptable sites shown on Figure I-1. These sites were identified in accordance with provisions of the Nuclear Waste Policy Act of 1982. 196 refs., 21 figs., 11 tabs. 15. The Effect of Hydrostatic Pressure and Temperature on Sonoluminescence of Metal Atoms from Aqueous Salt Solutions NASA Astrophysics Data System (ADS) Gordeychuk, Tatyana V.; Kazachek, Mikhail V. 2008-06-01 We have measured the spectra of multi-bubble sonoluminescence from aqueous LiCl, NaCl, MgCl2, CaCl2, CuSO4 solutions, saturated with argon, at the frequency 22 kHz. The spectra were collected at different hydrostatic pressures for all salts and in the ambient temperature range of 1-20 C for NaCl solutions. Total absorbed ultrasonic power was 41 W. The results show that line emission from exited state metal atoms grows rapidly with the increase of hydrostatic pressure and at low ambient temperature. In all cases the increase of the intensity of underlying continuum and OH-radical emission was much lower than that of atomic emission at the same experimental conditions. The observed effects can be explained assuming higher temperature inside the cavitation bubbles during collapse in case of increased static pressure or decreased vapor pressure. 16. Infrared spectroscopy of HCN-salt complexes formed in liquid-helium nanodroplets. PubMed Lewis, William K; Miller, Roger E 2006-02-14 Rotationally resolved infrared spectra are reported for the binary complexes of HCN and LiF, LiCl, NaF, and NaCl, formed in helium nanodroplets. Stark spectroscopy is used to determine the dipole moments for these complexes. Ab initio calculations are also reported for these complexes, revealing the existence of several different isomers of these binary systems. In the frequency region examined in this experimental study we only observe one of these, corresponding to the salt binding to the nitrogen end of the HCN molecule. The experimental rotational constants, dipole moments, and vibrational frequency shifts are all compared with the results from ab initio calculations for this isomer. PMID:16483201 17. Mass transfer in a salt repository SciTech Connect Pigford, T.H.; Chambre, P.L. 1985-05-01 To meet regulatory requirements for radioactive waste in a salt repository it is necessary to predict the rates of corrosion of the waste container, the release rates of radionuclides from the waste package, and the cumulative release of radionuclides into the accessible environment. The mechanisms that may control these rates and an approach to predicting these rates from mass-transfer theory are described. This new mechanistic approach is suggested by three premises: (a) a brine inclusion originally in a salt crystal moves along grain boundaries after thermal-induced migration out of the crystal, (b) brine moves along a grain boundary under the influence of a pressure gradient, and (c) salt surrounding a heat-generating waste package will soon creep and consolidate as a monolithic medium surrounding and in contact with the waste package. After consolidation there may be very little migration of intergranular and intragranular brine to the waste package. The corrosion rate of the waste container may then be limited by the rate at which brine reaches the container and may be calculable from mass-transfer theory, and the rate at which dissolved radionuclides leave the waste package may be limited by molecular diffusion in intragranular brine and may be calculable from mass-transfer theory. If porous nonsalt interbeds intersect the waste-package borehole, the release rate of dissolved radionuclides to interbed brine may also be calculable from mass-transfer theory. The logic of these conclusions is described, as an aid in formulating the calculations that are to be made. 18. Characterization of an extremely salt-tolerant and thermostable phytase from Bacillus amyloliquefaciens US573. PubMed Boukhris, Ines; Farhat-Khemakhem, Ameny; Blibech, Monia; Bouchaala, Kameleddine; Chouayekh, Hichem 2015-09-01 The extracellular phytase produced by the Bacillus amyloliquefaciens US573 strain, isolated from geothermal soil located in Southern Tunisia was purified and characterized. This calcium-dependent and bile-stable enzyme (PHY US573) was optimally active at pH 7.5 and 70 °C. It showed a good stability at pH ranging from 4 to 10, and especially, an exceptional thermostability as it recovered 50 and 62% of activity after heating for 10 min at 100 and 90 °C, respectively. In addition, PHY US573 was found to be extremely salt-tolerant since it preserved 80 and 95% of activity in the presence of 20 g/l of NaCl and LiCl, respectively. The gene corresponding to PHY US573 was cloned. It encodes a 383 amino acids polypeptide exhibiting 99% identity with the highly thermostable phytases from Bacillus sp. MD2 and B. amyloliquefaciens DS11 (3 and 5 residues difference, respectively), suggesting the existence of common molecular determinants responsible for their remarkable heat stability. Overall, our findings illustrated that in addition to its high potential for application in feed industry, the salt tolerance of the PHY US573 phytase, may represent an exciting new avenue for improvement of phosphorus-use efficiency of salt-tolerant plants in soils with high salt and phytate content. PMID:26188308 19. Pyrochemical process for extracting plutonium from an electrolyte salt DOEpatents Mullins, L.J.; Christensen, D.C. 1982-09-20 A pyrochemical process for extracting plutonium from a plutonium-bearing salt is disclosed. The process is particularly useful in the recovery of plutonium for electrolyte salts which are left over from the electrorefining of plutonium. In accordance with the process, the plutonium-bearing salt is melted and mixed with metallic calcium. The calcium reduces ionized plutonium in the salt to plutonium metal, and also causes metallic plutonium in the salt, which is typically present as finely dispersed metallic shot, to coalesce. The reduced and coalesced plutonium separates out on the bottom of the reaction vessel as a separate metallic phase which is readily separable from the overlying salt upon cooling of the mixture. Yields of plutonium are typically on the order of 95%. The stripped salt is virtually free of plutonium and may be discarded to low-level waste storage. 20. Pyrochemical process for extracting plutonium from an electrolyte salt DOEpatents Mullins, Lawrence J. (Los Alamos, NM); Christensen, Dana C. (Los Alamos, NM) 1984-01-01 A pyrochemical process for extracting plutonium from a plutonium-bearing salt is disclosed. The process is particularly useful in the recovery of plutonium from electrolyte salts which are left over from the electrorefining of plutonium. In accordance with the process, the plutonium-bearing salt is melted and mixed with metallic calcium. The calcium reduces ionized plutonium in the salt to plutonium metal, and also causes metallic plutonium in the salt, which is typically present as finely dispersed metallic shot, to coalesce. The reduced and coalesced plutonium separates out on the bottom of the reaction vessel as a separate metallic phase which is readily separable from the overlying salt upon cooling of the mixture. Yields of plutonium are typically on the order of 95%. The stripped salt is virtually free of plutonium and may be discarded to low-level waste storage. 1. Salt: a sacred substance. PubMed De Santo, N G; Bisaccia, C; De Santo, R M; De Santo, L S; Petrelli, L; Gallo, L; Cirillo, M; Capasso, G 1997-11-01 Salt is the last relic of the ocean where life was born. Its presence has influenced the whole gamut of history and its name is linked to hundred of geographical locations. Its importance for nutrition is supported by the discovery of Aeneolithic salt cellars. Salt cellars and pyramids of salt have been included in paintings and other works of art. In Japan where salt was and still is obtained from the sea, a salt culture has developed that can be traced in the rituals of everyday life, including meal preparation, sports, and Shinto ceremonies. PMID:9350697 2. Solubility issues related to the performance of a high-level nuclear waste repository in bedded salt: A survey of data needs for solubilities and aqueous species of technetium, radium, and iodine: Nuclear and Chemical Waste Program SciTech Connect O'Kelley, G.D. 1987-07-01 A survey of the literature on the solubility and aqueous species of technetium, radium, and iodine was conducted, with special emphasis on salt repository brines. The elements reviewed here were three of a total of seventeen elements reviewed by participating laboratories. Comprehensive tables of solid compounds and aqueous species for each elements were screened according to a uniform set of procedures. Technetium solid compounds selected for further solubility studies in repository brines were TcO/sub 2//center dot/nH/sub 2/O, TcS/sub 2/, and Tc/sub 2/S/sub 7/. Aqueous species of interest were those associated with solubility measurements and complexing ligands such as phosphate and carbonate species. The least soluble compounds of radium are expected to be RaSO/sub 4/ and Ra/sub 3/(PO/sub 4/)/sub 2/, although thermodynamic estimates predict that RaF/sub 2/ and RaCO/sub 3/ may also be of some concern under particular circumstances. In cases where concentrations are too low to precipitate a pure radium phase, it appears probable that the concentration of Ra/sup 2 +/ in a brine may be limited by formation of solid solutions with insoluble compounds of other Group II A elements. The aqueous species of most concern are those associated with the least soluble solid compounds of radium, e.g., phosphate, sulfate, fluoride, and carbonate. No solid species of iodine were identified as important in repository brines. The species of iodine of major interest are those involved in the hydrolysis of I/sub 2/ to I/sup /minus// and IO/sub 3//sup /minus//, including trihalide ions and HOI. 40 refs., 5 figs., 21 tabs. 3. Removal of uranium from spent salt from the moltensalt oxidation process SciTech Connect Summers, L.; Hsu, P. C.; Holtz, E. V.; Hipple, D.; Wang, F.; Adamson, M. 1997-03-01 Molten salt oxidation (MSO) is a thermal process that has the capability of destroying organic constituents of mixed wastes, hazardous wastes, and energetic materials. In this process, combustible waste and air are introduced into the molten sodium carbonate salt. The organic constituents of the waste materials are oxidized to carbon dioxide and water, while most of the inorganic constituents, including toxic metals, minerals, and radioisotopes, are retained in the molten salt bath. As these impurities accumulate in the salt, the process efficiency drops and the salt must be replaced. An efficient process is needed to separate these toxic metals, minerals, and radioisotopes from the spent carbonate to avoid generating a large volume of secondary waste. Toxic metals such as cadmium, chromium, lead, and zinc etc. are removed by a method described elsewhere. This paper describes a separation strategy developed for radioisotope removal from the mixed spent salt, as well as experimental results, as part of the spent salt cleanup. As the MSO system operates, inorganic products resulting from the reaction of halides, sulfides, phosphates, metals and radionuclides with carbonate accumulate in the salt bath. These must be removed to prevent complete conversion of the sodium carbonate, which would result in eventual losses of destruction efficiency and acid scrubbing capability. There are two operational modes for salt removal: (1) during reactor operation a slip-stream of molten salt is continuously withdrawn with continuous replacement by carbonate, or (2) the spent salt melt is discharged completely and the reactor then refilled with carbonate in batch mode. Because many of the metals and/or radionuclides captured in the salt are hazardous and/or radioactive, spent salt removed from the reactor would create a large secondary waste stream without further treatment. A spent salt clean up/recovery system is necessary to segregate these materials and minimize the amount of secondary waste. These materials can then be encapsulated for final disposal. 4. Salt tectonics on Venus SciTech Connect Wood, C.A.; Amsbury, D. 1986-05-01 The discovery of a surprisingly high deuterium/hydrogen ratio on Venus immediately led to the speculation that Venus may have once had a volume of surface water comparable to that of the terrestrial oceans. The authors propose that the evaporation of this putative ocean may have yielded residual salt deposits that formed various terrain features depicted in Venera 15 and 16 radar images. By analogy with models for the total evaporation of the terrestrial oceans, evaporite deposits on Venus should be at least tens to hundreds of meters thick. From photogeologic evidence and in-situ chemical analyses, it appears that the salt plains were later buried by lava flows. On Earth, salt diapirism leads to the formation of salt domes, anticlines, and elongated salt intrusions - features having dimensions of roughly 1 to 100 km. Due to the rapid erosion of salt by water, surface evaporite landforms are only common in dry regions such as the Zagros Mountains of Iran, where salt plugs and glaciers exist. Venus is far drier than Iran; extruded salt should be preserved, although the high surface temperature (470/sup 0/C) would probably stimulate rapid salt flow. Venus possesses a variety of circular landforms, tens to hundreds of kilometers wide, which could be either megasalt domes or salt intrusions colonizing impact craters. Additionally, arcurate bands seen in the Maxwell area of Venus could be salt intrusions formed in a region of tectonic stress. These large structures may not be salt features; nonetheless, salt features should exist on Venus. 5. Molten salt electrolyte separator DOEpatents Kaun, Thomas D. (New Lenox, IL) 1996-01-01 A molten salt electrolyte/separator for battery and related electrochemical systems including a molten electrolyte composition and an electrically insulating solid salt dispersed therein, to provide improved performance at higher current densities and alternate designs through ease of fabrication. 6. What Are Bath Salts? MedlinePLUS ... Are bath salts becoming more popular? Marsha Lopez Hi, Lauren. Nope! Actually quite the opposite! This family ... and how dangerous for your body? Michelle Rankin Hi ParkerPanella - Bath salts are drugs known as synthetic ... 7. Retrospective salt tectonics SciTech Connect Jackson, M.P.A. 1996-12-31 The conceptual breakthroughs in understanding salt tectonics can be recognized by reviewing the history of salt tectonics, which divides naturally into three parts: the pioneering era, the fluid era, and the brittle era. The pioneering era (1856-1933) featured the search for a general hypothesis of salt diapirism, initially dominated by bizarre, erroneous notions of igneous activity, residual islands, in situ crystallization, osmotic pressures, and expansive crystallization. Gradually data from oil exploration constrained speculation. The effects of buoyancy versus orogeny were debated, contact relations were characterized, salt glaciers were discovered, and the concepts of downbuilding and differential loading were proposed as diapiric mechanisms. The fluid era (1933-{approximately}1989) was dominated by the view that salt tectonics resulted from Rayleigh-Taylor instabilities in which a dense fluid overburden having negligible yield strength sinks into a less dense fluid salt layer, displacing it upward. Density contrasts, viscosity contrasts, and dominant wavelengths were emphasized, whereas strength and faulting of the overburden were ignored. During this era, palinspastic reconstructions were attempted; salt upwelling below thin overburdens was recognized; internal structures of mined diapirs were discovered; peripheral sinks, turtle structures, and diapir families were comprehended; flow laws for dry salt were formulated; and contractional belts on divergent margins and allochthonous salt sheets were recognized. The 1970s revealed the basic driving force of salt allochthons, intrasalt minibasins, finite strains in diapirs, the possibility of thermal convection in salt, direct measurement of salt glacial flow stimulated by rainfall, and the internal structure of convecting evaporites and salt glaciers. The 1980s revealed salt rollers, subtle traps, flow laws for damp salt, salt canopies, and mushroom diapirs. 8. Salt weathering on Mars NASA Technical Reports Server (NTRS) Malin, M. C. 1974-01-01 Mariner 9 photographs of Mars indicate that significant erosion has occurred on that planet. Although several possible erosion mechanisms have been proposed, most terrestrial weathering mechanisms cannot function in the present Martian environment. Salt weathering, believed to be active in the Antarctic dry valleys, is especially suited to Mars, given the presence of salts and small amounts of water. Volcanic salts are probably available, and the association of salts and water is likely from both thermodynamic and geologic considerations. 9. Brines formed by multi-salt deliquescence SciTech Connect Carroll, S; Rard, J; Alai, M; Staggs, K 2005-11-04 The FY05 Waste Package Environment testing program at Lawrence Livermore National Laboratory focused on determining the temperature, relative humidity, and solution compositions of brines formed due to the deliquescence of NaCl-KNO{sub 3}-NaNO{sub 3} and NaCl-KNO{sub 3}-NaNO{sub 3}-Ca(NO{sub 3}){sub 2} salt mixtures. Understanding the physical and chemical behavior of these brines is important because they define conditions under which brines may react with waste canister surfaces. Boiling point experiments show that NaCl-KNO{sub 3}-NaNO{sub 3} and NaCl-KNO{sub 3}-NaNO{sub 3}-Ca(NO{sub 3}){sub 2} salt mixtures form brines that transform to hydrous melts that do not truly 'dry out' until temperatures exceed 300 and 400 C, respectively. Thus a conducting solution is present for these salt assemblages over the thermal history of the repository. The corresponding brines form at lower relative humidity at higher temperatures. The NaCl-KNO{sub 3}-NaNO{sub 3} salt mixture has a mutual deliquescence relative humidity (MDRH) of 25.9% at 120 C and 10.8% at 180 C. Similarly, the KNO{sub 3}-NaNO{sub 3} salt mixture has MDRH of 26.4% at 120 C and 20.0% at 150 C. The KNO{sub 3}-NaNO{sub 3} salt mixture salts also absorb some water (but do not appear to deliquesce) at 180 C and thus may also contribute to the transfer of electrons at interface between dust and the waste package surface. There is no experimental evidence to suggest that these brines will degas and form less deliquescent salt assemblages. Ammonium present in atmospheric and tunnel dust (as the chloride, nitrate, or sulfate) will readily decompose in the initial heating phase of the repository, and will affect subsequent behavior of the remaining salt mixture only through the removal of a stoichiometric equivalent of one or more anions. Although K-Na-NO{sub 3}-Cl brines form at high temperature and low relative humidity, these brines are dominated by nitrate, which is known to inhibit corrosion at lower temperature. Nitrate to chloride ratios of the NaCl-KNO{sub 3}-NaNO{sub 3} salt mixture are about NO{sub 3}:Cl = 19:1. The role of nitrate on corrosion at higher temperatures is addressed in a companion report (Dixit et al., 2005). 10. Utah: Salt Lake Region Atmospheric Science Data Center 2014-05-15 article title: Winter and Summer Views of the Salt Lake Region ... Salt Lake City, Utah are captured in these winter and summer images from the Multi-angle Imaging SpectroRadiometer (MISR) ... the obvious difference in snow cover between the winter and summer views, water color changes in parts of the Great Salt Lake are apparent ... 11. 21 CFR 100.155 - Salt and iodized salt. Code of Federal Regulations, 2011 CFR 2011-04-01 ... 21 Food and Drugs 2 2011-04-01 2011-04-01 false Salt and iodized salt. 100.155 Section 100.155... FOR HUMAN CONSUMPTION GENERAL Specific Administrative Rulings and Decisions 100.155 Salt and iodized salt. (a) For the purposes of this section, the term iodized salt or iodized table salt is... 12. 21 CFR 100.155 - Salt and iodized salt. Code of Federal Regulations, 2010 CFR 2010-04-01 ... 21 Food and Drugs 2 2010-04-01 2010-04-01 false Salt and iodized salt. 100.155 Section 100.155... FOR HUMAN CONSUMPTION GENERAL Specific Administrative Rulings and Decisions § 100.155 Salt and iodized salt. (a) For the purposes of this section, the term iodized salt or iodized table salt is... 13. 21 CFR 100.155 - Salt and iodized salt. Code of Federal Regulations, 2014 CFR 2014-04-01 ... 21 Food and Drugs 2 2014-04-01 2014-04-01 false Salt and iodized salt. 100.155 Section 100.155... FOR HUMAN CONSUMPTION GENERAL Specific Administrative Rulings and Decisions 100.155 Salt and iodized salt. (a) For the purposes of this section, the term iodized salt or iodized table salt is... 14. Integrated Salt Basin Evaluation NASA Astrophysics Data System (ADS) Kukla, P. A. 2012-04-01 Salt tectonics plays a major role in the development of many sedimentary basins. Basins containing salt thus frequently display a complex geodynamic evolution characterized by several phases of halokinesis and associated sedimentation. One classic area of salt tectonics is the Central European Basin System (CEBS). Here, the mobile Permian Zechstein salt formed a large number of salt structures such as anticlines, diapirs, pillows, sheets, stocks, and walls during an extended period of salt tectonic activity in Mesozoic and Cenozoic times. Major changes in sedimentation patterns and structural regimes are associated and common in this setting. Increasingly complex subsurface evaluation therefore requires an approach to study salt basins including analogue and numerical models, field studies and laboratory studies which combine seismic, structural and sedimentary studies with analysis of rheological properties, and geomechanic modelling. This concept can be demonstrated using case studies from Permian Salt Basins in Europe and the Late Neoproterozoic to Early Cambrian South Oman Salt Basin. There salt-influenced sedimentary responses to renewed phases of tectonism can be clearly discerned from detailed sequence analysis based on seismic and log data combined with retrodeformation modelling studies. High quality 3-D seismic data integrated with structural modelling improves the definition of the internal dynamics of salt structures and associated sediment architecture in salt-controlled sequences. Paleo-caprocks inside the diapirs point to long phases of dissolution. Salt wedges formed by extrusion and lateral flow of salt glaciers during periods of diapir emergence and reduced sediment accumulation can be accurately modelled. Although salt is widely regarded as a perfect seal, it can become permeable for one- or two-phase fluids under certain conditions of fluid pressure, temperature and deviatoric stress. The fluid pathways can be either along zones of diffuse grain boundary dilatancy, or along open fractures, depending on the fluid overpressure and deviatoric stress. 15. Impulse radar scanning of intact salt at the Avery Island Mine SciTech Connect Cook, C.W. 1980-05-01 A series of experiments was run in the Avery Island Mine to evaluate the capability of an impulse radar to locate anomalies and simulated waste targets in intact dome salt. Voids in salt were difficult to detect. On the positive side, metal targets and simulated waste (glass) were easily located in intact salt. Radar scanning at ranges of greater than 25 meters and short-range resolution of target positions to within a few centimeters were achieved. 16. The salts of Mars NASA Astrophysics Data System (ADS) Clark, B. C.; van Hart, D. C. 1981-02-01 Salt compounds are apparently an important component of the fine-grained regolith on Mars. Salt enrichment may be explained either as a secondary concentration of chemical weathering products or as direct incorporation of planetary released volatiles. Geochemical measurements and chemical relationships constrain the salt species and resultant physicochemical consequences. A likely assemblage is dominated by (Mg,Na)SO4, NaCl, and (Mg,Ca)CO3. Formation of brine in equilibrium with such a salt mixture is unlikely under the temperature and water-vapor restrictions prevalent over most, if not all, of the Martian surface. Acidic conditions, accompanying salt formation, favor the preferential destruction of susceptible igneous minerals. 17. Correlation of Creep Behavior of Domal Salts SciTech Connect Munson, D.E. 1999-02-16 The experimentally determined creep responses of a number of domal salts have been reported in, the literature. Some of these creep results were obtained using standard (conventional) creep tests. However, more typically, the creep data have come from multistage creep tests, where the number of specimens available for testing was small. An incremental test uses abrupt changes in stress and temperature to produce several time increments (stages) of different creep conditions. Clearly, the ability to analyze these limited data and to correlate them with each other could be of considerable potential value in establishing the mechanical characteristics of salt domes, both generally and specifically. In any analysis, it is necessary to have a framework of rules to provide consistency. The basis for the framework is the Multimechanism-Deformation (M-D) constitutive model. This model utilizes considerable general knowledge of material creep deformation to supplement specific knowledge of the material response of salt. Because the creep of salt is controlled by just a few micromechanical mechanisms, regardless of the origin of the salt, certain of the material parameters are values that can be considered universal to salt. Actual data analysis utilizes the methodology developed for the Waste Isolation Pilot Plant (WIPP) program, and the response of a bedded pure WIPP salt as the baseline for comparison of the domal salts. Creep data from Weeks Island, Bryan Mound, West Hackberry, Bayou Choctaw, and Big Hill salt domes, which are all sites of Strategic Petroleum Reserve (SPR) storage caverns, were analyzed, as were data from the Avery Island, Moss Bluff, and Jennings salt domes. The analysis permits the parameter value sets for the domal salts to be determined in terms of the M-D model with various degrees of completeness. In turn this permits detailed numerical calculations simulating cavern response. Where the set is incomplete because of the sparse database, reasonable assumptions permit the set to be completed. From the analysis, two distinct response groups were evident, with the salts of one group measurably more creep resistant than the other group. Interestingly, these groups correspond well with the indirectly determined creep closure of the SPR storage caverns, a correlation that probably should be expected. Certainly, the results suggest a simple laboratory determination of the creep characteristics of a salt material from a dome site can indicate the relative behavior of any potential cavern placed within that dome. 18. Waste Encapsulation and Storage Facility (WESF) Waste Analysis Plan SciTech Connect SIMMONS, F.M. 2000-12-01 The purpose of this waste analysis plan (WAP) is to document waste analysis activities associated with the Waste Encapsulation and Storage Facility (WESF) to comply with Washington Administrative Code (WAC) 173-303-300(1), (2), (3), (4), (5), and (6). WESF is an interim status other storage-miscellaneous storage unit. WESF stores mixed waste consisting of radioactive cesium and strontium salts. WESF is located in the 200 East Area on the Hanford Facility. Because dangerous waste does not include source, special nuclear, and by-product material components of mixed waste, radionuclides are not within the scope of this documentation. The information on radionuclides is provided only for general knowledge. 19. Large-scale dynamic compaction of natural salt SciTech Connect Hansen, F.D.; Ahrens, E.H. 1996-05-01 A large-scale dynamic compaction demonstration of natural salt was successfully completed. About 40 m{sup 3} of salt were compacted in three, 2-m lifts by dropping a 9,000-kg weight from a height of 15 m in a systematic pattern to achieve desired compaction energy. To enhance compaction, 1 wt% water was added to the relatively dry mine-run salt. The average compacted mass fractional density was 0.90 of natural intact salt, and in situ nitrogen permeabilities averaged 9X10{sup -14}m{sup 2}. This established viability of dynamic compacting for placing salt shaft seal components. The demonstration also provided compacted salt parameters needed for shaft seal system design and performance assessments of the Waste Isolation Pilot Plant. 20. Molten salts and nuclear energy production NASA Astrophysics Data System (ADS) Le Brun, Christian 2007-01-01 Molten salts (fluorides or chlorides) were considered near the beginning of research into nuclear energy production. This was initially due to their advantageous physical and chemical properties: good heat transfer capacity, radiation insensitivity, high boiling point, wide range solubility for actinides. In addition it was realised that molten salts could be used in numerous situations: high temperature heat transfer, core coolants with solid fuels, liquid fuel in a molten salt reactor, solvents for spent nuclear solid fuel in the case of pyro-reprocessing and coolant and tritium production in the case of fusion. Molten salt reactors, one of the six innovative concepts chosen by the Generation IV international forum, are particularly interesting for use as either waste incinerators or thorium cycle systems. As the neutron balance in the thorium cycle is very tight, the possibility to perform online extraction of some fission product poisons from the salt is very attractive. In this article the most important questions that must be addressed to demonstrate the feasibility of molten salt reactor will be reviewed. 1. Process to separate transuranic elements from nuclear waste DOEpatents Johnson, T.R.; Ackerman, J.P.; Tomczuk, Z.; Fischer, D.F. 1988-07-12 A process for removing transuranic elements from a waste chloride electrolytic salt containing transuranic elements in addition to rare earth and other fission product elements so the salt waste may be disposed of more easily and the valuable transuranic elements may be recovered for reuse. The salt is contacted with a cadmium-uranium alloy which selectively extracts the transuranic elements from the salt. The waste salt is generated during the reprocessing of nuclear fuel associated with the Integral Fast Reactor (IFR). 2 figs. 2. Process to separate transuranic elements from nuclear waste DOEpatents Johnson, T.R.; Ackerman, J.P.; Tomczuk, Z.; Fischer, D.F. 1989-03-21 A process is described for removing transuranic elements from a waste chloride electrolytic salt containing transuranic elements in addition to rare earth and other fission product elements so the salt waste may be disposed of more easily and the valuable transuranic elements may be recovered for reuse. The salt is contacted with a cadmium-uranium alloy which selectively extracts the transuranic elements from the salt. The waste salt is generated during the reprocessing of nuclear fuel associated with the Integral Fast Reactor (IFR). 2 figs. 3. Effect of salts on the electrospinning of poly(vinyl alcohol) NASA Astrophysics Data System (ADS) Stanger, Jonathan J.; Tucker, Nick; Staiger, Mark; Kirwan, Kerry; Coles, Stuart; Jacobs, Daniel; Larsen, Nigel 2009-07-01 Fibres with a diameter in the nanometer range were electrospun from aqueous poly(vinyl alcohol) (PVOH). In order to improve the mass deposition rate and decrease the final fibre diameter salts (NaCl, LiCl, LiBr and LiF) were added to the solution. The aim was to increase the charge density and hence increase the electrostatic forces on the fluid. It was found that with increasing salt concentration the charge density did increase. However the mass deposition rate was found to decrease and the final fibre diameter was found to increase. The decrease in mass deposition rate is explained by considering the concept of a virtual orifice. The increase in the final fibre diameter is explained by considering the charge distribution in the jet when it behaves like a conductor compared to when it behaves like an insulator. Both mechanisms result from the increase in conductivity of the PVOH solution without significantly modifying other solution properties when salt is added. 4. Specific secretion of proline-rich proteins by salt-adapted winged bean cells. PubMed Esaka, M; Hayakawa, H 1995-04-01 Six proteins, designated SAP1 through SAP6, were secreted specifically by salt-adapted cells of winged bean (Psophocarpus tetragonolobus) in suspension cultures. The amino-terminal amino acid sequences of SAP2 (57 kDa), SAP4 (21 kDa), SAP5 (19 kDa) and SAP6 (17 kDa) were homologous to the sequences of proline-rich proteins, indicating that proline-rich proteins are secreted specifically by these salt-adapted cells. In addition, the amino-terminal amino acid sequence of SAP2 was identical to that of SAP4, and the amino-terminal sequence of SAP5 was identical to that of SAP6. Secretion of SAP2 was significantly enhanced by addition of AlCl3 but not of KCl, LiCl, CaCl2, MgCl2, mannitol or sucrose to suspension cultures. Furthermore, secretion of SAP4, SAP5 and SAP6 was stimulated by addition of abscisic acid to cultures, suggesting that these proteins might be secreted in response to salt or osmotic stress. PMID:7757337 5. Effect of salts on the water sorption kinetics of dried pasta. PubMed Ogawa, Takenobu; Adachi, Shuji 2013-01-01 The water sorption kinetics of dried pasta were measured in the 20-90 C range in 1.83 mol/L of NaCl and at 80 C in 1.83 mol/L of LiCl, KCl, NaBr and NaI solutions in order to elucidate the role of salt in the kinetics. At the temperatures higher than 70.8 C, the change in the enthalpy of sorption, ?H, in the 1.83 mol/L NaCl solution was 33.1 kJ/mol, which was greater than the ?H value in water, and the activation energy for the sorption, E, in the salt solution was 25.6 kJ/mol, which was slightly lower than the E value in water. The Hofmeister series of ions was an index for their effect on the equilibrium amount of the sorbed solution of pasta. The apparent diffusion coefficient of water into pasta was not correlated with the crystal radius of the salts, but was with the Stokes radius of the hydrated ions. Equations were formulated to predict the amount of sorbed solution under any condition of temperature and NaCl concentration. PMID:23391911 6. Microstructural observations of reconsolidated granular salt to 250C NASA Astrophysics Data System (ADS) Mills, M. M.; Hansen, F.; Bauer, S. J.; Stormont, J. 2014-12-01 Very low permeability is a principal reason salt formations are considered viable hosts for disposal of nuclear waste and spent nuclear fuel. Granular salt is likely to be used as back-fill material and as a seal system component. Granular salt is expected to reconsolidate to a low permeability condition because of external pressure from the surrounding salt formation. Understanding the consolidation processes--known to depend on the stress state, moisture availability and temperature--is important for predicting achievement of sealing functions and long-term repository performance. As granular salt consolidates, initial void reduction is accomplished by brittle processes of grain rearrangement and cataclastic flow. At porosities of less than 10%, grain boundary processes and crystal-plastic mechanisms govern further porosity reduction. We investigate the micro-mechanisms operative in granular salt that has been consolidated under high temperatures to relatively low porosity. These conditions would occur proximal to heat-generating canisters. Mine-run salt from the Waste Isolation Pilot Plant was used to create cylindrical samples which were consolidated at 250C and stresses to 20 MPa. From samples consolidated to fractional densities of 86% and 97% polished thin sections, etched cleavage chips, and fragments were fabricated. Microstructural techniques included scanning electron and optical microscopy. Microstructure of undeformed mine-run salt was compared to the deformed granular salt. Observed deformation mechanisms include glide, cross slip, climb, fluid-assisted creep, pressure-solution redeposition, and annealing. Documentation of operative deformation mechanisms within the consolidating granular salt, particularly at grain boundaries, is essential to establish effects of moisture, stress, and temperature. Future work will include characterization of pore structures. Information gleaned in these studies supports evaluation of a constitutive model for reconsolidating granular salt, which will be used to predict the thermal-mechanical-hydrologic response of salt repository seal structures and backfilled rooms. 7. Photochemistry of triarylsulfonium salts SciTech Connect Dektar, J.L.; Hacker, N.P. ) 1990-08-01 The photolysis of triphenylsulfonium, tris(4-methylphenyl)sulfonium, tris(4-chlorophenyl)sulfonium, several monosubstituted (4-F, 4-Cl, 4-Me, 4-MeO, 4-PhS, and 4-PhCO), and disubstituted (4,4{prime}-Me{sub 2} and 4,4{prime}-(MeO){sub 2}) triphenylsulfonium salts was examined in solution. It was found that direct irradiation of triphenylsulfonium salts produced new rearrangement products, phenylthiobiphenyls, along with diphenyl sulfide, which had been previously reported. Similarly, the triarylsulfonium salts, with the exception of the (4-(phenylthio)phenyl)diphenylsulfonium salts produced new rearrangement products, phenylthiobiphenyls, along with diphenyl sulfide, which had been previously reported. Similarly, the triarylsulfonium salts, with the exception of the (4-(phenylthio)phenyl)diphenylsulfonium salts, gave the new rearrangement products. The mechanism for direct photolysis is proposed to occur from the singlet excited states to give a predominant heterolytic cleavage along with some homolytic cleavage. 8. Preconceptual design of a salt splitting process using ceramic membranes SciTech Connect Kurath, D.E.; Brooks, K.P.; Hollenberg, G.W.; Clemmer, R.; Balagopal, S.; Landro, T.; Sutija, D.P. 1997-01-01 Inorganic ceramic membranes for salt splitting of radioactively contaminated sodium salt solutions are being developed for treating U. S. Department of Energy tank wastes. The process consists of electrochemical separation of sodium ions from the salt solution using sodium (Na) Super Ion Conductors (NaSICON) membranes. The primary NaSICON compositions being investigated are based on rare- earth ions (RE-NaSICON). Potential applications include: caustic recycling for sludge leaching, regenerating ion exchange resins, inhibiting corrosion in carbon-steel tanks, or retrieving tank wastes; reducing the volume of low-level wastes volume to be disposed of; adjusting pH and reducing competing cations to enhance cesium ion exchange processes; reducing sodium in high-level-waste sludges; and removing sodium from acidic wastes to facilitate calcining. These applications encompass wastes stored at the Hanford, Savannah River, and Idaho National Engineering Laboratory sites. The overall project objective is to supply a salt splitting process unit that impacts the waste treatment and disposal flowsheets and meets user requirements. The potential flowsheet impacts include improving the efficiency of the waste pretreatment processes, reducing volume, and increasing the quality of the final waste disposal forms. Meeting user requirements implies developing the technology to the point where it is available as standard equipment with predictable and reliable performance. This report presents two preconceptual designs for a full-scale salt splitting process based on the RE-NaSICON membranes to distinguish critical items for testing and to provide a vision that site users can evaluate. 9. CRITICALITY SAFETY OF PROCESSING SALT SOLUTION AT SRS SciTech Connect Stephens, K; Davoud Eghbali, D; Michelle Abney, M 2008-01-15 High level radioactive liquid waste generated as a result of the production of nuclear material for the United States defense program at the Savannah River Site has been stored as 36 million gallons in underground tanks. About ten percent of the waste volume is sludge, composed of insoluble metal hydroxides primarily hydroxides of Mn, Fe, Al, Hg, and most radionuclides including fission products. The remaining ninety percent of the waste volume is saltcake, composed of primarily sodium (nitrites, nitrates, and aluminates) and hydroxides. Saltcakes account for 30% of the radioactivity while the sludge accounts for 70% of the radioactivity. A pilot plant salt disposition processing system has been designed at the Savannah River Site for interim processing of salt solution and is composed of two facilities: the Actinide Removal Process Facility (ARPF) and the Modular Caustic Side Solvent Extraction Unit (MCU). Data from the pilot plant salt processing system will be used for future processing salt at a much higher rate in a new salt processing facility. Saltcake contains significant amounts of actinides, and other long-lived radioactive nuclides such as strontium and cesium that must be extracted prior to disposal as low level waste. The extracted radioactive nuclides will be mixed with the sludge from waste tanks and vitrified in another facility. Because of the presence of highly enriched uranium in the saltcake, there is a criticality concern associated with concentration and/or accumulation of fissionable material in the ARP and MCU. 10. Characterization of salt cake from secondary aluminum production. PubMed Huang, Xiao-Lan; Badawy, Amro El; Arambewela, Mahendranath; Ford, Robert; Barlaz, Morton; Tolaymat, Thabet 2014-05-30 Salt cake is a major waste component generated from the recycling of secondary aluminum processing (SAP) waste. Worldwide, the aluminum industry produces nearly 5 million tons of waste annually and the end-of-life management of these wastes is becoming a challenge in the U.S. and elsewhere. In this study, the mineral phases, metal content and metal leachability of 39 SAP waste salt cake samples collected from 10 different facilities across the U.S. were determined. The results showed that aluminum (Al), aluminum oxide, aluminum nitride and its oxides, spinel and elpasolite are the dominant aluminum mineral phases in salt cake. The average total Al content was 14% (w/w). The overall percentage of the total leachable Al in salt cake was 0.6% with approximately 80% of the samples leaching at a level less than 1% of the total aluminum content. The extracted trace metal concentrations in deionized water were relatively low (μgL(-1) level). The toxicity characteristic leaching procedure (TCLP) was employed to further evaluate leachability and the results indicated that the leached concentrations of toxic metals from salt cake were much lower than the EPA toxicity limit set by USEPA. PMID:24747373 11. Fluoridated salt in France. PubMed Hescot, P; Roland, E; Desfontaine, J 1995-07-01 Since November 28, 1985, fluoridated salt at a concentration of 250 mg KF per Kg has been authorized in food in France. It was introduced in the market at the end of 1986 and represented, in 1993, more than 60% of the market of salt. The effect of this Public Health measure will be shown. The efficiency is evaluated in epidemiological studies concerning DMF in the whole of France and in terms of safety by measurement of urinary excretion of fluoride in children who consume potassium fluoride in the salt and fluoride tablets. Last, the prevention policy in France will be described, including the use of fluoridated salt. PMID:7546136 12. Dosimetry using silver salts DOEpatents Warner, Benjamin P. (Los Alamos, NM) 2003-06-24 The present invention provides a method for detecting ionizing radiation. Exposure of silver salt AgX to ionizing radiation results in the partial reduction of the salt to a mixture of silver salt and silver metal. The mixture is further reduced by a reducing agent, which causes the production of acid (HX) and the oxidized form of the reducing agent (R). Detection of HX indicates that the silver salt has been exposed to ionizing radiation. The oxidized form of the reducing agent (R) may also be detected. The invention also includes dosimeters employing the above method for detecting ionizing radiation. 13. Brine flow in heated geologic salt. SciTech Connect Kuhlman, Kristopher L.; Malama, Bwalya 2013-03-01 This report is a summary of the physical processes, primary governing equations, solution approaches, and historic testing related to brine migration in geologic salt. Although most information presented in this report is not new, we synthesize a large amount of material scattered across dozens of laboratory reports, journal papers, conference proceedings, and textbooks. We present a mathematical description of the governing brine flow mechanisms in geologic salt. We outline the general coupled thermal, multi-phase hydrologic, and mechanical processes. We derive these processes' governing equations, which can be used to predict brine flow. These equations are valid under a wide variety of conditions applicable to radioactive waste disposal in rooms and boreholes excavated into geologic salt. 14. Catalytic oxidation of waste materials NASA Technical Reports Server (NTRS) Jagow, R. B. 1977-01-01 Aqueous stream of human waste is mixed with soluble ruthenium salts and is introduced into reactor at temperature where ruthenium black catalyst forms on internal surfaces of reactor. This provides catalytically active surface to convert oxidizable wastes into breakdown products such as water and carbon dioxide. 15. Laboratory investigation of crushed salt consolidation and fracture healing SciTech Connect Not Available 1987-01-01 A laboratory test program was conducted to investigate the consolidation behavior of crushed salt and fracture healing in natural and artificial salt. Crushed salt is proposed for use as backfill in a nuclear waste repository in salt. Artificial block salt is proposed for use in sealing a repository. Four consolidation tests were conducted in a hydrostatic pressure vessel at a maximum pressure of 2500 psi (17.2 MPa) and at room temperature. Three 1-month tests were conducted on salt obtained from the Waste Isolation Pilot Plant and one 2-month test was conducted on salt from Avery Island. Permeability was obtained using argon and either a steady-state or transient method. Initial porosities ranged from 0.26 to 0.36 and initial permeabilities from 2000 to 50,000 md. Final porosities and permeabilities ranged from 0.05 to 0.19 and from <10/sup -5/ md to 110 md, respectively. The lowest final porosity (0.05) and permeability (<10/sup -5/ md) were obtained in a 1-month test in which 2.3% moisture was added to the salt at the beginning of the test. The consolidation rate was much more rapid than in any of the dry salt tests. The fracture healing program included 20 permeability tests conducted on fractured and unfractured samples. The tests were conducted in a Hoek cell at hydrostatic pressures up to 3000 psi (20.6 MPa) with durations up to 8 days. For the natural rock salt tested, permeability was strongly dependent on confining pressure and time. The effect of confining pressure was much weaker in the artificial salt. In most cases the combined effects of time and pressure were to reduce the permeability of fractured samples to the same order of magnitude (or less) as the permeability measured prior to fracturing. 16. Salt Repository Project transportation program plan SciTech Connect Fisher, R.L.; Greenberg, A.H.; Anderson, T.L.; Yates, K.R. 1987-11-02 The Salt Repository Project (SRP) has the responsibility to develop a comprehensive transportation program plan (TrPP) that treats the transportation of workers, supplies, and high-level radioactive waste to the site and the transportation of salt, low-level, and transuranic wastes from the site. The TrPP has developed a systematic approach to transportation which is directed towards satisfying statutes, regulations, and directives and is guided by a hierarchy of specific functional requirements, strategies, plans, and reports. The TrPP identifies and develops the planning process for transportation-related studies and provides guidance to staff in performing and documenting these activities. The TrPP also includes an explanation of the responsibilities of the organizational elements involved in these transportation studies. Several of the report chapters relate to identifying routes for transporting nuclear waste to the site. These include a chapter on identifying an access corridor for a new rail route leading to the site, identifying and evaluating emergency-response preparedness capabilities along candidate routes in the state, and identifying alternative routes from the state border, ports, or in-state reactors to the site. The TrPP also includes plans for identifying salt disposal routes and a discussion of repository/transportation interface requirements. 89 refs., 6 figs. 17. Reference repository design concept for bedded salt SciTech Connect Carpenter, D.W.; Martin, R.W. 1980-10-08 A reference design concept is presented for the subsurface portions of a nuclear waste repository in bedded salt. General geologic, geotechnical, hydrologic and geochemical data as well as descriptions of the physical systems are provided for use on generic analyses of the pre- and post-sealing performance of repositories in this geologic medium. The geology of bedded salt deposits and the regional and repository horizon stratigraphy are discussed. Structural features of salt beds including discontinuities and dissolution features are presented and their effect on repository performance is discussed. Seismic hazards and the potential effects of earthquakes on underground repositories are presented. The effect on structural stability and worker safety during construction from hydrocarbon and inorganic gases is described. Geohydrologic considerations including regional hydrology, repository scale hydrology and several hydrological failure modes are presented in detail as well as the hydrological considerations that effect repository design. Operational phase performance is discussed with respect to operations, ventilation system, shaft conveyances, waste handling and retrieval systems and receival rates of nuclear waste. Performance analysis of the post sealing period of a nuclear repository is discussed, and parameters to be used in such an analysis are presented along with regulatory constraints. Some judgements are made regarding hydrologic failure scenarios. Finally, the design and licensing process, consistent with the current licensing procedure is described in a format that can be easily understood. 18. New public information resources on salt caverns. SciTech Connect Tomasko, D.; Veil, J. A. 1999-08-25 For the past decade, interest has been growing in using underground salt caverns for disposing of wastes. The Railroad Commission of Texas has permitted a few caverns for disposal of nonhazardous oil field waste (NOW) and one cavern for disposal of naturally occurring radioactive materials (NORM) from oil field activities. Several salt caverns in Canada have also been permitted for disposal of NOW. In addition, oil and gas agencies in Louisiana and New Mexico are developing cavern disposal regulations. The US Department of Energy (DOE) has funded several studies to evaluate the technical feasibility, legality, economic viability, and risk of disposing of NOW and NORM in caverns. The results of these studies have been disseminated to the scientific and regulatory communities. However, as use of caverns for waste disposal increases, more government and industry representatives and members of the public will become aware of this practice and will need adequate information about how disposal caverns operate and the risks they pose. In anticipation of this need, DOE has funded Argonne National Laboratory to develop a salt cavern public outreach program. Key components of this program are an informational brochure designed for nontechnical persons and a website that provides greater detail on cavern operations and allows downloadable access to the reports on the topic funded by DOE. This paper provides an overview of the public outreach program. 19. New public information resources on salt caverns. SciTech Connect Tomasko, D.; Veil, J. A. 1999-08-25 For the past decade, interest has been growing in using underground salt caverns for disposing of wastes. The Railroad Commission of Texas has permitted a few caverns for disposal of nonhazardous oil field waste (NOW) and one cavern for disposal of naturally occurring radioactive materials (NORM) from oil field activities. Several salt caverns in Canada have also been permitted for disposal of NOW. In addition, oil and gas agencies in Louisiana and New Mexico are developing cavern disposal regulations. The US Department of Energy (DOE) has funded several studies to evaluate the technical feasibility, legality, economic viability, and risk of disposing of NOW and NORM in caverns. The results of these studies have been disseminated to the scientific and regulatory communities. However, as use of caverns for waste disposal increases, more government and industry representatives and members of the public will become aware of this practice and will need adequate information about how disposal caverns operate and the risks they pose. In anticipation of this need, DOE has fi.mded Argonne National Laboratory to develop a salt cavern public outreach program. Key components of this program are an informational brochure designed for nontechnical persons and a website that provides greater detail on cavern operations and allows downloadable access to the reports on the topic funded by DOE. This paper provides an overview of the public outreach program. 20. Waste immobilization process development at the Savannah River Plant SciTech Connect Charlesworth, D L 1986-01-01 Processes to immobilize various wasteforms, including waste salt solution, transuranic waste, and low-level incinerator ash, are being developed. Wasteform characteristics, process and equipment details, and results from field/pilot tests and mathematical modeling studies are discussed. 1. Utah: Salt Lake City Atmospheric Science Data Center 2014-05-15 article title: Snow-Covered Peaks of the Wasatch and Uinta Mountains ... backdrops for the 2002 Winter Olympics, to be held in Salt Lake City, Utah. The mountains surrounding Salt Lake City are renowned for the dry, powdery snow that results from the arid climate and location at the western edge of the ... 2. Integrated Salt Studies NASA Astrophysics Data System (ADS) Urai, Janos L.; Kukla, Peter A. 2015-04-01 The growing importance of salt in the energy, subsurface storage, and chemical and food industries also increases the challenges with prediction of geometries, kinematics, stress and transport in salt. This requires an approach, which integrates a broader range of knowledge than is traditionally available in the different scientific and engineering disciplines. We aim to provide a starting point for a more integrated understanding of salt, by presenting an overview of the state of the art in a wide range of salt-related topics, from (i) the formation and metamorphism of evaporites, (ii) rheology and transport properties, (iii) salt tectonics and basin evolution, (iv) internal structure of evaporites, (v) fluid flow through salt, to (vi) salt engineering. With selected case studies we show how integration of these domains of knowledge can bring better predictions of (i) sediment architecture and reservoir distribution, (ii) internal structure of salt for optimized drilling and better cavern design, (iii) reliable long-term predictions of deformations and fluid flow in subsurface storage. A fully integrated workflow is based on geomechanical models, which include all laboratory and natural observations and links macro- and micro-scale studies. We present emerging concepts for (i) the initiation dynamics of halokinesis, (ii) the rheology and deformation of the evaporites by brittle and ductile processes, (iii) the coupling of processes in evaporites and the under- and overburden, and (iv) the impact of the layered evaporite rheology on the structural evolution. 3. Molten salt electrolyte separator DOEpatents Kaun, T.D. 1996-07-09 The patent describes a molten salt electrolyte/separator for battery and related electrochemical systems including a molten electrolyte composition and an electrically insulating solid salt dispersed therein, to provide improved performance at higher current densities and alternate designs through ease of fabrication. 5 figs. 4. Mountain of Salt USGS Multimedia Gallery This photo shows a road salt storage stockpile that is awaiting use for pavement deicing, Port of Milwaukee, WI. The use of salt to deice pavement can be harmful to aquatic life in urban streams. The USGS is involved in studies that focus on the influence of winter runoff on aquatic ecosystems.... 5. Hydroxycarboxylic acids and salts SciTech Connect Kiely, Donald E; Hash, Kirk R; Kramer-Presta, Kylie; Smith, Tyler N 2015-02-24 Compositions which inhibit corrosion and alter the physical properties of concrete (admixtures) are prepared from salt mixtures of hydroxycarboxylic acids, carboxylic acids, and nitric acid. The salt mixtures are prepared by neutralizing acid product mixtures from the oxidation of polyols using nitric acid and oxygen as the oxidizing agents. Nitric acid is removed from the hydroxycarboxylic acids by evaporation and diffusion dialysis. 6. Clinical salt deficits. PubMed Luft, Friedrich C 2015-03-01 Salt retention or salt deficit has a bearing on the body fluid volume. Both states are clinically difficult to recognize and quantitate. Salt deficit is particularly cumbersome in that regard since orthostatic blood pressure, heart rate changes, and simple physical inspection are inaccurate and unreliable. Salt deficit can be acute such as after hemorrhage or massive diarrhea, or more chronic as observed in Addison's disease, failure of renal sodium chloride transporters, drug-related effects, or distal nephron disease. Molecular genetics has given us important new insights into salt deficit syndromes. Recent recognition of a novel sodium storage compartment involving sodium binding to proteoglycans adds to the overall complexity of these syndromes. PMID:25471347 7. Salt and nephrolithiasis. PubMed Ticinesi, Andrea; Nouvenne, Antonio; Maalouf, Naim M; Borghi, Loris; Meschi, Tiziana 2016-01-01 Dietary sodium chloride intake is nowadays globally known as one of the major threats for cardiovascular health. However, there is also important evidence that it may influence idiopathic calcium nephrolithiasis onset and recurrence. Higher salt intake has been associated with hypercalciuria and hypocitraturia, which are major risk factors for calcium stone formation. Dietary salt restriction can be an effective means for secondary prevention of nephrolithiasis as well. Thus in this paper, we review the complex relationship between salt and nephrolithiasis, pointing out the difference between dietary sodium and salt intake and the best methods to assess them, highlighting the main findings of epidemiologic, laboratory and intervention studies and focusing on open issues such as the role of dietary salt in secondary causes of nephrolithiasis. PMID:25031016 8. [Salt and cancer]. PubMed Strnad, Marija 2010-05-01 Besides cardiovascular disease, a high salt intake causes other adverse health effects, i.e., gastric and some other cancers, obesity (risk factor for many cancer sites), Meniere's disease, worsening of renal disease, triggering an asthma attack, osteoporosis, exacerbation of fluid retention, renal calculi, etc. Diets containing high amounts of food preserved by salting and pickling are associated with an increased risk of cancers of the stomach, nose and throat. Because gastric cancer is still the most common cancer in some countries (especially in Japan), its prevention is one of the most important aspects of cancer control strategy. Observations among Japanese immigrants in the U.S.A. and Brazil based on the geographic differences, the trend in cancer incidence with time, and change in the incidence patterns indicate that gastric cancer is closely associated with dietary factors such as the intake of salt and salted food. The findings of many epidemiological studies suggest that high dietary salt intake is a significant risk factor for gastric cancer and this association was found to be strong in the presence of Helicobacter (H.) pylori infection with atrophic gastritis. A high-salt intake strips the lining of the stomach and may make infection with H. pylori more likely or may exacerbate the infection. Salting, pickling and smoking are traditionally popular ways of preparing food in Japan and some parts of Asia. In addition to salt intake, cigarette smoking and low consumption of fruit and vegetables increase the risk of stomach cancer. However, it is not known whether it is specifically the salt in these foods or a combination of salt and other chemicals that can cause cancer. One study identified a mutagen in nitrite-treated Japanese salted fish, and chemical structure of this mutagen suggests that it is derived from methionine and that salt and nitrite are precursors for its formation. Working under conditions of heat stress greatly increased the workers' salt excretion through perspiration. Workers exposed to heat stress consumed as much as 13-38 g salt daily. As salt strongly enhances and promotes chemical gastric carcinogenesis and H. pylori infection in both humans and animals, there is an association between work, salt intake, and development of stomach cancer. Reducing salt intake, especially during pregnancy, also reduces the risk of developing breast cancer and many other diseases, as well as obesity. The risk of most cancers is reduced by losing weight. The geographical data and analyses currently available suggest that road salt (road salting in winter) may be associated with elevated mortality from cancer of the breast, lung, esophagus, throat, larynx, large intestine, rectum and bladder. There is no available literature on the health impacts of road salt. The cause and effect relationships cannot be established without further studies. PMID:20649083 9. INSPECTIONS OF THE WASTE ISOLATION PILOT PLANT. EPA Science Inventory The Waste Isolation Pilot Plant (WIPP) is a disposal system for radioactive wastes. Developed by the Department of Energy (DOE), the WIPP is located near Carlsbad in southeastern New Mexico. The DOE is burying radioactive waste 2150 feet underground in an ancient layer of salt ... 10. Concept for Underground Disposal of Nuclear Waste NASA Technical Reports Server (NTRS) Bowyer, J. M. 1987-01-01 Packaged waste placed in empty oil-shale mines. Concept for disposal of nuclear waste economically synergistic with earlier proposal concerning backfilling of oil-shale mines. New disposal concept superior to earlier schemes for disposal in hard-rock and salt mines because less uncertainty about ability of oil-shale mine to contain waste safely for millenium. 11. Deformation of underground deep cavities in rock salts at their long-term operations SciTech Connect Zhuravleva, T.; Shafarenko, E. 1995-12-01 The underground deep cavities are created in rock salts of various morphological types with the purpose of storage of petroleum, gas and nuclear wastes. It is well known that the rock salt has rheological properties, which can result in closure of caverns and loss of their stability. In the evaporitic rocks, especially those containing halite, time-dependent deformation is pronounced even at comparatively low stress levels. At high stress levels this creep becomes a dominant feature of the mechanical behavior of salt rocks. So the knowledge of creep behavior of rock salt is of paramount importance in underground storage application of gas, petroleum products and nuclear wastes. 12. Tank 37H Salt Removal Batch Process and Salt Dissolution Mixing Study SciTech Connect Kwon, K.C. 2001-09-18 Tank 30H is the receipt tank for concentrate from the 3H Evaporator. Tank 30H has had problems, such as cooling coil failure, which limit its ability to receive concentrate from the 3H Evaporator. SRS High Level Waste wishes to use Tank 37H as the receipt tank for the 3H Evaporator concentrate. Prior to using Tank 37H as the 3H Evaporator concentrate receipt tank, HLW must remove 50 inches of salt cake from the tank. They requested SRTC to evaluate various salt removal methods for Tank 37H. These methods include slurry pumps, Flygt mixers, the modified density gradient method, and molecular diffusion. 13. Water purification using organic salts DOEpatents Currier, Robert P. 2004-11-23 Water purification using organic salts. Feed water is mixed with at least one organic salt at a temperature sufficiently low to form organic salt hydrate crystals and brine. The crystals are separated from the brine, rinsed, and melted to form an aqueous solution of organic salt. Some of the water is removed from the aqueous organic salt solution. The purified water is collected, and the remaining more concentrated aqueous organic salt solution is reused. 14. Fission product removal from molten salt using zeolite SciTech Connect Pereira, C.; Babcock, B.D. 1996-10-01 Spent nuclear fuel (SNF) can be treated in a molten salt electrorefiner for conversion into metal and mineral waste forms for geologic disposal. The fuel is dissolved in molten chloride salt. Non-transuranic fission products in the molten salt are ion-exchanged into zeolite A, which is subsequently mixed with glass and consolidated. Zeolite was found to be effective in removing fission product cations from the molten salt. Breakthrough of cesium and the alkaline earths occurred more rapidly than was observed for the rare earths. The effluent composition as a function of time is presented, as well as results for the distribution of fission products along the length of the column. Effects of temperature and salt flow rate are also discussed. 15. Linking external and internal salt geometries - a key to understanding salt dynamics NASA Astrophysics Data System (ADS) Kukla, Peter; Urai, Janos 2014-05-01 Considering the growing importance of salt in the energy, food and waste disposal industries, this paper reviews the status quo and major developments in salt research over the last decade. As a way forward in order to close identified gaps in knowledge, an integrated salt basin evaluation concept is proposed appreciating both external and internal geometries and properties. Examples of key studies in the Central European Basin and the South Oman Salt basin show that such a model may improve our understanding of the multi-scale processes operating in salt terrains. The workflow proposed allows to better asses (i) the initiation and maintenance of salt dynamics, (ii) the evolution of the internal structure of evaporites during halokinesis in salt giants, (iii) the coupling of processes in the evaporites and the salt's under- and overburden. It will lead to a better integration of the different data sets and resulting models, which will provide new insights into the structural evolution of salt giants. Finally it will also stimulate new concepts for (i) the initiation dynamics of halokinesis, (ii) the rheology and mechanics of the evaporites by brittle and ductile processes, (iii) the coupling of processes in the evaporites and the under- and overburden, and (iv) the impact of the layered evaporite rheology on the structural evolution. As an outlook for future research to be initiated in salt terrains we still need to improve our database on evaporite rocks especially the ones which take changes of properties in time into account. This includes for example the dependencies of thermal and mechanical properties on changes in strain, pressure and temperature or external and internal geometry changes relating to slow geological processes. Also geomechanical modelling efforts can be significantly improved by making full use of the data available on the effects of water, and some of the discrepancies seen in experimental data on different salts can probably be explained in terms of these effects. This all will contribute to the development of new integrated techniques for investigating and predicting salt structures from multiple data sets. 16. Electrochemistry of LiCl-Li2O-H2O Molten Salt Systems SciTech Connect Natalie J. Gese; Batric Pesic 2013-03-01 Uranium can be recovered from uranium oxide (UO2) spent fuel through the combination of the oxide reduction and electrorefining processes. During oxide reduction, the spent fuel is introduced to molten LiCl-Li2O salt at 650 degrees C and the UO2 is reduced to uranium metal via two routes: (1) electrochemically, and (2) chemically by lithium metal (Li0) that is produced electrochemically. However, the hygroscopic nature of both LiCl and Li2O leads to the formation of LiOH, contributing hydroxyl anions (OH-), the reduction of which interferes with the Li0 generation required for the chemical reduction of UO2. In order for the oxide reduction process to be an effective method for the treatment of uranium oxide fuel, the role of moisture in the LiCl-Li2O system must be understood. The behavior of moisture in the LiCl-Li2O molten salt system was studied using cyclic voltammetry, chronopotentiometry and chronoamperometry, while reduction to hydrogen was confirmed with gas chromatography. 17. Bedded-salt repository analysis. Final report SciTech Connect Guiffre, M.S.; Kaplan, M.F.; Ensminger, D.A.; Oston, S.G.; Nalbandian, J.Y. 1980-03-31 This report contains a description of an analysis of generic nuclear waste repository in bedded salt. This analysis was performed by TASC for inclusion in a major Lawrence Livermore Laboratory report to NRC; this report therefore should be viewed as providing more complete and detailed information about this analysis than was possible to include in the LLL report. The analysis is performed with the NUTRAN computer codes which are described in the report. The model to be analyzed is defined, and the results of a series of possible waste migration scenarios are presented. Several of these scenarios are used as the basis for a sensitivity analysis, and an uncertainty analysis utilizing Monte Carlo techniques is also performed. A new method for defining the consequences to users of a well drilled near the repository is also described, and results are presented based on two of the waste migration scenarios. 18. Cooking without salt MedlinePLUS ... flavor and nutrition. Plant-based foods -- carrots, spinach, apples, and peaches -- are naturally salt-free. Sun-dried ... types of pepper, including black, white, green, and red. Experiment with vinegars (white and red wine, rice ... 19. Amine salts of nitroazoles DOEpatents Kienyin Lee; Stinecipher, M.M. 1993-10-26 Compositions of matter, a method of providing chemical energy by burning said compositions, and methods of making said compositions are described. These compositions are amine salts of nitroazoles. 1 figure. 20. 24. VIEW SHOWING WASTE GATES ON GRAND CANAL AT JUNCTION ... Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey 24. VIEW SHOWING WASTE GATES ON GRAND CANAL AT JUNCTION WITH OLD CROSSCUT NE/4, Sec. 7, TIN, R4E; LOOKING WEST. OLD CROSSCUT CANAL ENTERS FROM RIGHT. WASTE GATE ON LEFT EMPTIES INTO SALT RIVER BED Photographer: Kevin Kreisel-Coons, May 1990 - Grand Canal, North side of Salt River, Tempe, Maricopa County, AZ 1. Novel Ternary Molten Salt Electrolytes for intermediate-temperature sodium/nickel chloride batteries SciTech Connect Li, Guosheng; Lu, Xiaochuan; Coyle, Christopher A.; Kim, Jin Yong; Lemmon, John P.; Sprenkle, Vincent L.; Yang, Zhenguo 2012-12-15 The sodium-nickel chloride (ZEBRA) battery is typically operated at relatively high temperature (250~350°C) to achieve adequate electrochemical performance. Reducing the operating temperature in the range of 150 to 200°C can lead to enhanced cycle life by suppressing temperature related degradation mechanisms. The reduced temperature range also allows for lower cost materials of construction such as elastomeric sealants and gaskets. To achieve adequate electrochemical performance at lower operating temperatures requires an overall reduction in ohmic losses associated with temperature. This includes reducing the ohmic resistance of ”-alumina solid electrolyte (BASE) and the incorporation of low melting point molten salt as the secondary electrolyte. In present work, planar-type Na/NiCl2 cells with a thin flat plate BASE (600 m) and low melting point secondary electrolyte were evaluated at reduced temperatures. Molten salt formulation for use as secondary electrolytes were fabricated by the partial replace of NaCl in the standard secondary electrolyte (NaAlCl4) with other lower melting point alkali metal salts such as NaBr, LiCl, and LiBr. Electrochemical characterization of the ternary molten salts demonstrated , improved ionic conductivity, and sufficient electrochemical window at reduced temperatures. Furthermore, Na/NiCl2 cells with 50 mol% NaBr-containing secondary electrolyte exhibited reduced polarizations at 175°C compared to the cell with the standard NaAlCl4 catholyte. The cells also exhibited stable cycling performance even at 150oC. 2. Studies of 6Li-NMR properties in different salt solutions in low magnetic fields NASA Astrophysics Data System (ADS) Gordji-Nejad, A.; Colell, J.; Glggler, S.; Blmich, B.; Appelt, S. 2012-01-01 In this article we report the longitudinal relaxation times ( T1) of various 6Li salts ( 6LiI, 6LiCl and 6LiNO 3) in D 2O and H 2O, measured in low magnetic fields ( B0 = 3.5 mT). This investigation serves the purpose of clarifying the relaxation behavior of different 6Li solutions and different concentrations. The measurement were undertaken to establish a framework for future applications of hyperpolarized 6Li in medical imaging, biological studies and investigations of lithium ion batteries. Time will pass during the transport of hyperpolarized lithium ions to the sample, which leads to a polarization loss. In order to store polarization as long as possible, it is necessary to examine which 6Li salt solution has the longest relaxation time T1. Longitudinal relaxation times of 6Li salts in D 2O and H 2O were investigated as a function of concentration and the most extended T1 was found for 6LiI in D 2O and H 2O. In agreement with the theory the relaxation time T1 of all 6Li salts increase with decreasing concentration. In the case of 6LiI in H 2O an inverse behavior was observed. We assume that the prolonged T1 times occur due to formation of 6LiOH upon the solution of 6LiI in H 2O, which settles as a precipitate. By diluting the solution, the precipitate continuously dissolves and approaches T1 of 6LiOH ( T1 28 s), leading to a shorter T1 relaxation time. 3. Salt Lake City, Utah NASA Technical Reports Server (NTRS) 2002-01-01 Salt Lake City, Utah, will host the 2002 Winter Olympic Games. The city is located on the southeastern shore of the Great Salt Lake and sits to the west of the Wasatch Mountains, which rise more than 3,500 meters (10,000 feet) above sea level. The city was first settled in 1847 by pioneers seeking relief from religious persecution. Today Salt Lake City, the capital of Utah, is home to more than 170,000 residents. This true-color image of Salt Lake City was acquired by the Enhanced Thematic Mapper Plus (ETM+), flying aboard Landsat 7, on May 26, 2000. The southeastern tip of the Great Salt Lake is visible in the upper left of the image. The furrowed green and brown landscape running north-south is a portion of the Wasatch Mountains, some of which are snow-capped (white pixels). The greyish pixels in the center of the image show the developed areas of the city. A number of water reservoirs can be seen east of the mountain range. Salt Lake City International Airport is visible on the northwestern edge of the city. About 20 miles south of the airport is the Bingham Canyon Copper Mine (tan pixels), the world's largest open pit excavation. See also this MODIS image of Utah. Image courtesy NASA Landsat7 Science Team and USGS Eros Data Center 4. Modeling of Pilot-Scale Salt-cake Dissolution SciTech Connect Toghiani, R.K.; Smith, L.T.; Lindner, J.S.; Tachiev, G.I.; Yaari, G. 2006-07-01 Large portions of the high-level waste present at the Hanford Site and Savannah River Site are comprised of porous salts with associated interstitial liquors. Various processes have been proposed wherein the aqueous phase is removed followed by dissolution of the salt with further mixing or blending of the resulting stream in a receiver tank. This leads to a large reduction in the radioactivity for the dissolved salt-cake; however, the interstitial retrieval process is hindered by capillary forces within the salt-cake pores and large aqueous phase fractions may remain. Thus, the interim stabilized or low-curie salt processes may have less separation effectiveness than desired. In addition, based on the initial extent of pretreatment of the waste, the salt-cake may be either unsaturated or hydraulically saturated. Different interactions are expected based on the contact of the diluent with the salt and/or on mixing the diluent with the salt and some fraction of interstitial liquid. The initial approximation is that the dissolution is governed by the associated thermodynamics of the system. This may be correct assuming sufficient time for contact between the salt and diluent has occurred. Pilot-scale simulant salt-cake dissolution experiments have been conducted by the Applied Research Center (ARC) at Florida International University. As part of a companion program, these experiments have been modeled at the Diagnostic Instrumentation and Analysis Laboratory (DIAL, Mississippi State University) using the Environmental Simulation Program (ESP, OLI Systems, Inc.). Hanford simulant compositions were examined under unsaturated and saturated conditions. To account for channeling that occurred during the unsaturated experiment, additional operations were required for the process flowsheet. Direct modeling of the saturated bed was possible without this consideration. The results have impacts on the salt-cake retrieval process. First, depending on the extent of interstitial liquid contained in the waste, recycling may be necessary; removal of the resulting aqueous stream at the largest specific gravity consistent with the operating safety basis ensures productive use of water. Secondly, direct modeling of a given waste dissolution must consider variations in the extent of channeling such that limits can be established on anticipated concentrations expected during the course of the retrieval. Finally, the ability to account for heterogeneous dissolution has been accounted for. Details regarding the development of the modeling strategy as well as knowledge gained regarding flowsheet development are provided. (authors) 5. Not salt taste perception but self-reported salt eating habit predicts actual salt intake. PubMed Lee, Hajeong; Cho, Hyun-Jeong; Bae, Eunjin; Kim, Yong Chul; Kim, Suhnggwon; Chin, Ho Jun 2014-09-01 Excessive dietary salt intake is related to cardiovascular morbidity and mortality. Although dietary salt restriction is essential, it is difficult to achieve because of salt palatability. However, the association between salt perception or salt eating habit and actual salt intake remains uncertain. In this study, we recruited 74 healthy young individuals. We investigated their salt-eating habits by questionnaire and salt taste threshold through a rating scale that used serial dilution of a sodium chloride solution. Predicted 24-hr urinary salt excretions using Kawasaki's and Tanaka's equations estimated dietary salt intake. Participants' mean age was 35 yr, and 59.5% were male. Salt sense threshold did not show any relationship with actual salt intake and a salt-eating habit. However, those eating "salty" foods showed higher blood pressure (P for trend=0.048) and higher body mass index (BMI; P for trend=0.043). Moreover, a salty eating habit was a significant predictor for actual salt intake (regression coefficient [β] for Kawasaki's equation 1.35, 95% confidence interval [CI] 10-2.69, P=0.048; β for Tanaka's equation 0.66, 95% CI 0.01-1.31, P=0.047). In conclusion, a self-reported salt-eating habit, not salt taste threshold predicts actual salt intake. PMID:25317023 6. Synergistic Interactions of Sugars/Polyols and Monovalent Salts with Phospholipids Depend upon Sugar/Polyol Complexity and Anion Identity. PubMed Clark, Ginevra A; Henderson, J Michael; Heffern, Charles; Akgn, Blent; Majewski, Jaroslaw; Lee, Ka Yee C 2015-11-24 We found that interactions of dipalmitoylphosphatidylcholine (DPPC) lipid monolayers with sugars are influenced by addition of NaCl. This work is of general importance in understanding how sugar-lipid-salt interactions impact biological systems. Using Langmuir isothermal compressions, fluorescence microscopy, atomic force microscopy, and neutron reflectometry, we examined DPPC monolayers upon addition of sugars/polyols and/or monovalent salts. Sugar-lipid interactions in the presence of NaCl increased with increasing complexity of the sugar/polyol in the order glycerol ? glucose < trehalose. When the anion was altered in the series NaF, NaCl, and NaBr, only minor differences were observed. When comparing LiCl, NaCl, and KCl, sodium chloride had the greatest influence on glucose and trehalose interactions with DPPC. We propose that heterogeneity created by cation binding allows for sugars to bind the lipid headgroups. While cation binding increases in the order K(+) < Na(+) < Li(+), lithium ions may also compete with glucose for binding sites. Thus, both cooperative and competitive factors contribute to the overall influence of salts on sugar-lipid interactions. PMID:26496486 7. Treatment of plutonium process residues by molten salt oxidation SciTech Connect Stimmel, J.; Wishau, R.; Ramsey, K.B.; Montoya, A.; Brock, J.; Heslop, M.; Wernly, K. 1999-04-01 Molten Salt Oxidation (MSO) is a thermal process that can remove more than 99.999% of the organic matrix from combustible {sup 238}Pu material. Plutonium processing residues are injected into a molten salt bed with an excess of air. The salt (sodium carbonate) functions as a catalyst for the conversion of the organic material to carbon dioxide and water. Reactive species such as fluorine, chlorine, bromine, iodine, sulfur, phosphorous and arsenic in the organic waste react with the molten salt to form the corresponding neutralized salts, NaF, NaCl, NaBr, NaI, Na{sub 2}SO{sub 4}, Na{sub 3}PO{sub 4} and NaAsO{sub 2} or Na{sub 3}AsO4. Plutonium and other metals react with the molten salt and air to form metal salts or oxides. Saturated salt will be recycled and aqueous chemical separation will be used to recover the {sup 238}Pu. The Los Alamos National Laboratory system, which is currently in the conceptual design stage, will be scaled down from current systems for use inside a glovebox. 8. Solubility of uranium in alkaline salt solutions SciTech Connect Hobbs, D.T.; Edwards, T.B. 1994-03-29 The solubility of uranium in alkaline salt solutions was investigated to screen for significant factors and interactions among the major salt components and temperature. The components included in the study were the sodium salts of hydroxide, nitrate, nitrite, aluminate, sulfate, and carbonate. General findings from the study included: (1) uranium solubilities are very low (1-20 mg/L) for all solution compositions at hydroxide concentrations from 0.1 to 17 molar (2) carbonate, sulfate, and aluminate are not effective complexants for uranium at high hydroxide concentration, (3) uranium solubility decreases with increasing temperature for most alkaline salt solutions, and (4) uranium solubility increases with changes in solution chemistry that reflect aging of high level waste (increase in nitrite and carbonate concentrations, decrease in nitrate and hydroxide concentrations). A predictive model for the concentration of uranium as a function of component concentrations and temperature was fitted to the data. All of the solution components and temperature were found to be significant. There is a significant lack of fit for the model, which suggests that the dependence on the uranium solubility over the wide range of solution compositions is non-linear and/or that there are other uncontrolled parameters which are important to the uranium solubility. 9. Accelerators for Subcritical Molten-Salt Reactors SciTech Connect Johnson, Roland 2011-08-03 Accelerator parameters for subcritical reactors have usually been based on using solid nuclear fuel much like that used in all operating critical reactors as well as the thorium burning accelerator-driven energy amplifier proposed by Rubbia et al. An attractive alternative reactor design that used molten salt fuel was experimentally studied at ORNL in the 1960s, where a critical molten salt reactor was successfully operated using enriched U235 or U233 tetrafluoride fuels. These experiments give confidence that an accelerator-driven subcritical molten salt reactor will work better than conventional reactors, having better efficiency due to their higher operating temperature, having the inherent safety of subcritical operation, and having constant purging of volatile radioactive elements to eliminate their accumulation and potential accidental release in dangerous amounts. Moreover, the requirements to drive a molten salt reactor can be considerably relaxed compared to a solid fuel reactor, especially regarding accelerator reliability and spallation neutron targetry, to the point that much of the required technology exists today. It is proposed that Project-X be developed into a prototype commercial machine to produce energy for the world by, for example, burning thorium in India and nuclear waste from conventional reactors in the USA. 10. Crushed-salt constitutive model update SciTech Connect Callahan, G.D.; Loken, M.C.; Mellegard, K.D. 1998-01-01 Modifications to the constitutive model used to describe the deformation of crushed salt are presented in this report. Two mechanisms--dislocation creep and grain boundary diffusional pressure solutioning--defined previously but used separately are combined to form the basis for the constitutive model governing the deformation of crushed salt. The constitutive model is generalized to represent three-dimensional states of stress. New creep consolidation tests are combined with an existing database that includes hydrostatic consolidation and shear consolidation tests conducted on Waste Isolation Pilot Plant and southeastern New Mexico salt to determine material parameters for the constitutive model. Nonlinear least-squares model fitting to data from the shear consolidation tests and a combination of the shear and hydrostatic consolidation tests produced two sets of material parameter values for the model. The change in material parameter values from test group to test group indicates the empirical nature of the model but demonstrates improvement over earlier work with the previous models. Key improvements are the ability to capture lateral strain reversal and better resolve parameter values. To demonstrate the predictive capability of the model, each parameter value set was used to predict each of the tests in the database. Based on the fitting statistics and the ability of the model to predict the test data, the model appears to capture the creep consolidation behavior of crushed salt quite well. 11. Constitutive behavior of reconsolidating crushed salt SciTech Connect Callahan, G.D.; Mellegard, K.D. 1998-02-01 The constitutive model used to describe deformation of crushed salt is presented in this paper. Two mechanisms--dislocation creep and grain boundary diffusional pressure solutioning--are combined to form the basis for the constitutive model governing deformation of crushed salt. The constitutive model is generalized to represent three-dimensional states of stress. Recently completed creep consolidation tests are combined with an existing database that includes hydrostatic consolidation and shear consolidation tests conducted on Waste Isolation Pilot Plant (WIPP) and southeastern New Mexico salt to determine material parameters for the constitutive model. Nonlinear least-squares model fitting to data from shear consolidation tests and a combination of shear and hydrostatic tests produces two sets of material parameter values for the model. Changes in material parameter values from test group to test group indicate the empirical nature of the model but show significant improvement over earlier work. To demonstrate the predictive capability of the model, each parameter value set was used to predict each of the tests in the database. Based on fitting statistics and ability of the model to predict test data, the model appears to capture the creep consolidation behavior of crushed salt quite well. 12. Mass transport in salt repositories: Steady-state transport through interbeds SciTech Connect Hwang, Y.; Lee, W.W.-L.; Chambre, P.L.; Pigford, T.H. . Dept. of Nuclear Engineering) 1989-03-01 Salt has long been a candidate for geologic disposal of nuclear waste. Because salt is extremely soluble in water, the existence of rock salt in the ground atest to the long-term stability of the salt. Both bedded salt and salt domes have been considered for nuclear waste disposal in the United States and Europe. While the salt is known to be quite pure in salt domes, bedded salt is interlaced with beds of sediments. Traditionally rock salt has not been considered water-conducting, but sediments layers would be classical porous media, capable of conducting water. Therefore there is interest in determining whether interbeds in bedded salt constitute pathway for radionuclide migration. In this report we consider steady-state migration of radionuclides from a single waste cylinder into a single interbed. Two approaches are used. In 1982 Neretnieks proposed an approach for calculating the steady-state transport of oxidants to a copper container. We have adapted that approach for calculating steady-state radionuclide migration away from the waste package, as a first approximation. We have also analyzed the problem of time-dependent radionuclide diffusion from a container through a backfill layer into a fracture, and we used the steady-state solution from that problem for comparison. Section 2 gives a brief summary of the geology of interbeds in bedded salt. Section 3 presents the mass transfer resistances approach of Neretnieks, summarizing the formulation and giving numerical illustrations of the steady-state two-dimensional diffusion analysis. Section 4 gives a brief statement of the steady-state result from a related analysis. Conclusions are stated in Section 5. 13 refs., 5 figs., 2 tabs. 13. Tank waste chemistry: A new understanding of waste aging SciTech Connect Babad, H.; Camaioni, D.M.; Lilga, M.A.; Samuels, W.D.; Strachan, D.M. 1993-02-01 There is concern about the risk of uncontrolled exothermic reaction(s) in Hanford Site waste tanks containing NO{sub 3}{sup {minus}}/NO{sub 2} based salts and/or metal hydroxide sludges in combination with organics or ferrocyanides. However, gradual oxidation of the waste in the tanks to less reactive species appears to have reduced the risk. In addition, wastes sampled to date contain sufficiently large quantities of water so that propagation reactions are highly unlikely. This paper investigates various aspects of the aging of Hanford tank wastes. 14. Results of water quality sampling near Richton, Cypress Creek and Lampton Salt Domes, Mississippi USGS Publications Warehouse Gandl, L.A.; Spiers, C.A. 1980-01-01 In the Mississippi salt basin in southern Mississippi, chemical quality studies of surface water and ground water have been made to determine present water-quality conditions near three salt domes being studied by the Department of Energy as potential repositories for radioactive wastes. Chloride concentrations in excess of 60 milligrams per liter in surface water and ground water in Perry County indicate that contamination could be occurring from industrial wastes, oil test wells, or dissolution of Richton or Cypress Creek domes. (USGS) 15. Hanford's Simulated Low Activity Waste Cast Stone Processing SciTech Connect Kim, Young 2013-08-20 Cast Stone is undergoing evaluation as the supplemental treatment technology for Hanford’s (Washington) high activity waste (HAW) and low activity waste (LAW). This report will only cover the LAW Cast Stone. The programs used for this simulated Cast Stone were gradient density change, compressive strength, and salt waste form phase identification. Gradient density changes show a favorable outcome by showing uniformity even though it was hypothesized differently. Compressive strength exceeded the minimum strength required by Hanford and greater compressive strength increase seen between the uses of different salt solution The salt waste form phase is still an ongoing process as this time and could not be concluded. 16. Fundamental Properties of Salts SciTech Connect Toni Y Gutknecht; Guy L Fredrickson 2012-11-01 Thermal properties of molten salt systems are of interest to electrorefining operations, pertaining to both the Fuel Cycle Research & Development Program (FCR&D) and Spent Fuel Treatment Mission, currently being pursued by the Department of Energy (DOE). The phase stability of molten salts in an electrorefiner may be adversely impacted by the build-up of fission products in the electrolyte. Potential situations that need to be avoided, during electrorefining operations, include (i) fissile elements build up in the salt that might approach the criticality limits specified for the vessel, (ii) electrolyte freezing at the operating temperature of the electrorefiner due to changes in the liquidus temperature, and (iii) phase separation (non-homogenous solution). The stability (and homogeneity) of the phases can be monitored by studying the thermal characteristics of the molten salts as a function of impurity concentration. Simulated salt compositions consisting of the selected rare earth and alkaline earth chlorides, with a eutectic mixture of LiCl-KCl as the carrier electrolyte, were studied to determine the melting points (thermal characteristics) using a Differential Scanning Calorimeter (DSC). The experimental data were used to model the liquidus temperature. On the basis of the this data, it became possible to predict a spent fuel treatment processing scenario under which electrorefining could no longer be performed as a result of increasing liquidus temperatures of the electrolyte. 17. Salt Restriction in Diabetes. PubMed Clifton, Peter M; Keogh, Jennifer B 2015-09-01 There are no long-term interventions examining the effects of salt reduction in people with diabetes, and these are urgently required. Sodium reduction is controversial as it appears that an intake below 2.5g and above 6g/day of salt is associated with increased cardiovascular disease risk. However, pre-existing illness leading to a lower salt intake may confound the findings. Only a few studies have prospectively collected data on the sodium intake and excretion of people with diabetes and examined hard end points. In addition, future studies need to collect more data on food intake as well as coexistent illnesses to address potential confounding. The World Health Organization recommends a reduction to less than 5g/day salt in adults. Given that the available evidence suggests that the salt intake of people with type 2 diabetes is generally well above 6g/day it seems reasonable to ensure individuals with diabetes have an intake below 6g/day. However, such recommendations need to be individualized. PMID:26194154 18. HLW Salt Disposition Alternatives Preconceptual Phase II Summary Report SciTech Connect Piccolo, S.F. 1999-07-09 The purpose of the report is to summarize the process used to identify the Short List alternatives that will be evaluated during Phase III and to document the results of the selection process. The Phase III evaluation will result in the determination of the preferred alternative(s) to be used for final disposition of the HLW salt to a permitted waste form. 19. Technology for treatment of salt residue stored at NPPs SciTech Connect Kobelev, A.P.; Savkin, A.E.; Sinjakin, O.G.; Kachalova, E.A.; Sorokoletova, A.N.; Nechaev, V.R. 2007-07-01 At Moscow SIA 'Radon', three (3) options for NPP salt residue treatment were developed and tested. Option 1 consists of dissolving the salt residue and subsequent treatment by ozonization, separation of the deposits formed from ozonization and selective cleaning by ferrocyanide sorbents. Option 2 consists of fusion of the salt residue, addition of glass-forming additives and melting of borosilicate glass in a melter such as a 'cold crucible'. Option 3 consists of dissolving the salt residue, oxidation of the solution obtained, removal of radionuclides by collectors and the separate handling of formed deposits and the solution. The deposits containing more than 99 % of the activity are directed to vitrification and the solution is directed either to a concentrates dryer or to cementation. The vitrified waste product is placed in repository for solid radioactive waste storage and the solidified product from the solution goes to an industrial waste disposal site or a repository specially developed at NPP sites for 'exempt waste' products by IAEA classification. (authors) 20. Stress measurements in rock salt using hydraulic fracturing SciTech Connect Wawersik, W.R.; Stone, C.M. 1986-01-01 Hydraulic fracturing was applied in horizontal drillholes in the Salado salt formation near Carlsbad, New Mexico. Testing took place approximately 650 m below surface in order to support the design of a Waste Isolation Pilot Plant (WIPP) for the disposal of radioactive waste from defense activities of the United States. Hydraulic fracturing was performed primarily to determine whether the virgin in situ stress state at the WIPP site is isotropic and whether the magnitudes of the the virgin in situ stresses correspond to the weight of the overburden. Beyond these limited objectives, measurements are being analyzed to evaluate the usefulness of hydraulic fracturing in salt formations in general. Such measurements are desirable to determine stresses induced by mining and to monitor time-dependent stress changes around underground excavations in salt masses. Hydraulic fracturing measurements are also relevant to the evaluation of allowable pressures before fracturing is induced in pressurized boreholes and storage caverns. 1. Waste package/repository impact study: Final report SciTech Connect Not Available 1985-09-01 The Waste Package/Repository Impact Study was conducted to evaluate the feasibility of using the current reference salt waste package in the salt repository conceptual design. All elements of the repository that may impact waste package parameters, i.e., (size, weight, heat load) were evaluated. The repository elements considered included waste hoist feasibility, transporter and emplacement machine feasibility, subsurface entry dimensions, feasibility of emplacement configuration, and temperature limits. The evaluations are discussed in detail with supplemental technical data included in Appendices to this report, as appropriate. Results and conclusions of the evaluations are discussed in light of the acceptability of the current reference waste package as the basis for salt conceptual design. Finally, recommendations are made relative to the salt project position on the application of the reference waste package as a basis for future design activities. 31 refs., 11 figs., 11 tabs. 2. Gas releases from salt SciTech Connect Ehgartner, B.; Neal, J.; Hinkebein, T. 1998-06-01 The occurrence of gas in salt mines and caverns has presented some serious problems to facility operators. Salt mines have long experienced sudden, usually unexpected expulsions of gas and salt from a production face, commonly known as outbursts. Outbursts can release over one million cubic feet of methane and fractured salt, and are responsible for the lives of numerous miners and explosions. Equipment, production time, and even entire mines have been lost due to outbursts. An outburst creates a cornucopian shaped hole that can reach heights of several hundred feet. The potential occurrence of outbursts must be factored into mine design and mining methods. In caverns, the occurrence of outbursts and steady infiltration of gas into stored product can effect the quality of the product, particularly over the long-term, and in some cases renders the product unusable as is or difficult to transport. Gas has also been known to collect in the roof traps of caverns resulting in safety and operational concerns. The intent of this paper is to summarize the existing knowledge on gas releases from salt. The compiled information can provide a better understanding of the phenomena and gain insight into the causative mechanisms that, once established, can help mitigate the variety of problems associated with gas releases from salt. Outbursts, as documented in mines, are discussed first. This is followed by a discussion of the relatively slow gas infiltration into stored crude oil, as observed and modeled in the caverns of the US Strategic Petroleum Reserve. A model that predicts outburst pressure kicks in caverns is also discussed. 3. Development of Polymeric Waste Forms for the Encapsulation of Toxic Wastes Using an Emulsion-Encapsulation Based Process SciTech Connect Evans, R.; Quach, A.; Birnie, D. P.; Saez, A. E.; Ela, W. P.; Zeliniski, B. J. J.; Xia, G.; Smith, H. 2003-01-01 Developed technologies in vitrification, cement, and polymeric materials manufactured using flammable organic solvents have been used to encapsulate solid wastes, including low-level radioactive materials, but are impractical for high salt-content waste streams (Maio, 1998). In this work, we investigate an emulsification process for producing an aqueous-based polymeric waste form as a preliminary step towards fabricating hybrid organic/inorganic polyceram matrices. The material developed incorporates epoxy resin and polystyrene-butadiene (PSB) latex to produce a waste form that is non-flammable, light weight, of relatively low cost, and that can be loaded to a relatively high weight content of waste materials. Sodium nitrate was used as a model for the salt waste. Small-scale samples were manufactured and analyzed using leach tests designed to measure the diffusion coefficient and leachability index for the fastest diffusing species in the waste form, the salt ions. The microstructure and composition of the samples were probed using SEM/EDS techniques. The results show that some portion of the salt migrates towards the exterior surfaces of the waste forms during the curing process. A portion of the salt in the interior of the sample is contained in polymer corpuscles or sacs. These sacs are embedded in a polymer matrix phase that contains fine, well-dispersed salt crystals. The diffusion behavior observed in sections of the waste forms indicates that samples prepared using this emulsion process meet or exceed the leachability criteria suggested for low level radioactivity waste forms. 4. [Salt intake in children]. PubMed Girardet, J-P; Rieu, D; Bocquet, A; Bresson, J-L; Briend, A; Chouraqui, J-P; Darmaun, D; Dupont, C; Frelut, M-L; Hankard, R; Goulet, O; Simeoni, U; Turck, D; Vidailhet, M 2014-05-01 Very early in life, sodium intake correlates with blood pressure level. This warrants limiting the consumption of sodium by children. However, evidence regarding exact sodium requirements in that age range is lacking. This article focuses on the desirable sodium intake according to age as suggested by various groups of experts, on the levels of sodium intake recorded in consumption surveys, and on the public health strategies implemented to reduce salt consumption in the pediatric population. Practical recommendations are given by the Committee on nutrition of the French Society of Pediatrics in order to limit salt intake in children. PMID:24686038 5. Constitutive modeling of salt behavior: State of the technology SciTech Connect Munson, D.E.; Wawersik, W.R. 1991-01-01 The modern investigation of the thermochemical behavior of salt started in the mid-1930's and, for what appears to be a very narrow discipline, salt mechanics'' has acquired considerable technical depth and sophistication. The last three decades have been especially productive in constitutive model development and laboratory investigations of time-dependent creep behavior. This has been largely due to anticipated use of domal or bedded salt deposits as sites for radioactive waste repositories and to expanded need for hydrocarbon and feedstock storage caverns. Salt is an interesting material, in that it is metal like''; and, therefore, constitutive modeling can draw upon a large body of metal deformation information to arrive at appropriate models of behavior. Testing apparatus and methods have centered on either uniaxial or triaxial compression to obtain steady state and transient creep responses. Flow and fracture potentials have been defined. Validation attempts of the models against field data, although limited, have proved promising. 27 refs. 6. Effects of waste content of glass waste forms on Savannah River high-level waste disposal costs SciTech Connect McDonell, W R; Jantzen, C M 1985-01-01 Effects of the waste content of glass waste forms of Savannah River high-level waste disposal costs are evaluated by their impact on the number of waste canisters produced. Changes in waste content affect onsite Defense Waste Processing Facility (DWPF) costs as well as offsite shipping and repository emplacement charges. A nominal 1% increase over the 28 wt % waste loading of DWPF glass would reduce disposal costs by about$50 million for Savannah River wastes generated to the year 2000. Waste form modifications under current study include adjustments of glass frit content to compensate for added salt decontamination residues and increased sludge loadings in the DWPF glass. Projected cost reductions demonstrate significant incentives for continued optimization of the glass waste loadings. 13 refs., 3 figs., 3 tabs. 7. Analysis of Multistage and Other Creep Data for Domal Salts SciTech Connect Munson, D.E. 1998-10-01 There have existed for some time relatively sparse creep databases for a number of domal salts. Although all of these data were analyzed at the time they were reported, to date there has not been a comprehensive, overall evaluation within the same analysis framework. Such an evaluation may prove of value. The analysis methodology is based on the Multimechanism Deformation (M-D) description of salt creep and the corresponding model parameters determined from conventional creep tests. The constitutive model of creep wss formulated through application of principles involved in micromechanical modeling. It was possible, at minimum, to obtain the steady state parameters of the creep model from the data on the domal salts. When this was done, the creep of the domal salts, as compared to the well-defined Waste Isolation Pilot Plant (WIPP) bedded clean salt, was either essentially identical to, or significantly harder (more creep resistant) than WIPP salt. Interestingly, the domal salts form two distinct groups, either sofl or hard, where the difference is roughly a factor often in creep rate between the twcl groups. As might be expected, this classification corresponds quite well to the differences in magnitude of effective creep volume losses of the Strategic Petroleum Reserve (SPR) caverns as determined by the CAVEMAN cavern pressure history analysis, depending upon the specific dome or region within the dome. Creep response shoulcl also correlate to interior cavern conditions that produce salt falls. WMle, in general, the caverns in hard sah have a noticeably greater propensity for salt falls, a smaller number of similar events are exhibited even in the caverns in soft salt. 8. Safe actinide disposition in molten salt reactors SciTech Connect Gat, U. 1997-03-01 Safe molten salt reactors (MSR) can readily accommodate the burning of all fissile actinides. Only minor compromises associated with plutonium are required. The MSRs can dispose safely of actinides and long lived isotopes to result in safer and simpler waste. Disposing of actinides in MSRs does increase the source term of a safety optimized MSR. It is concluded that the burning and transmutation of actinides in MSRs can be done in a safe manner. Development is needed for the processing to handle and separate the actinides. Calculations are needed to establish the neutron economy and the fuel management. 9 refs. 9. Permeability of WIPP Salt During Damage Evolution and Healing SciTech Connect BODNER,SOL R.; CHAN,KWAI S.; MUNSON,DARRELL E. 1999-12-03 The presence of damage in the form of microcracks can increase the permeability of salt. In this paper, an analytical formulation of the permeability of damaged rock salt is presented for both initially intact and porous conditions. The analysis shows that permeability is related to the connected (i.e., gas accessible) volumetric strain and porosity according to two different power-laws, which may be summed to give the overall behavior of a porous salt with damage. This relationship was incorporated into a constitutive model, known as the Multimechanism Deformation Coupled Fracture (MDCF) model, which has been formulated to describe the inelastic flow behavior of rock salt due to coupled creep, damage, and healing. The extended model was used to calculate the permeability of rock salt from the Waste Isolation Pilot Plant (WIPP) site under conditions where damage evolved with stress over a time period. Permeability changes resulting from both damage development under deviatoric stresses and damage healing under hydrostatic pressures were considered. The calculated results were compared against experimental data from the literature, which indicated that permeability in damaged intact WIPP salt depends on the magnitude of the gas accessible volumetric strain and not on the total volumetric strain. Consequently, the permeability of WIPP salt is significantly affected by the kinetics of crack closure, but shows little dependence on the kinetics of crack removal by sintering. 10. Nickel, soluble salts Integrated Risk Information System (IRIS) Nickel , soluble salts ; CASRN Various Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarcinogenic 11. Dalapon, sodium salt Integrated Risk Information System (IRIS) Dalapon , sodium salt ; CASRN 75 - 99 - 0 Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarcinoge 12. Thallium (I), soluble salts Integrated Risk Information System (IRIS) Thallium ( I ) , soluble salts ; CASRN Various Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarc 13. SALT and Spelling Achievement. ERIC Educational Resources Information Center Nelson, Joan A study investigated the effects of suggestopedic accelerative learning and teaching (SALT) on the spelling achievement, attitudes toward school, and memory skills of fourth-grade students. Subjects were 20 male and 28 female students from two self-contained classrooms at Kennedy Elementary School in Rexburg, Idaho. The control classroom and the 14. Chlorite (sodium salt) Integrated Risk Information System (IRIS) Chlorite ( sodium salt ) ; CASRN 7758 - 19 - 2 Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarc 15. Uranium, soluble salts Integrated Risk Information System (IRIS) Uranium , soluble salts ; no CASRN Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarcinogenic Eff 16. Unitized paramagnetic salt thermometer SciTech Connect Abraham, B.M. 1982-06-01 The details of construction and assembly of a cerous magnesium nitrate (CMN) paramagnetic thermometer are presented. The thermometer is a small unit consisting of a primary, two secondaries, the salt pill, and thermal links. The thermometer calibration changes very little on successive coolings and is reliable to 35 mK. A typical calibration curve is also presented. 17. 21 CFR 100.155 - Salt and iodized salt. Code of Federal Regulations, 2013 CFR 2013-04-01 ... 21 Food and Drugs 2 2013-04-01 2013-04-01 false Salt and iodized salt. 100.155 Section 100.155 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN CONSUMPTION GENERAL Specific Administrative Rulings and Decisions 100.155 Salt and... 18. 21 CFR 100.155 - Salt and iodized salt. Code of Federal Regulations, 2012 CFR 2012-04-01 ... 21 Food and Drugs 2 2012-04-01 2012-04-01 false Salt and iodized salt. 100.155 Section 100.155 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN CONSUMPTION GENERAL Specific Administrative Rulings and Decisions 100.155 Salt and... 19. Salt Marshes at Chincoteague Island USGS Multimedia Gallery Salt marshes at Chincoteague Island. The salt marshes that make up Chincoteague Island are important habitat for migrating waterfowl. In addition, they serve an important role in protecting inland ecosystems and communities from oceanic storms.... 20. Testing of stripping columns for the removal of benzene from aqueous radioactive salt solution SciTech Connect Georgeton, G.K.; Taylor, G.A.; Gaughan, T.P. 1995-06-27 Radioactive high level wastes (HLW) generated from production of special nuclear materials at the Savannah River Site (SRS) are held in interim storage in 51 underground, million gallon tanks. Radioactive cesium ({sup 137}Cs) is segregated by evaporation of aqueous waste solution for interim storage in a salt matrix comprised of Na and K salts or in concentrated salt solution. The saltcake will be dissolved and {sup 137}Cs will be separated from the nonradioactive salts in solution in the In-Tank Precipitation (ITP) Process. The cesium will be combined with other radioactive species and glass formers to be melted and poured into stainless steel canisters in the Defense Waste Processing Facility (DWPF). The salt solution remaining after decontamination in the ITP process will be incorporated into grout for disposal at the sites Saltstone facility. In the ITP facility, sodium tetraphenylborate (STPB) will be added to precipitate the cesium. Potassium in the waste solution also reacts with STPB and precipitates. Due to radiolytic and chemical degradation of the tetraphenylborate (TPB) precipitate, benzene is generated. The benzene dissolves into the decontaminated salt solution (DSS) and into water (WW) used to {open_quotes}wash{close_quotes} the precipitate to lower the soluble salt content of the slurry. Safety and processing requirements for disposal of the DSS and for temporary storage of the WW dictate that the benzene concentration be reduced. 1. Salt acclimation processes in wheat. PubMed Janda, Tibor; Darko, Éva; Shehata, Sami; Kovács, Viktória; Pál, Magda; Szalai, Gabriella 2016-04-01 Young wheat plants (Triticum aestivum L. cv. Mv Béres) were exposed to 0 or 25 mM NaCl for 11 days (salt acclimation). Thereafter the plants were irrigated with 500 mM NaCl for 5 days (salt stress). Irrigating the plants with a low concentration of NaCl successfully led to a reduction in chlorotic symptoms and in the impairment of the photosynthetic processes when the plants were exposed to subsequent high-dose salt treatment. After exposure to a high concentration of NaCl there was no difference in leaf Na content between the salt-acclimated and non-acclimated plants, indicating that salt acclimation did not significantly modify Na transport to the shoots. While the polyamine level was lower in salt-treated plants than in the control, salt acclimation led to increased osmotic potential in the leaves. Similarly, the activities of certain antioxidant enzymes, namely glutathione reductase, catalase and ascorbate peroxidase, were significantly higher in salt-acclimated plants. The results also suggest that while SOS1, SOS2 or NHX2 do not play a decisive role in the salt acclimation processes in young wheat plants; another stress-related gene, WALI6, may contribute to the success of the salt acclimation processes. The present study suggested that the responses of wheat plants to acclimation with low level of salt and to treatment with high doses of salt may be fundamentally different. PMID:26854409 2. Nature and location of the receptors for salt-gland secretion in the goose PubMed Central Hanwell, Ann; Linzell, J. L.; Peaker, M. 1972-01-01 1. The nature and location of the receptors which stimulate salt-gland secretion in the goose have been investigated. 2. The rapid injection of homologous blood (sufficient to raise the blood volume by 16 and 9%) into the right atrium failed to induce secretion. In contrast, hypertonic sucrose, Na2SO4 and LiCl initiated secretion. 3. These results support the theory that osmoreceptors initiate secretion by detecting an increase in plasma tonicity. 4. The minimal amount of hypertonic NaCl required to initiate secretion when infusions were made into a carotid artery or into various arteries and veins in the splanchnic region was not less than that required by an I.V. route. 5. Cross-circulation and perfusion studies also showed that a raised [NaCl] in the blood perfusing the head was ineffective in evoking secretion and thus that plasma tonicity must be raised elsewhere in the body. 6. Secretion in response to salt-loading was abolished or prevented by cutting the vagus nerves or blocking them with local anaesthetic. Stimulation of the cephalic end of the cut vagi in an isolated, perfused decerebrate head induced secretion, indicating that the afferent fibres from the receptors to the C.N.S. lie in the vagus nerves. Cutting the vagi below the heart, however, had no effect on the secretory response. 7. Blocking nerves in the crop with local anaesthetic had no effect on secretion induced by salt-loading but when local anaesthetic was injected into the pericardial sac, secretion decreased immediately, stopped, and recovered with a time course similar to that seen after blocking the vagus nerves. 8. Section of the vagi in the neck abolished the tachycardia observed in response to the injection of hypertonic NaCl into the right atrium. 9. As in other species, stimulation of the secretory nerve' induced secretion in anaesthetized or decerebrate geese. 10. Hexamethonium given I.V. or applied topically to the secretory nerve ganglion' blocked secretion in response to salt-loading or to secretory nerve stimulation. 11. It appears that the receptors for salt-gland secretion are located in or near the heart and that afferent fibres from these receptors travel in the vagus nerves to the C.N.S. 12. A possible scheme of the secretory reflex which initiates and maintains salt-gland activity is proposed. PMID:5085337 3. Solvent Extraction Batch Distribution Coefficients with Savannah River Site Dissolved Salt Cake SciTech Connect Walker, D.D. 2002-05-22 Researchers characterized high-level waste derived from dissolved salt cake from the Savannah River Site (SRS) tank farm and measured the cesium distribution coefficients (DCs) for extraction, scrub, and stripping steps of the caustic-side solvent extraction (CSSX) flowsheet. The measurements used two SRS high-level waste samples derived entirely or in part from salt cake. The chemical compositions of both samples are reported. Dissolved salt cake waste contained less Cs-137 and more dianions than is typical of supernate samples. Extraction, scrub, and strip DCs values for both samples exceeded process requirements and agreed well with model predictions. The results indicate no significant problems processing dissolved salt cake compared to supernate. During the course of testing, researchers observed sorption of cesium on glass sample vials containing strip solutions. The problem was detected in the material balance calculated for each organic/aqueous contact. Methods and recommendations for avoiding this problem are discussed. 4. Frost formation with salt NASA Astrophysics Data System (ADS) Guadarrama-Cetina, J.; Mongruel, A.; González-Viñas, W.; Beysens, D. 2015-06-01 The formation of frost in presence of salt (NaCl) crystal is experimentally investigated on a hydrophobic surface. It presents several remarkable features due to the interplay of salty-water saturation pressure evolution, initially lower than the saturation pressure of ice and water, and the percolating propagation of ice dendrites from defects throughout the supercooled water droplet pattern. In particular, it is remarkable that nucleation of supercooled water and/or ice is prevented around the salty drop in a region of inhibited condensation where the substrate remains dry. As condensation proceeds, salt concentration decreases to eventually become lower than ice's, allowing ice dendrites to hit the salty drop. Salty water then melts ice but eventually freezes as an effect of dilution. 5. SALT IN AYURVEDA I PubMed Central Mooss, N S 1987-01-01 In basic Ayurveda texts, Susruta, Caraka and Vagbhata, some quite specific Salts (Lavanam) have been described and their properties and actions are enumerated. By comparing those accounts with the present methods of preparation, conclusions have been made and evidently spurious methods are pointed out. The reported properties of Saindhava, Samudra, Vida, Sauvarcha, Romaka, Audbhida, Gutika, the Katu Group, Krsna and Pamsuja Lavanas are discussed in terms of their chemical constituents here and, thus, the authors establish its inter-connections. PMID:22557573 6. Calculation of density and permeability of compacted crushed salt within an engineered shaft sealing system SciTech Connect Loken, M. 1997-07-01 Crushed salt from the host Salado Formation is proposed as a sealing material in one component of a multicomponent seal system design for the shafts of the Waste Isolation Pilot Plant (WIPP), a mined geological repository for storage and disposal of transuranic radioactive wastes located near Carlsbad, New Mexico. The crushed salt will be compacted and placed at a density approaching 90% of the intact density of the host Salado salt. Creep closure of the shaft will further compact the crushed salt over time, thereby reducing the crushed-salt permeability from the initial state and creating an effective long-term seal. A structural model and a fluid flow model have been developed to provide an estimate of crushed-salt reconsolidation rate as a function of depth, time, and pore pressure. Model results are obtained in terms of crushed-salt permeability as a function of time and depth within the salt column. Model results indicate that average salt column permeability will be reduced to 3.3 {times} 10{sup {minus}20} m{sup 2} in about 100 years, which provides for an acceptable long-term seal component. 7. Time-dependent expression of hypertonic effects on bullfrog taste nerve responses to salts and bitter substances. PubMed Mashiyama, Kazunori; Nozawa, Yuhei; Ohtubo, Yoshitaka; Kumazawa, Takashi; Yoshii, Kiyonori 2014-03-27 We previously showed that the hypertonicity of taste stimulating solutions modified tonic responses, the quasi-steady state component following the transient (phasic) component of each integrated taste nerve response. Here we show that the hypertonicity opens tight junctions surrounding taste receptor cells in a time-dependent manner and modifies whole taste nerve responses in bullfrogs. We increased the tonicity of stimulating solutions with non-taste substances such as urea or ethylene glycol. The hypertonicity enhanced phasic responses to NaCl>0.2M, and suppressed those to NaCl<0.1M, 1mM CaCl2, and 1mM bitter substances (quinine, denatonium and strychnine). The hypertonicity also enhanced the phasic responses to a variety of 0.5M salts such as LiCl and KCl. The enhancing effect was increased by increasing the difference between the ionic mobilities of the cations and anions in the salt. A preincubation time >20s in the presence of 1M non-taste substances was needed to elicit both the enhancing and suppressing effects. Lucifer Yellow CH, a paracellular marker dye, diffused into bullfrog taste receptor organs in 30s in the presence of hypertonicity. These results agreed with our proposed mechanism of hypertonic effects that considered the diffusion potential across open tight junctions. PMID:24513402 8. Treatment for hydrazine-containing waste water solution NASA Technical Reports Server (NTRS) 1986-01-01 The treatment for waste solutions containing hydrazine is presented. The invention attempts oxidation and decomposition of hydrazine in waste water in a simple and effective processing. The method adds activated charcoal to waste solutions containing hydrazine while maintaining a pH value higher than 8, and adding iron salts if necessary. Then, the solution is aerated. 9. Feed Basis for Processing Relatively Low Radioactivity Waste Tanks SciTech Connect Jones , R.T. 2003-02-18 This paper presents the characterization of potential feed for processing relatively low radioactive waste tanks. The feed characterization is based on waste characterization data extracted from the waste characterization system. This data is compared to salt cake sample results from Tanks 37, 38, and 41. 10. Feed Basis for Processing Relatively Low Radioactivity Waste Tanks SciTech Connect Pike, J.A. 2002-07-09 This paper presents the characterization of potential feed for processing relatively low radioactive waste tanks. The feed characterization is based on waste characterization data extracted from the waste characterization system. This data is compared to salt cake sample results from Tanks 37, 38 and 41. 11. A Trail of Salts NASA Technical Reports Server (NTRS) 2004-01-01 This graph shows the relative abundances of sulfur (in the form of sulfur tri-oxide) and chlorine at three Meridiani Planum sites: soil measured in the small crater where Opportunity landed; the rock dubbed 'McKittrick' in the outcrop lining the inner edge of the crater; and the rock nicknamed 'Guadalupe,' also in the outcrop. The 'McKittrick' data shown here were taken both before and after the rover finished grinding the rock with its rock abrasion tool to expose fresh rock underneath. The 'Guadalupe' data were taken after the rover grounded the rock. After grinding both rocks, the sulfur abundance rose to high levels, nearly five times higher than that of the soil. This very high sulfur concentration reflects the heavy presence of sulfate salts (approximately 30 percent by weight) in the rocks. Chloride and bromide salts are also indicated. Such high levels of salts strongly suggest the rocks contain evaporite deposits, which form when water evaporates or ice sublimes into the atmosphere. 12. Excitonic improvement of colloidal nanocrystals in salt powder matrix for quality lighting and color enrichment. PubMed Erdem, Talha; Soran-Erdem, Zeliha; Kelestemur, Yusuf; Gaponik, Nikolai; Demir, Hilmi Volkan 2016-01-25 Here we report excitonic improvement in color-converting colloidal nanocrystal powders enabled by co-integrating nonpolar green- and red-emitting nanocrystal energy transfer pairs into a single LiCl salt matrix. This leads to nonradiative energy transfer (NRET) between the co-integrated nanocrystals in the host matrix. Here we systematically studied the resulting NRET process by varying donor and acceptor concentrations in the powders. We observed that NRET is a strong function of both of the nanocrystal concentrations and that NRET efficiency increases with increasing acceptor concentration. Nevertheless, with increasing donor concentration in the powders, NRET efficiency was found to first increase (up to a maximum level of 53.9%) but then to decrease. As a device demonstrator, we employed these NRET-improved nanocrystal powders as color-converters on blue light-emitting diodes (LEDs), with the resulting hybrid LED exhibiting a luminous efficiency >70 lm/Welect. The proposed excitonic nanocrystal powders potentially hold great promise for quality lighting and color enrichment applications. PMID:26832600 13. Brine Inclusions Migration in Intact Salt Crystals under Thermal Gradient NASA Astrophysics Data System (ADS) Caporuscio, F.; Boukhalfa, H. 2013-12-01 The behavior of water contained in rock salt under the influence of thermal gradients is critical to the performance of salt as a medium for the disposal of nuclear waste. Water contained in salt can be present as discrete inclusions within intact salt crystals, at the interface between salt crystals and aggregates, and also as hydration water and structural water present in accessory minerals present in salt. Water content in pure halite salt usually rages from 0.1 to 0.5 wt. % but is significantly higher in clay rich salt, for which water content can be up to several wt. %. Under the influence of thermal gradients brine inclusions and water associated to the accessory mineral is mobilized. Previous investigations have shown brine inclusions tend to move towards the heat source through a mechanism that involves the dissolution of salt at the hot face of the brine inclusion and its precipitation at the colder side of the inclusion. Uncertainties remain on the exact parameters that define the rate of brine migration and whether it truly migrates to towards the heat source. We performed studies under controlled thermal gradients to examine the behavior of brine inclusions in single salt crystals obtained from the underground salt mine at the Waste Isolation Power Plant (WIPP). We found that the behavior of the brine inclusions under thermal gradients is dependent on the thermal gradient magnitude and the nature of the inclusion. Full inclusions (liquid only) migrate predominantly towards the heat source, but when the inclusions are large and close to the surface they fracture the salt and release water near the surface. Inclusions that migrate towards the heat source migrate through a mechanism that involves the dissolution of salt at the hot side of the inclusion and its deposition along the migration path. SEM analysis of the migration pathways shows that brine migrates through the creation of a network of square shaped hollow channels of about 10 micron diameter. The behavior of two phase inclusions (liquid and gas) in a temperature gradient is distinctly different from that of full inclusions. The brine in the two phase inclusions still migrates towards the heat source; however, the vapor phase moves away from the heat source through narrow square shaped channels. The resulting salt crystals are much more fractured and have a distinct appearance compared to salt crystals with full inclusion. The composition of the salt deposited along the migration channels changes along migration pathway. At the start of the inclusion migration pathway the deposited salt is composed of a mixture of NaCl, MgCl2, and CaCl2, with minor other elements. However, as the brine migrates towards the heat source its composition changes and it gets enriched in NaCl. SEM images and migration channels 14. Tank waste chemistry: A new understanding of waste aging SciTech Connect Babad, H.; Camaioni, D.M.; Lilga, M.A.; Samuels, W.D.; Strachan, D.M. 1993-02-01 There is concern about the risk of uncontrolled exothermic reactions(s) in Hanford Site waste tanks containing NO{sub 3{sup minus}}/NO{sub 2{sup minus}} based salts and/or metal hydroxide sludges in combination with organics or ferrocyanides. However, gradual oxidation of the waste in the tanks to less reactive species appears to have reduced the risk. In addition, wastes sampled to date contain sufficiently large quantities of water so that propagation reactions are highly unlikely. This report details an investigation into the risk of an uncontrolled exothermic reaction in Hanford Site high-activity water tanks. 15. Contribution of the TRPV1 channel to salt taste quality in mice as assessed by conditioned taste aversion generalization and chorda tympani nerve responses PubMed Central Smith, Kimberly R.; Treesukosol, Yada; Paedae, A. Brennan; Contreras, Robert J. 2012-01-01 In rodents, at least two transduction mechanisms are involved in salt taste: 1) the sodium-selective epithelial sodium channel, blocked by topical amiloride administration, and 2) one or more amiloride-insensitive cation-nonselective pathways. Whereas electrophysiological evidence from the chorda tympani nerve (CT) has implicated the transient receptor potential vanilloid-1 (TRPV1) channel as a major component of amiloride-insensitive salt taste transduction, behavioral results have provided only equivocal support. Using a brief-access taste test, we examined generalization profiles of water-deprived C57BL/6J (WT) and TRPV1 knockout (KO) mice conditioned (via LiCl injection) to avoid 100 μM amiloride-prepared 0.25 M NaCl and tested with 0.25 M NaCl, sodium gluconate, KCl, NH4Cl, 6.625 mM citric acid, 0.15 mM quinine, and 0.5 M sucrose. Both LiCl-injected WT and TRPV1 KO groups learned to avoid NaCl+amiloride relative to controls, but their generalization profiles did not differ; LiCl-injected mice avoided the nonsodium salts and quinine suggesting that a TRPV1-independent pathway contributes to the taste quality of the amiloride-insensitive portion of the NaCl signal. Repeating the experiment but doubling all stimulus concentrations revealed a difference in generalization profiles between genotypes. While both LiCl-injected groups avoided the nonsodium salts and quinine, only WT mice avoided the sodium salts and citric acid. CT responses to these stimuli and a concentration series of NaCl and KCl with and without amiloride did not differ between genotypes. Thus, in our study, TRPV1 did not appear to contribute to sodium salt perception based on gustatory signals, at least in the CT, but may have contributed to the oral somatosensory features of sodium. PMID:23054171 16. Development of iron phosphate ceramic waste form to immobilize radioactive waste solution SciTech Connect Choi, Jongkwon; Um, Wooyong; Choung, Sungwook 2014-05-09 The objective of this research was to develop an iron phosphate ceramic (IPC) waste form using converter slag obtained as a by-product of the steel industry as a source of iron instead of conventional iron oxide. Both synthetic off-gas scrubber solution containing technetium-99 (or Re as a surrogate) and LiCl-KCl eutectic salt, a final waste solution from pyrochemical processing of spent nuclear fuel, were used as radioactive waste streams. The IPC waste form was characterized for compressive strength, reduction capacity, chemical durability, and contaminant leachability. Compressive strengths of the IPC waste form prepared with different types of waste solutions were 16 MPa and 19 MPa for LiCl-KCl eutectic salt and the off-gas scrubber simulant, respectively, which meet the minimum compressive strength of 3.45 MPa (500 psi) for waste forms to be accepted into the radioactive waste repository. The reduction capacity of converter slag, a main dry ingredient used to prepare the IPC waste form, was 4,136 meq/kg by the Ce(IV) method, which is much higher than those of the conventional Fe oxides used for the IPC waste form and the blast furnace slag materials. Average leachability indexes of Tc, Li, and K for the IPC waste form were higher than 6.0, and the IPC waste form demonstrated stable durability even after 63-day leaching. In addition, the Toxicity Characteristic Leach Procedure measurements of converter slag and the IPC waste form with LiCl-KCl eutectic salt met the universal treatment standard of the leachability limit for metals regulated by the Resource Conservation and Recovery Act. This study confirms the possibility of development of the IPC waste form using converter slag, showing its immobilization capability for radionuclides in both LiCl-KCl eutectic salt and off-gas scrubber solutions with significant cost savings. 17. Development of iron phosphate ceramic waste form to immobilize radioactive waste solution NASA Astrophysics Data System (ADS) Choi, Jongkwon; Um, Wooyong; Choung, Sungwook 2014-09-01 The objective of this research was to develop an iron phosphate ceramic (IPC) waste form using converter slag obtained as a by-product of the steel industry as a source of iron instead of conventional iron oxide. Both synthetic off-gas scrubber solution containing technetium-99 (or Re as a surrogate) and LiCl-KCl eutectic salt, a final waste solution from pyrochemical processing of spent nuclear fuel, were used as radioactive waste streams. The IPC waste form was characterized for compressive strength, reduction capacity, chemical durability, and contaminant leachability. Compressive strengths of the IPC waste form prepared with different types of waste solutions were 16 MPa and 19 MPa for LiCl-KCl eutectic salt and the off-gas scrubber simulant, respectively, which meet the minimum compressive strength of 3.45 MPa (500 psi) for waste forms to be accepted into the radioactive waste repository. The reduction capacity of converter slag, a main dry ingredient used to prepare the IPC waste form, was 4136 meq/kg by the Ce(IV) method, which is much higher than those of the conventional Fe oxides used for the IPC waste form and the blast furnace slag materials. Average leachability indexes of Tc, Li, and K for the IPC waste form were higher than 6.0, and the IPC waste form demonstrated stable durability even after 63-day leaching. In addition, the Toxicity Characteristic Leach Procedure measurements of converter slag and the IPC waste form with LiCl-KCl eutectic salt met the universal treatment standard of the leachability limit for metals regulated by the Resource Conservation and Recovery Act. This study confirms the possibility of development of the IPC waste form using converter slag, showing its immobilization capability for radionuclides in both LiCl-KCl eutectic salt and off-gas scrubber solutions with significant cost savings. 18. Engineered waste-package-system design specification SciTech Connect Not Available 1983-05-01 This report documents the waste package performance requirements and geologic and waste form data bases used in developing the conceptual designs for waste packages for salt, tuff, and basalt geologies. The data base reflects the latest geotechnical information on the geologic media of interest. The parameters or characteristics specified primarily cover spent fuel, defense high-level waste, and commercial high-level waste forms. The specification documents the direction taken during the conceptual design activity. A separate design specification will be developed prior to the start of the preliminary design activity. 19. Results of Analyses of Tank 37H Criticality Salt Samples (HTK-493 and 494) SciTech Connect Swingle, R.F. 2002-07-10 High Level Waste Division (HLWD) personnel requested Savannah River Technology Center (SRTC) support in analyzing samples of salt cake from Tank 37H as input to a Nuclear Criticality Safety Evaluation (NCSE) in support of dissolution of the salt in the tank. Two salt core samples (HTK-493 and 494) were pulled from Tank 37H and transferred to the SRTC Shielded Cells Facility on April 2, 2002 for analyses of species that may affect criticality as the salt is dissolved. The core samples were composited and prepared for analysis of the as-received sample and of insoluble solids separated from as-received sample. Results of the analyses are presented herein. Results are given for the as-received salt cake, for insoluble solids separated from the core sample both with and without a correction for soluble solids remaining after dissolution, and for wash water produced by dissolving the soluble salts in inhibited water. 20. Brine reuse in ion-exchange softening: salt discharge, hardness leakage, and capacity tradeoffs. PubMed Flodman, Hunter R; Dvorak, Bruce I 2012-06-01 Ion-exchange water softening results in the discharge of excess sodium chloride to the aquatic environment during the regeneration cycle. In order to reduce sodium chloride use and subsequent discharge from ion-exchange processes, either brine reclaim operations can be implemented or salt application during regeneration can be reduced. Both result in tradeoffs related to loss of bed volumes treated per cycle and increased hardness leakage. An experimentally validated model was used to compare concurrent water softening operations at various salt application quantities with and without the direct reuse of waste brine for treated tap water of typical midwestern water quality. Both approaches were able to reduce salt use and subsequent discharge. Reducing salt use and discharge by lowering the salt application rate during regeneration consequently increased hardness leakage and decreased treatment capacity. Single or two tank brine recycling systems are capable of reducing salt use and discharge without increasing hardness leakage, although treatment capacity is reduced. PMID:22866395 1. Concentration and precipitation of NaCl and KCl from salt cake leach solutions by electrodialysis SciTech Connect Sreenivasarao, K; Patsiogiannis, F.; Hryn, J.N. 1997-02-09 Electrodialysis was investigated for cost-effective recovery of salt from salt cake leach solutions. (Salt cake is a waste stream generated by the aluminum industry during treatment of aluminum drosses and scrap.) We used a pilot-scale electrodialysis stack of 5 membrane pairs, each with an effective area of 0.02 m{sup 2}. The diluate stream contained synthetic NaCl, KCl,mixtures of NaCl and KCl, and actual salt cake leach solutions (mainly NaCl and KCl, with small amounts of MgCl{sub 2}). We concentrated and precipitated NaCl and KCl salts from the concentrate steam when the initial diluate stream concentration was 21.5 to 28.8 wt% NaCl and KCl. We found that water transferring through the membranes was a significant factor in overall efficiency of salt recovery by electrodialysis. 2. Savannah River Site Salt Processing Final Supplemental Environmental Impact Statement SciTech Connect N /A 2001-07-20 DOE prepared this Supplemental Environmental Impact Statement (SEIS) on alternatives for separating the high-activity fraction from the low-activity fraction of the high-level radioactive waste salt solutions now stored in underground tanks at the Savannah River Site (SRS) near Aiken, South Carolina. The high-activity fraction of the high-level waste (HLW) salt solution would then be vitrified in the Defense Waste Processing Facility (DWPF) and stored until it could be disposed of as HLW in a geologic repository. The low-activity fraction would be disposed of as low-level waste (saltstone) in vaults at SRS. A process to separate the high-activity and low-activity waste fractions of the HLW salt solutions is needed to replace the In-Tank Precipitation (ITP) process which, as presently configured, cannot achieve production goals and safety requirements for processing HLW. This SEIS analyzes the impacts of constructing and operating facilities for four alternative processing technologies--Small Tank Precipitation, Ion Exchange, Solvent Extraction, and Direct Disposal in Grout--and the No Action Alternative. Solvent Extraction is DOE's preferred alternative. Sites for locating processing facilities within S and Z Areas at SRS are identified. Because replacing the ITP process constitutes a substantial change to the HLW salt processing operation of the DWPF, as evaluated in a 1994 SEIS (DOE/SEIS-0082-S) to the 1982 DWPF EIS (DOE/EIS-0082), DOE prepared this second SEIS to evaluate the potential environmental impacts of alternatives to the ITP process. 3. Constitutive modeling of salt behavior: State of the technology SciTech Connect Munson, D.E.; Wawersik, W.R. 1992-01-01 The modern investigation of the thermomechanical behavior of salt started in the mid-1930s and, for what appears to be a very narrow discipline, salt mechanics has acquired considerable technical depth and sophistication. The last three decades have been especially productive in constitutive model development and laboratory investigations of time-dependent creep behavior. This has been largely due ot anticipated use of domal or bedded salt deposits as sites for radioactive waste repositories and to expanded need for hydrocarbon and feedback storage caverns. Salt is an interesting material, in that it is metal-like; and, therefore, constitutive modeling can draw upon a large body of metal deformation information to arrive at appropriate models of behavior. Testing apparatus and methods have centered on either uniaxial or triaxial compression to obtain steady state and transient creep responses. Flow and fracture potentials have been defined. Validation attempts of the models against field data, although limited, have proved promising. The objective here is to summarize the state-of-the-technology of the constitutive modeling of salt behavior or salt mechanics. 4. Constitutive modeling of salt behavior: State of the technology SciTech Connect Munson, D.E.; Wawersik, W.R. 1992-01-01 The modern investigation of the thermomechanical behavior of salt started in the mid-1930's and, for what appears to be a very narrow discipline, salt mechanics'' has acquired considerable technical depth and sophistication. The last three decades have been especially productive in constitutive model development and laboratory investigations of time-dependent creep behavior. This has been largely due ot anticipated use of domal or bedded salt deposits as sites for radioactive waste repositories and to expanded need for hydrocarbon and feedback storage caverns. Salt is an interesting material, in that it is metal-like''; and, therefore, constitutive modeling can draw upon a large body of metal deformation information to arrive at appropriate models of behavior. Testing apparatus and methods have centered on either uniaxial or triaxial compression to obtain steady state and transient creep responses. Flow and fracture potentials have been defined. Validation attempts of the models against field data, although limited, have proved promising. The objective here is to summarize the state-of-the-technology of the constitutive modeling of salt behavior or salt mechanics.'' 5. Reduction of perchlorate and nitrate by salt tolerant bacteria. PubMed Okeke, Benedict C; Giblin, Tara; Frankenberger, William T 2002-01-01 Spent regenerant brine from ion-exchange technology for the removal of perchlorate and nitrate produces a high salt waste stream, which requires remediation before disposal. Bioremediation is an attractive treatment option. In this study, we enriched for salt tolerant bacteria from sediments from Cargill salt evaporation facility (California, USA), the Salton Sea (California, USA), and a high density hydrocarbon oxidizing bacterial cocktail. The bacterial cocktail enrichment culture reduced ClO4- from 500 to 260 mg 1 in 4 weeks. Salt tolerant bacterial isolates from the enrichment cultures and two denitrifying salt tolerant bacteria, Haloferax denitrificans and Parococcus halodenitricans, substantially reduced perchlorate. The highest rate of perchlorate removal was recorded with the isolate, Citrobacter sp.: 32% reduction in 1 week. This bacterium substantially reduced perchlorate in 0-5% NaCl solutions and maximally at 30 degrees C and at an initial pH 7.5. In simulated brines containing 7.5% total solids, the Citrobacter sp. significantly reduced both perchlorate and nitrate with 34.9 and 15.6% reduction, respectively, in 1 week. Coculture of a potent perchlorate reducing, non-salt tolerant (non-saline) bacterium, perclace and the Citrobacter sp. proved most effective for perchlorate removal in the brine (46.4% in 1 week). This study demonstrates that both anions can be reduced in treatment of brines from ion exchange systems. PMID:12009133 6. Burn wastes cost-effectively SciTech Connect Leite, O.C. 1998-05-01 Chemical processes generate side reaction compounds and other wastes. Many of these residuals are now considered toxic and hazardous. Consequently, hydrocarbon processing companies are seeking safe and economical methods to treat and discard unwanted byproducts. Burning/incinerating hazardous liquid wastes is an acceptable alternative to converting residuals such as halogenated compounds, inorganic salts, sulfur compounds, etc., into safely disposed of compounds. Another benefit from incineration is recovering waste-heat energy from hot incinerator flue gases. Several examples describe how to design an incineration unit that cost-effectively recovers thermal energy as steam. Equally important, these designs must meet stringent air emission requirements and handle corrosive compounds at extreme temperatures. 7. Salt associated with oil and gas activities: Remediation and related issues SciTech Connect Deuel, L.E. Jr. 1995-12-31 Salt, or brine, are generic terms for the chemical compound sodium chloride. We all know it as a seasoning agent and preservative for foods. Ancient cultures and civilizations actually used salt as money for barter and trade for goods and services. Presently, salt in the oil patch can still be equated with money, but in terms of capital outlay rather than barter and trade. Hazardous wastes, heavy metals, hydrocarbons, and other substances, are perceived by the public as a greater environmental threat than salt, and thus receive greater corporate attention. However, in terms of actual damage to the environment and capital outlay necessary for remediation, all other constituents pale in comparison to salt. The primary source of salt associated with oil and gas activities is produced water, with produced solids, drilling muds, and cuttings a very distant second. Produced waters contain salt concentrations ranging from trace levels to 2.6 pounds per gallon of water. The average for production provinces in the continental United States is about 0.7 pounds per gallon. Salt wasting of land is considered complete and irreparable at a salt concentration in soil of 7,500 ppm. Salt damaged soils are classed as either saline, saline/sodic, or sodic. The only way to reclaim a saline soil is to flush it with fresh water. There is no miracle product or magic elixir that can do this job. A saline/sodic soil contains both excess salt and excess exchangeable (adsorbed) sodium. Remediation of a saline/sodic soil requires the addition of amendments to address the exchangeable sodium, prior to the addition of fresh water to flush the excess salt. Sodic soils are the most difficult to remediate because of their dispersed nature, and the innate difficulty of placing amendments in contact with the soil particles necessary for the exchange processes to occur. 8. Determination of Thermal Conductivity of Reconsolidated Crushed Salt to 200C NASA Astrophysics Data System (ADS) Urquhart, A.; Bauer, S. J.; Hansen, F. D. 2012-12-01 Thermal, mechanical, and fluid transport properties of reconsolidating granular salt are important for design, analysis and performance assessment of potential salt repositories for heat-generating nuclear waste. Properties such as thermal conductivity (?) and permeability (k) are functions of porosity. To inform salt repository evaluations, where salt creep at elevated pressures and temperatures will reduce the porosity of salt enclosures, we have undertaken an experimental program to determine ? of reconsolidated granular salt as a function of porosity, with the secondary intent of quantifying temperature dependence over a range of temperatures from 100-200C. Mine-run salt from the Waste Isolation Pilot Program (WIPP) was first dried at 100C until weight loss ceased. The disaggregated salt was compacted at room temperature by quasistatic die compression into samples with porosities ranging from 40% (unconsolidated) to 5%. Thermal conductivity was also measured on intact domal salt, WIPP bedded salt, and commercially available salt licks in order to determine ? in salt with porosity as low as 1-8%. For every sample, measurements were made at 100, 125, 150, 175, and 200C using a guarded heat flow meter. Thermal conductivity was found to decrease by a factor of about 4-5 for the porosity range studied. The ? versus porosity relationship is well represented by a simple mixture model. A temperature dependence is also observed, the effect of which increases at lower porosities. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000. 9. Aqueous Biphasic Systems Based on Salting-Out Polyethylene Glycol or Ionic Solutions: Strategies for Actinide or Fission Product Separations SciTech Connect Rogers, Robin D.; Gutowski, Keith E.; Griffin, Scott T.; Holbrey, John D. 2004-03-29 Aqueous biphasic systems can be formed by salting-out (with kosmotropic, waterstructuring salts) water soluble polymers (e.g., polyethylene glycol) or aqueous solutions of a wide range of hydrophilic ionic liquids based on imidazolium, pyridinium, phosphonium and ammonium cations. The use of these novel liquid/liquid biphases for separation of actinides or other fission products associated with nuclear wastes (e.g., pertechnetate salts) has been demonstrated and will be described in this presentation. 10. Molten salt lithium cells DOEpatents Raistrick, Ian D.; Poris, Jaime; Huggins, Robert A. 1982-02-09 Lithium-based cells are promising for applications such as electric vehicles and load-leveling for power plants since lithium is very electropositive and light weight. One type of lithium-based cell utilizes a molten salt electrolyte and is operated in the temperature range of about 400.degree.-500.degree. C. Such high temperature operation accelerates corrosion problems and a substantial amount of energy is lost through heat transfer. The present invention provides an electrochemical cell (10) which may be operated at temperatures between about 100.degree.-170.degree. C. Cell (10) comprises an electrolyte (16), which preferably includes lithium nitrate, and a lithium or lithium alloy electrode (12). 11. Molten salt lithium cells DOEpatents Raistrick, Ian D.; Poris, Jaime; Huggins, Robert A. 1983-01-01 Lithium-based cells are promising for applications such as electric vehicles and load-leveling for power plants since lithium is very electropositive and light weight. One type of lithium-based cell utilizes a molten salt electrolyte and is operated in the temperature range of about 400.degree.-500.degree. C. Such high temperature operation accelerates corrosion problems and a substantial amount of energy is lost through heat transfer. The present invention provides an electrochemical cell (10) which may be operated at temperatures between about 100.degree.-170.degree. C. Cell (10) comprises an electrolyte (16), which preferably includes lithium nitrate, and a lithium or lithium alloy electrode (12). 12. Molten salt lithium cells DOEpatents Raistrick, I.D.; Poris, J.; Huggins, R.A. 1980-07-18 Lithium-based cells are promising for applications such as electric vehicles and load-leveling for power plants since lithium is very electropositive and light weight. One type of lithium-based cell utilizes a molten salt electrolyte and is operated in the temperature range of about 400 to 500/sup 0/C. Such high temperature operation accelerates corrosion problems and a substantial amount of energy is lost through heat transfer. The present invention provides an electrochemical cell which may be operated at temperatures between about 100 to 170/sup 0/C. The cell is comprised of an electrolyte, which preferably includes lithium nitrate, and a lithium or lithium alloy electrode. 13. Thermal-hydraulics of internally heated molten salts and application to the Molten Salt Fast Reactor NASA Astrophysics Data System (ADS) Fiorina, Carlo; Cammi, Antonio; Luzzi, Lelio; Mikityuk, Konstantin; Ninokata, Hisashi; Ricotti, Marco E. 2014-04-01 The Molten Salt Reactors (MSR) are an innovative kind of nuclear reactors and are presently considered in the framework of the Generation IV International Forum (GIF-IV) for their promising performances in terms of low resource utilization, waste minimization and enhanced safety. A unique feature of MSRs is that molten fluoride salts play the distinctive role of both fuel (heat source) and coolant. The presence of an internal heat generation perturbs the temperature field and consequences are to be expected on the heat transfer characteristics of the molten salts. In this paper, the problem of heat transfer for internally heated fluids in a straight circular channel is first faced on a theoretical ground. The effect of internal heat generation is demonstrated to be described by a corrective factor applied to traditional correlations for the Nusselt number. It is shown that the corrective factor can be fully characterized by making explicit the dependency on Reynolds and Prandtl numbers. On this basis, a preliminary correlation is proposed for the case of molten fluoride salts by interpolating the results provided by an analytic approach previously developed at the Politecnico di Milano. The experimental facility and the related measuring procedure for testing the proposed correlation are then presented. Finally, the developed correlation is used to carry out a parametric investigation on the effect of internal heat generation on the main out-of-core components of the Molten Salt Fast Reactor (MSFR), the reference circulating-fuel MSR design in the GIF-IV. The volumetric power determines higher temperatures at the channel wall, but the effect is significant only in case of large diameters and/or low velocities. 14. Electrolyte salts for nonaqueous electrolytes DOEpatents Amine, Khalil; Zhang, Zhengcheng; Chen, Zonghai 2012-10-09 Metal complex salts may be used in lithium ion batteries. Such metal complex salts not only perform as an electrolyte salt in a lithium ion batteries with high solubility and conductivity, but also can act as redox shuttles that provide overcharge protection of individual cells in a battery pack and/or as electrolyte additives to provide other mechanisms to provide overcharge protection to lithium ion batteries. The metal complex salts have at least one aromatic ring. The aromatic moiety may be reversibly oxidized/reduced at a potential slightly higher than the working potential of the positive electrode in the lithium ion battery. The metal complex salts may also be known as overcharge protection salts. 15. Radioactive air emissions notice of construction use of a portable exhauster on single shell tanks (SSTs) during salt well pumping SciTech Connect GRANDO, C.J. 1999-11-18 This document serves as a notice of construction (NOC), pursuant to the requirements of Washington Administrative Code (WAC) 246-247-060, and as a request for approval to construct, pursuant to 40 Code of Federal Regulations (CFR) 61.07, portable exhausters for use on single-shell tanks (SSTs) during salt well pumping. Table 1-1 lists 18 SSTs covered by this NOC. This NOC also addresses other activities that are performed in support of salt well pumping but do not require the application of a portable exhauster. Specifically this NOC analyzes the following three activities that have the potential for emissions. (1) Salt well pumping (i.e., the actual transferring of waste from one tank to another) under nominal tank operating conditions. Nominal tank operating conditions include existing passive breathing rates. (2) Salt well pumping (the actual transferring of waste from one tank to another) with use of a portable exhauster. (3) Use of a water lance on the waste to facilitate salt well screen and salt well jet pump installation into the waste. This activity is to be performed under nominal (existing passive breathing rates) tank operating conditions. The use of portable exhausters represents a cost savings because one portable exhauster can be moved back and forth between SSTs as schedules for salt well pumping dictate. A portable exhauster also could be used to simultaneously exhaust more than one SST during salt well pumping. 16. Batteries using molten salt electrolyte DOEpatents Guidotti, Ronald A. (Albuquerque, NM) 2003-04-08 An electrolyte system suitable for a molten salt electrolyte battery is described where the electrolyte system is a molten nitrate compound, an organic compound containing dissolved lithium salts, or a 1-ethyl-3-methlyimidazolium salt with a melting temperature between approximately room temperature and approximately 250.degree. C. With a compatible anode and cathode, the electrolyte system is utilized in a battery as a power source suitable for oil/gas borehole applications and in heat sensors. 17. Electrochromic Salts, Solutions, and Devices DOEpatents Burrell, Anthony K. (Los Alamos, NM); Warner, Benjamin P. (Los Alamos, NM); McClesky, T. Mark (Los Alamos, NM) 2008-11-11 Electrochromic salts. Electrochromic salts of dicationic viologens such as methyl viologen and benzyl viologen associated with anions selected from bis(trifluoromethylsulfonyl)imide, bis(perfluoroethylsulfonyl)imide, and tris(trifluoromethylsulfonyl)methide are produced by metathesis with the corresponding viologen dihalide. They are highly soluble in molten quarternary ammonium salts and together with a suitable reductant provide electrolyte solutions that are used in electrochromic windows. 18. Salt appetite in the elderly. PubMed Hendi, Khadeja; Leshem, Micah 2014-11-28 The present study investigated whether salt appetite in the elderly is impaired similar to thirst because of the commonality of their physiological substrates and whether alterations in salt appetite are related to mood. Elderly (65-85 years, n 30) and middle-aged (45-58 years, n 30) men and women were compared in two test sessions. Thirst, psychophysical ratings of taste solutions, dietary Na and energy intakes, seasoning with salt and sugar, number of salty and sweet snacks consumed, preferred amounts of salt in soup and sugar in tea, and an overall measure of salt appetite and its relationship with mood, nocturia and sleep were measured. Elderly participants were found to be less thirsty and respond less to thirst. In contrast, no impairment of salt appetite was found in them, and although they had a reduced dietary Na intake, it dissipated when corrected for their reduced dietary energy intake. Diet composition and Na intake were found to be similar in middle-aged and elderly participants, despite the lesser intake in elderly participants. There were no age-related differences in the intensity of taste or hedonic profile of Na, in salting habits, in tests of salting soup, or number of salty snacks consumed. No relationship of any measure of salt appetite with mood measured by the Positive and Negative Affect Schedule, frequency of nocturia, or sleep duration was observed. The age-related impairment of the physiology of mineralofluid regulation, while compromising thirst and fluid intake, spares salt appetite, suggesting that salt appetite in humans is not regulated physiologically. Intact salt appetite in the elderly might be utilised judiciously to prevent hyponatraemia, increase thirst and improve appetite. PMID:25287294 19. Electrochromic salts, solutions, and devices DOEpatents Burrell, Anthony K. (Los Alamos, NM); Warner, Benjamin P. (Los Alamos, NM); McClesky,7,064,212 T. Mark (Los Alamos, NM) 2006-06-20 Electrochromic salts. Electrochromic salts of dicationic viologens such as methyl viologen and benzyl viologen associated with anions selected from bis(trifluoromethylsulfonyl)imide, bis(perfluoroethylsulfonyl)imide, and tris(trifluoromethylsulfonyl)methide are produced by metathesis with the corresponding viologen dihalide. They are highly soluble in molten quarternary ammonium salts and together with a suitable reductant provide electrolyte solutions that are used in electrochromic windows. 20. Electrochromic Salts, Solutions, and Devices DOEpatents Burrell, Anthony K. (Los Alamos, NM); Warner, Benjamin P. (Los Alamos, NM); McClesky, T. Mark (Los Alamos, NM) 2008-10-14 Electrochromic salts. Electrochromic salts of dicationic viologens such as methyl viologen and benzyl viologen associated with anions selected from bis(trifluoromethylsulfonyl)imide, bis(perfluoroethylsulfonyl)imide, and tris(trifluoromethylsulfonyl)methide are produced by metathesis with the corresponding viologen dihalide. They are highly soluble in molten quarternary ammonium salts and together with a suitable reductant provide electrolyte solutions that are used in electrochromic windows. 1. ENVIROCARE OF UTAH: EXPANDING WASTE ACCEPTANCE CRITERIA TO PROVIDE LOW-LEVEL AND MIXED WASTE DISPOSAL OPTIONS SciTech Connect Rogers, B.; Loveland, K. 2003-02-27 Envirocare of Utah operates a low-level radioactive waste disposal facility 80 miles west of Salt Lake City in Clive, Utah. Accepted waste types includes NORM, 11e2 byproduct material, Class A low-level waste, and mixed waste. Since 1988, Envirocare has offered disposal options for environmental restoration waste for both government and commercial remediation projects. Annual waste receipts exceed 12 million cubic feet. The waste acceptance criteria (WAC) for the Envirocare facility have significantly expanded to accommodate the changing needs of restoration projects and waste generators since its inception, including acceptable physical waste forms, radiological acceptance criteria, RCRA requirements and treatment capabilities, PCB acceptance, and liquids acceptance. Additionally, there are many packaging, transportation, and waste management options for waste streams acceptable at Envirocare. Many subcontracting vehicles are also available to waste generators for both government and commercial activities. 2. Hydrogen production under salt stress conditions by a freshwater Rhodopseudomonas palustris strain. PubMed Adessi, Alessandra; Concato, Margherita; Sanchini, Andrea; Rossi, Federico; De Philippis, Roberto 2016-03-01 Hydrogen represents a possible alternative energy carrier to face the growing request for energy and the shortage of fossil fuels. Photofermentation for the production of H2 constitutes a promising way for integrating the production of energy with waste treatments. Many wastes are characterized by high salinity, and polluted seawater can as well be considered as a substrate. Moreover, the application of seawater for bacterial culturing is considered cost-effective. The aims of this study were to assess the capability of the metabolically versatile freshwater Rhodopseudomonas palustris 42OL of producing hydrogen on salt-containing substrates and to investigate its salt stress response strategy, never described before. R. palustris 42OL was able to produce hydrogen in media containing up to 3 % added salt concentration and to grow in media containing up to 4.5 % salinity without the addition of exogenous osmoprotectants. While the hydrogen production performances in absence of sea salts were higher than in their presence, there was no significant difference in performances between 1 and 2 % of added sea salts. Nitrogenase expression levels indicated that the enzyme was not directly inhibited during salt stress, but a regulation of its expression may have occurred in response to salt concentration increase. During cell growth and hydrogen production in the presence of salts, trehalose was accumulated as a compatible solute; it protected the enzymatic functionality against salt stress, thus allowing hydrogen production. The possibility of producing hydrogen on salt-containing substrates widens the range of wastes that can be efficiently used in production processes. PMID:26762392 3. Dietary salt intake and hypertension. PubMed Ha, Sung Kyu 2014-06-01 Over the past century, salt has been the subject of intense scientific research related to blood pressure elevation and cardiovascular mortalities. Moderate reduction of dietary salt intake is generally an effective measure to reduce blood pressure. However, recently some in the academic society and lay media dispute the benefits of salt restriction, pointing to inconsistent outcomes noted in some observational studies. A reduction in dietary salt from the current intake of 9-12 g/day to the recommended level of less than 5-6 g/day will have major beneficial effects on cardiovascular health along with major healthcare cost savings around the world. The World Health Organization (WHO) strongly recommended to reduce dietary salt intake as one of the top priority actions to tackle the global non-communicable disease crisis and has urged member nations to take action to reduce population wide dietary salt intake to decrease the number of deaths from hypertension, cardiovascular disease and stroke. However, some scientists still advocate the possibility of increased risk of CVD morbidity and mortality at extremes of low salt intake. Future research may inform the optimal sodium reduction strategies and intake targets for general populations. Until then, we have to continue to build consensus around the greatest benefits of salt reduction for CVD prevention, and dietary salt intake reduction strategies must remain at the top of the public health agenda. PMID:25061468 4. Dietary Salt Intake and Hypertension PubMed Central 2014-01-01 Over the past century, salt has been the subject of intense scientific research related to blood pressure elevation and cardiovascular mortalities. Moderate reduction of dietary salt intake is generally an effective measure to reduce blood pressure. However, recently some in the academic society and lay media dispute the benefits of salt restriction, pointing to inconsistent outcomes noted in some observational studies. A reduction in dietary salt from the current intake of 9-12 g/day to the recommended level of less than 5-6 g/day will have major beneficial effects on cardiovascular health along with major healthcare cost savings around the world. The World Health Organization (WHO) strongly recommended to reduce dietary salt intake as one of the top priority actions to tackle the global non-communicable disease crisis and has urged member nations to take action to reduce population wide dietary salt intake to decrease the number of deaths from hypertension, cardiovascular disease and stroke. However, some scientists still advocate the possibility of increased risk of CVD morbidity and mortality at extremes of low salt intake. Future research may inform the optimal sodium reduction strategies and intake targets for general populations. Until then, we have to continue to build consensus around the greatest benefits of salt reduction for CVD prevention, and dietary salt intake reduction strategies must remain at the top of the public health agenda. PMID:25061468 5. Dietary salt and pullet development. PubMed Leeson, S; Summers, J D 1980-04-01 White Leghorn pullets were fed corn-soybean meal diets containing .25 (control), .175, or .1% supplemental salt to 20 weeks of age, followed by a regular laying diet to 36 weeks of age. Body weight and feed intake of the growing pullet was significantly reduced with .1% salt compared to other treatments, while in early production these same birds produced fewer eggs compared with birds fed .25% salt during rearing. It is concluded that manipulation of dietary salt levels may be used to delay sexual maturity. PMID:7375441 6. From science to compliance: Geomechanics studies of the Waste Isolation Pilot Plant SciTech Connect HANSEN,FRANCIS D. 2000-06-05 Mechanical and hydrological properties of salt provide excellent bases for geological isolation of hazardous materials. Regulatory certification of the Waste Isolation Pilot Plant (WIPP) testifies to the nearly ideal characteristics of bedded salt deposits in southeast New Mexico. The WIPP history includes decades of testing and scientific investigations, which have resulted in a comprehensive understanding of salt's mechanical deformational and hydrological properties over an applicable range of stresses and temperatures. Comprehensive evaluation of salt's favorable characteristics helped demonstrate regulatory compliance and ensure isolation of radioactive waste placed in a salt geological setting. 7. Salt deposits in Los Medanos area, Eddy and Lea counties, New Mexico USGS Publications Warehouse Jones, C.L.; with sections on Ground water hydrology, Cooley, Maurice E.; and Surficial Geology, Bachman, George Odell 1973-01-01 The salt deposits of Los Medanos area, in Eddy and Lea Counties, southeastern New Mexico, are being considered for possible use as a receptacle for radioactive wastes in a pilot-plant repository. The salt deposits of the area. are in three evaporite formations: the Castile, Salado, and Rustler Formations, in ascending order. The three formations are dominantly anhydrite and rock salt, but some gypsum, potassium ores, carbonate rock, and fine-grained clastic rocks are present. They have combined thicknesses of slightly more than 4,000 feet, of which roughly one-half belongs to the Salado. Both the Castile and the Rustler are-richer in anhydrite-and poorer in rock salt-than the Salado, and they provide this salt-rich formation with considerable Protection from any fluids which might be present in underlying or overlying rocks. The Salado Formation contains many thick seams of rock salt at moderate depths below the surface. The rock salt has a substantial cover of well-consolidated rocks, and it is very little deformed structurally. Certain geological details essential for Waste-storage purposes are unknown or poorly known, and additional study involving drilling is required to identify seams of rock salt suitable for storage purposes and to establish critical details of their chemistry, stratigraphy, and structure. 8. Schematic designs for penetration seals for a reference repository in bedded salt SciTech Connect Kelsall, P.C.; Case, J.B.; Meyer, D.; Coons, W.E. 1982-11-01 The isolation of radioactive wastes in geologic repositories requires that man-made penetrations such as shafts, tunnels, or boreholes are adequately sealed. This report describes schematic seal designs for a repository in bedded salt referenced to the straitigraphy of southeastern New Mexico. The designs are presented for extensive peer review and will be updated as site-specific conceptual designs when a site for a repository in salt has been selected. The principal material used in the seal system is crushed salt obtained from excavating the repository. It is anticipated that crushed salt will consolidate as the repository rooms creep close to the degree that mechanical and hydrologic properties will eventually match those of undisturbed, intact salt. For southeastern New Mexico salt, analyses indicate that this process will require approximately 1000 years for a seal located at the base of one of the repository shafts (where there is little increase in temperature due to waste emplacement) and approximately 400 years for a seal located in an access tunnel within the repository. Bulkheads composed of contrete or salt bricks are also included in the seal system as components which will have low permeability during the period required for salt consolidation. 9. Occlusion and ion exchange in the molten (lithium chloride + potassium chloride + alkaline earth chloride) salt + zeolite 4A system with alkaline earth chlorides of calcium and strontium, and in the molten (lithium chloride + potassium chloride + actinide chloride) salt + zeolite 4A system with the actinide chloride of uranium. SciTech Connect Lexa, D.; Chemical Engineering 2003-04-01 The interaction between molten salts of the type LiCl-KCl-MeCl n (Me=Ca, Sr, U; x{sub MeCLn} = to 0.45; and x {sub KCl}/x LiCl=0.69) and zeolite 4A have been studied at 823 K. The main interactions between these salts and zeolite are molten salt occlusion to form salt-loaded zeolite and ion exchange between the molten salt and salt-loaded zeolite. An irreversible chemical reaction has been observed in the LiCl-KCl-UCl{sup 3+}zeolite system. The extent of occlusion is a function of the concentration of MeCl n in the zeolite and is equal to 10{+-}1 Cl- per zeolite unit cell, (AlSiO{sub 4}){sub 12}, at infinite MeCl n dilution. The ion-exchange mole-fraction equilibrium constants (separation factors) with respect to Li are decreasing functions of the concentration of SrCl{sub 2} and UCl{sub 3}, but an increasing function of the concentration of CaCl{sub 2} in the zeolite. At infinite MeCl n dilution, they are equal to 0.9, 11.9, and 13 for CaCl{sub 2}, SrCl{sub 2}, and UCl{sub 3}, respectively. The standard ion-exchange chemical potentials are equal to -50.0, -84, and -101.1 kJ x mol-1 for Ca{sup 2+}, Sr{sup 2+}, and U{sup 3+}, respectively. 10. Hazardous solid waste from agriculture. PubMed Central Loehr, R C 1978-01-01 Large quantities of food processing, crop, forestry, and animal solid wastes are generated in the United States each year. The major components of these wastes are biodegradable. However, they also contain components such as nitrogen, human and animal pathogens, medicinals, feed additives, salts, and certain metals, that under uncontrolled conditions can be detrimental to aquatic, plant, animal, or human life. The most common method of disposal of these wastes is application to the land. Thus the major pathways for transmission of hazards are from and through the soil. Use of these wastes as animal feed also can be a pathway. While at this time there are no crises associated with hazardous materials in agricultural solid wastes, the potential for problems should not be underestimated. Manpower and financial support should be provided to obtain more detailed information in this area, esepcially to better delineate transport and dispersal and to determine and evaluate risks. PMID:367770 11. Salt pretreatment enhance salt tolerance in Zea mays L. seedlings. PubMed Tajdoost, S; Farboodnia, T; Heidari, R 2007-06-15 Recent molecular studies show that genetic factors of salt tolerance in halophytes exist in glycophytes too, but they are not active. If these plants expose to low level salt stress these factors may become active and cause plants acclimation to higher salt stresses. So because of the importance of these findings in this research the effect of salt pretreatment has been examined in Zea mays seedlings. To do the experiment four day old Zea mays seedlings (Var. single cross 704) pretreated with 50 mM NaCl for the period of 20 h. Then they were transferred to 200 and 300 mM NaCl for 48 h. At the end of treatment roots and shoots of seedlings were harvested separately. The changes of K+ -leakage, the amount of malondialdehyde, proline, soluble sugars and the Hill reaction rate were analyzed. The results indicated that the amount of K+ -leakage and malondialdehyde (MDA) have been increased because of salt-induced lipid peroxidation and membrane unstability. Soluble sugars and proline as osmoregulators has been increased in stress condition and in pretreated plants with NaCl were the highest. The rate of Hill reaction was reduced significantly in stressed plants. Therefore we concluded that salt stress causes serious physiological and biochemical damages in plants and salt pretreatment enhances tolerance mechanisms of plants and help them to tolerate salt stress and grow on salty environments. PMID:19093451 12. Water, Vapor, and Salt Dynamics in a Hot Repository SciTech Connect Bahrami, Davood; Danko, George; Walton, John 2007-07-01 The purpose of this paper is to report the results of a new model study examining the high temperature nuclear waste disposal concept at Yucca Mountain using MULTIFLUX, an integrated in-drift- and mountain-scale thermal-hydrologic model. The results show that a large amount of vapor flow into the drift is expected during the period of above-boiling temperatures. This phenomenon makes the emplacement drift a water/moisture attractor during the above-boiling temperature operation. The evaporation of the percolation water into the drift gives rise to salt accumulation in the rock wall, especially in the crown of the drift for about 1500 years in the example. The deposited salts over the drift footprint, almost entirely present in the fractures, may enter the drift either by rock fall or by water drippage. During the high temperature operation mode, the barometric pressure variation creates fluctuating relative humidity in the emplacement drift with a time period of approximately 10 days. Potentially wet and dry conditions and condensation on salt-laden drift wall sections may adversely affect the storage environment. Salt accumulations during the above-boiling temperature operation must be sufficiently addressed to fully understand the waste package environment during the thermal period. Until the questions are resolved, a below-boiling repository design is favored where the Alloy-22 will be less susceptible to localized corrosion. (authors) 13. Mercury separation from mixed wastes. Annual report SciTech Connect Taylor, P.A.; Klasson, K.T.; Corder, S.L.; Carlson, T.R.; McCandless, K.R. 1995-11-01 This is an assessment of new sorbents for removing Hg from wastes at US DOE sites. Four aqueous wastes were used for the laboratory tests: a simulant of a high-salt, acidic waste currently stored at INEL, a simulant of a high-salt, alkaline waste stored at Savannah River (SRS), a dilute LiOH solution stored at Y-12, and a low-salt, neutral groundwater generated at Y-12. Eight adsorbents covering a wide range of cost and capability were tested. Screening tests identified the most promising adsorbents, and column tests were performed using at least two adsorbents for each waste stream. No one adsorbent is effective in all of these waste streams. Based on loading capacity and compatibility, the most effect adsorbents to date are SuperLig 618 for the INEL tank waste simulant, Mersorb and Ionac SR-3 for the SRS tank waste simulant, Durasil 70 and Ionac SR-3 for the LiOH solution, and Ionac SR-3, followed by Ionac SR-4 and Mersorb, for the Y-12 groundwater. 14. TOTAL RECYCLE SYSTEMS FOR PETROCHEMICAL WASTE BRINES CONTAINING REFRACTORY CONTAMINANTS EPA Science Inventory Petrochemical wastewaters containing relatively high concentrations of salt and refractory organics were selected to study their feasibility for total recycle. A combination of reverse osmosis and electrodialysis was operated as a hybrid system using the pretreated wastes to prod... 15. Cloning and Functional Characterization of a Vacuolar Na+/H+ Antiporter Gene from Mungbean (VrNHX1) and Its Ectopic Expression Enhanced Salt Tolerance in Arabidopsis thaliana PubMed Central Mishra, Sagarika; Alavilli, Hemasundar; Lee, Byeong-ha; Panda, Sanjib Kumar; Sahoo, Lingaraj 2014-01-01 Plant vacuolar NHX exchangers play a significant role in adaption to salt stress by compartmentalizing excess cytosolic Na+ into vacuoles and maintaining cellular homeostasis and ionic equilibrium. We cloned an orthologue of the vacuolar Na+/H+ antiporter gene, VrNHX1 from mungbean (Vigna radiata), an important Asiatic grain legume. The VrNHX1 (Genbank Accession number JN656211.1) contains 2095 nucleotides with an open reading frame of 1629 nucleotides encoding a predicted protein of 542 amino acids with a deduced molecular mass of 59.6 kDa. The consensus amiloride binding motif (84LFFIYLLPPI93) was observed in the third putative transmembrane domain of VrNHX1. Bioinformatic and phylogenetic analysis clearly suggested that VrNHX1 had high similarity to those of orthologs belonging to Class-I clade of plant NHX exchangers in leguminous crops. VrNHX1 could be strongly induced by salt stress in mungbean as the expression in roots significantly increased in presence of 200 mM NaCl with concomitant accumulation of total [Na+]. Induction of VrNHX1 was also observed under cold and dehydration stress, indicating a possible cross talk between various abiotic stresses. Heterologous expression in salt sensitive yeast mutant AXT3 complemented for the loss of yeast vacuolar NHX1 under NaCl, KCl and LiCl stress indicating that VrNHX1 was the orthologue of ScNHX1. Further, AXT3 cells expressing VrNHX1 survived under low pH environment and displayed vacuolar alkalinization analyzed using pH sensitive fluorescent dye BCECF-AM. The constitutive and stress inducible expression of VrNHX1 resulted in enhanced salt tolerance in transgenic Arabidopsis thaliana lines. Our work suggested that VrNHX1 was a salt tolerance determinant in mungbean. PMID:25350285 16. The Mixed Waste Management Facility. Preliminary design review SciTech Connect 1995-12-31 This document presents information about the Mixed Waste Management Facility. Topics discussed include: cost and schedule baseline for the completion of the project; evaluation of alternative options; transportation of radioactive wastes to the facility; capital risk associated with incineration; radioactive waste processing; scaling of the pilot-scale system; waste streams to be processed; molten salt oxidation; feed preparation; initial operation to demonstrate selected technologies; floorplans; baseline revisions; preliminary design baseline; cost reduction; and project mission and milestones. 17. Update on cavern disposal of NORM-contaminated oil field wastes. SciTech Connect Veil, J. A. 1998-09-22 Some types of oil and gas production and processing wastes contain naturally occurring radioactive material (NORM). If NORM is present at concentrations above regulatory levels in oil field waste, the waste requires special disposal practices. The existing disposal options for wastes containing NORM are limited and costly. Argonne National Laboratory has previously evaluated the feasibility, legality, risk and economics of disposing of nonhazardous oil field wastes, other than NORM waste, in salt caverns. Cavern disposal of nonhazardous oil field waste, other than NORM waste, is occurring at four Texas facilities, in several Canadian facilities, and reportedly in Europe. This paper evaluates the legality, technical feasibility, economics, and human health risk of disposing of NORM-contaminated oil field wastes in salt caverns as well. Cavern disposal of NORM waste is technically feasible and poses a very low human health risk. From a legal perspective, a review of federal regulations and regulations from several states indicated that there are no outright prohibitions against NORM disposal in salt caverns or other Class II wells, except for Louisiana which prohibits disposal of radioactive wastes or other radioactive materials in salt domes. Currently, however, only Texas and New Mexico are working on disposal cavern regulations, and no states have issued permits to allow cavern disposal of NORM waste. On the basis of the costs currently charged for cavern disposal of nonhazardous oil field waste (NOW), NORM waste disposal in caverns is likely to be cost competitive with existing NORM waste disposal methods when regulatory agencies approve the practice. 18. Large-scale dynamic compaction demonstration using WIPP salt: Fielding and preliminary results SciTech Connect Ahrens, E.H.; Hansen, F.D. 1995-10-01 Reconsolidation of crushed rock salt is a phenomenon of great interest to programs studying isolation of hazardous materials in natural salt geologic settings. Of particular interest is the potential for disaggregated salt to be restored to nearly an impermeable state. For example, reconsolidated crushed salt is proposed as a major shaft seal component for the Waste Isolation Pilot Plant (WIPP) Project. The concept for a permanent shaft seal component of the WIPP repository is to densely compact crushed salt in the four shafts; an effective seal will then be developed as the surrounding salt creeps into the shafts, further consolidating the crushed salt. Fundamental information on placement density and permeability is required to ensure attainment of the design function. The work reported here is the first large-scale compaction demonstration to provide information on initial salt properties applicable to design, construction, and performance expectations. The shaft seals must function for 10,000 years. Over this period a crushed salt mass will become less permeable as it is compressed by creep closure of salt surrounding the shaft. These facts preclude the possibility of conducting a full-scale, real-time field test. Because permanent seals taking advantage of salt reconsolidation have never been constructed, performance measurements have not been made on an appropriately large scale. An understanding of potential construction methods, achievable initial density and permeability, and performance of reconsolidated salt over time is required for seal design and performance assessment. This report discusses fielding and operations of a nearly full-scale dynamic compaction of mine-run WIPP salt, and presents preliminary density and in situ (in place) gas permeability results. 19. DEVELOPMENT OF AN INSOLUBLE SALT SIMULANT TO SUPPORT ENHANCED CHEMICAL CLEANING TESTS SciTech Connect Eibling, R 2008-05-23 The closure process for high level waste tanks at the Savannah River Site will require dissolution of the crystallized salts that are currently stored in many of the tanks. The insoluble residue from salt dissolution is planned to be removed by an Enhanced Chemical Cleaning (ECC) process. Development of a chemical cleaning process requires an insoluble salt simulant to support evaluation tests of different cleaning methods. The Process Science and Engineering section of SRNL has been asked to develop an insoluble salt simulant for use in testing potential ECC processes (HLE-TTR-2007-017). An insoluble salt simulant has been developed based upon the residues from salt dissolution of saltcake core samples from Tank 28F. The simulant was developed for use in testing SRS waste tank chemical cleaning methods. Based on the results of the simulant development process, the following observations were developed: (1) A composition based on the presence of 10.35 grams oxalate and 4.68 grams carbonate per 100 grams solids produces a sufficiently insoluble solids simulant. (2) Aluminum observed in the solids remaining from actual waste salt dissolution tests is probably precipitated from sodium aluminate due to the low hydroxide content of the saltcake. (3) In-situ generation of aluminum hydroxide (by use of aluminate as the Al source) appears to trap additional salts in the simulant in a manner similar to that expected for actual waste samples. (4) Alternative compositions are possible with higher oxalate levels and lower carbonate levels. (5) The maximum oxalate level is limited by the required Na content of the insoluble solids. (6) Periodic mixing may help to limit crystal growth in this type of salt simulant. (7) Long term storage of an insoluble salt simulant is likely to produce a material that can not be easily removed from the storage container. Production of a relatively fresh simulant is best if pumping the simulant is necessary for testing purposes. The insoluble salt simulant described in this report represents the initial attempt to represent the material which may be encountered during final waste removal and tank cleaning. The final selected simulant was produced by heating and evaporation of a salt slurry sample to remove excess water and promote formation and precipitation of solids with solubility characteristics which are consistent with actual tank insoluble salt samples. The exact anion composition of the final product solids is not explicitly known since the chemical components in the final product are distributed between the solid and liquid phases. By combining the liquid phase analyses and total solids analysis with mass balance requirements a calculated composition of assumed simple compounds was obtained and is shown in Table 0-1. Additional improvements to and further characterization of the insoluble salt simulant are possible. During the development of these simulants it was recognized that: (1) Additional waste characterization on the residues from salt dissolution tests with actual waste samples to determine the amount of species such as carbonate, oxalate and aluminosilicate would allow fewer assumptions to be made in constructing an insoluble salt simulant. (2) The tank history will impact the amount and type of insoluble solids that exist in the salt dissolution solids. Varying the method of simulant production (elevated temperature processing time, degree of evaporation, amount of mixing (shear) during preparation, etc.) should be tested. 20. Salt resistant crop plants. PubMed Roy, Stuart J; Negro, Snia; Tester, Mark 2014-04-01 Soil salinity is a major constraint to agriculture. To improve salinity tolerance of crops, various traits can be incorporated, including ion exclusion, osmotic tolerance and tissue tolerance. We review the roles of a range of genes involved in salt tolerance traits. Different tissues and cells are adapted for specific and often diverse function, so it is important to express the genes in specific cell-types and to pyramid a range of traits. Modern biotechnology (marker-assisted selection or genetic engineering) needs to be increasingly used to introduce the correct combination of genes into elite crop cultivars. Importantly, the effects of introduced genes need to be evaluated in the field to determine their effect on salinity tolerance and yield improvement. PMID:24679267
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.48774707317352295, "perplexity": 7601.454466454978}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-18/segments/1461860116587.6/warc/CC-MAIN-20160428161516-00014-ip-10-239-7-51.ec2.internal.warc.gz"}
https://resinsight.org/3d-main-window/surfaces/
# Surfaces ResInsight is capable of mapping properties and simulation results onto 3D surfaces imported from GOCAD and Petrel as specified below. ## Import and appearance of surfaces 3D surfaces from GOCAD and Petrel are imported into ResInsight by the menu option File->Import->Import Surfaces in the 3D Main Window. Imported surfaces are listed under the entry in the Project Tree. You can click on a surface to change i.e. the name or color of the surface, or change the file to read the surface data from. You can also change the depth of the surface by adding a depth offset (positive or negative). This could be used i.e. to import the same surface file multiple times and assign each instance of the surface a separate depth offset. When a depth offset is set, the offset value will be shown as part of the surface name in the project tree. ## Using folders To make it easier to organize the surfaces you import into your project, ResInsight supports creating surface folders in the project tree. To create a new folder, right-click on the top level Surfaces folder in the project tree and choose Add Folder. ## Surface Results Per default, ResInsight maps current Cell Result on a given surface. If another result is to be displayed, specify the result under Surface Results as shown below. Clicking a surface of a View in the Project Tree activates the Property Editor for allowing view settings as shown below. To change result displayed on the surface, please specify the desired result in the Result Reference section of the Property Editor. If you have modified a surface file using an external program, you can easily load the changes into ResInsight by using the reload surface feature. Bring up the right-click menu for the surface you want to reload and choose Reload. The views will automatically update with the new data (could take a few seconds). ## Create a copy You can easily create a copy of an existing surface by choosing Create Copy in the surface right-click menu. A new, identical surface will show up at the same level in the project tree. You can now give it a new name, change depth offset etc. ## Grid Case Surfaces In addition to importing surfaces from file, ResInsight can also generate grid case surfaces. You do that by bringing up the right-click menu for the project tree entry and choose Create Grid Case Surfaces. A grid case surface has the same properties as a surface imported from file, but instead of choosing which file the data should come from, you choose a source case, a slice direction and a slice index. The slice direction and index will be added to the surface name shown in the project tree. ## Exporting Surfaces ResInsight can export surfaces to the GOCAD TSurf file format. Grid case surfaces can additionally be exported to Petrel Surface PTL format. Use the right-click menu for the surface you want to export and choose the export format you want to use. A file selection window will show up allowing you to choose where you want to save the exported data. ## Supported Surface Formats GOCAD is a computer application that allows you to import, create, modify, integrate, view, and export geological objects in 3D. The GOCAD export file format supported by ResInsight is TSURF (*.ts). A TSURF data file is a triangle based surface format containing vertex coordinates and triangle to vertex connectivities as exemplified below. ResInsight import vertex and triangle identifiers from the first TFACE section in such a file. GOCAD TSurf 1 name:MF_027_SU } NAME Default AXIS_NAME "X" "Y" "Z" AXIS_UNIT "m" "m" "m" ZPOSITIVE Depth END_ORIGINAL_COORDINATE_SYSTEM TFACE VRTX 1 458177.767090 7322538.712891 1643.655884 CNXYZ VRTX 2 458260.834961 7322392.890625 1596.685303 CNXYZ VRTX 3 457985.662109 7322783.783203 1542.060059 CNXYZ VRTX 4 459601.453125 7322511.427734 3639.000000 CNXYZ VRTX 5 459422.015625 7322689.230469 3639.000000 CNXYZ VRTX 6 459793.410156 7322338.230469 3639.000000 CNXYZ ... TRGL 2 61 98 TRGL 20 153 66 TRGL 152 19 65 END ### Petrel Surface PTL files ResInsight is capable of importing a surface defined by a PTL (*.ptl) file from Schlumberger Petrel. A PTL data file specifies the quads of a surface by x, y, z nodal coordinates and the i and j indices as exemplified below. As seen, # denotes comment lines. #Type: scattered data #Version: 6 #Description: No description #Format: free #Field: 1 x #Field: 2 y #Field: 3 z meters #Field: 4 column #Field: 5 row #Projection: Local Rectangular #Units: meters #End: #Information from grid #Grid_size: Not_avaiable #Grid_space: Not_available #Z_field: z #Vertical_faults: Not_available #History: No history #Z_units: meters 443479.500000 7305390.500000 -1000.000000 1 1 443479.500000 7305488.500000 -1000.000000 1 2 443479.500000 7305586.500000 -1000.000000 1 3 443479.500000 7305684.500000 -1000.000000 1 4 443479.500000 7305782.500000 -1000.000000 1 5 443479.500000 7305880.500000 -1000.000000 1 6 ... ### OpenWorks XYZ Surface files ResInsight is capable of importing a surface defined by a XYZ (*.dat) file from OpenWorks. A XYZ data file specifies the quads of a surface by x, y, z nodal coordinates organized in a regular grid. As seen, # and @ denotes comment lines. @File_Version: 4 @Coordinate_Type_is: 1 @Export_Type_is: 1 @Number_of_Projects 1 @Project_Type_Name: , 3,xxx, @Project_Unit_is: meters , xxx #File_Version____________-> 4 #Project_Name____________-> xxx #Project_Type____________-> 3 #Export_XY_Unit__________-> meters #OpenWorks_Project_______-> 'xxx' #Master_Project_______-> #Coordinate_type_________-> 1 #Number_of_points_in_hzd_-> 1 #Horizon_internal_id_____-> xxx #Horizon_extremes_are____-> xxx,xxx #Horizon_onset_is_Minimum_____-> 1 #Horizon_type_is_DEPTH_STRUCTURE______-> 2 #Horizon_color_is________-> 255 0 0 #Horizon_name____________-> xxx #Horizon_attribute_______-> DEPTH_STRUCTURE #Horizon_version_________-> UNKNOWN #Horizon_interp_status___-> defaultStat #Horizon_class___________-> defaultClass #Export_Z_Unit___________-> meters #Horizon_onset_type______-> Minimum #Horizon_data_domain_____-> DEPTH #Horizon_remark_size_____-> 50 Horizon from Grid on Fri Aug 14 13:42:10 CEST 2020
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.34199535846710205, "perplexity": 7775.187449429}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030333541.98/warc/CC-MAIN-20220924213650-20220925003650-00371.warc.gz"}
https://ask.sagemath.org/questions/50330/revisions/
# Revision history [back] ### How to do arithmetic with a basis element from modular forms space Hi, I'm new to Sage and coding in general. I want to look at the 1-dim'l space CuspForms(Gamma0(23), 3) but with character. Here's my code for that S = G = DirichletGroup(7, QQ) list(G) G.gens() e = G.0 S = CuspForms(e, 3, prec=20) print(S) print(Gamma0(7).cusps()) print(S.basis()) g = S.0 Here, g is the only basis element for S. I want to do F = g*(L+4) where L is some eta-quotient of level 7. However, I get a long error with 'NoneType' object is not iterable at the end. So I'm thinking I need to define g in a different way, since it seems to be that g is not a q-series type object the way I've defined it? ### How to do arithmetic with a basis element from modular forms space Hi, I'm new to Sage and coding in general. I want to look at the 1-dim'l space CuspForms(Gamma0(23), 3) but with character. Here's my code for that S = G = DirichletGroup(7, DirichletGroup(23, QQ) list(G) G.gens() e = G.0 S = CuspForms(e, 3, prec=20) print(S) print(Gamma0(7).cusps()) print(Gamma0(23).cusps()) print(S.basis()) g = S.0 Here, g is the only basis element for S. I want to do F = g*(L+4) where L is some eta-quotient of level 7. However, I get a long error with 'NoneType' object is not iterable at the end. So I'm thinking I need to define g in a different way, since it seems to be that g is not a q-series type object the way I've defined it? it? ### How to do arithmetic with a basis element from modular forms space Hi, I'm new to Sage and coding in general. I want to look at the 1-dim'l space CuspForms(Gamma0(23), CuspForms(Gamma0(7), 3) but with character. Here's my code for that S = G = DirichletGroup(23, DirichletGroup(7, QQ) list(G) G.gens() e = G.0 S = CuspForms(e, 3, prec=20) print(S) print(Gamma0(23).cusps()) print(Gamma0(7).cusps()) print(S.basis()) g = S.0 Here, g is the only basis element for S. I want to do F = g*(L+4) where L is some eta-quotient of level 7. However, I get a long error with 'NoneType' object is not iterable at the end. So I'm thinking I need to define g in a different way, since it seems to be that g is not a q-series type object the way I've defined it?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8717278838157654, "perplexity": 853.9891160706568}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178362741.28/warc/CC-MAIN-20210301151825-20210301181825-00115.warc.gz"}
http://mathhelpforum.com/statistics/120704-probabilities-more-questions.html
# Thread: Probabilities - more questions 1. ## Probabilities - more questions 3. A bag contains 10 balls, 4 are red and 6 are yellow. Three balls are drawn without replacement. What is the probability of drawing at least two red balls? 4. A town has two fire engines operating independently. The probability that a specific fire engine is available when need is 0.95. What is the probability that neither is available when needed? What is the probability that a fire engine is available when needed? OK could someone help out a little bit my brain is failing to work, I have done some working out but none of it seems to make sense and it seems easy. Don't know what's going on up there. 2. What is the working out you have done? 3. Hello, djmccabie! 3. A bag contains 10 balls, 4 are red and 6 are yellow. Three balls are drawn without replacement. What is the probability of drawing at least two red balls? We draw 3 balls from 10 balls. There are: . ${10\choose3} = 120$ possible outcomes. We want 3 Red balls, or 2 Red and 1 Yellow. . . 3R: . ${4\choose3} = 4$ ways. . . 2R, 1Y: . ${4\choose2}{6\choose1} = 6\cdot6 = 36$ ways. Hence, there are: . $4 + 36 \:=\:40$ ways to get at least 2 red balls. Therefore: . $P(\text{at least 2 R}) \;=\;\frac{40}{120} \;=\;\frac{1}{3}$ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ You could have solved it taking baby steps . . . There are four ways to get at least 2 red balls: .RRR, RRY, RYR, YRR. . . $\begin{array}{ccccc}P(RRR) &=& \frac{4}{10}\cdot\frac{3}{9}\cdot\frac{2}{8} &=& \frac{1}{30} \\ \\[-3mm] P(RRY) &=&\frac{4}{10}\cdot\frac{3}{9}\cdot\frac{6}{8} &=& \frac{1}{10} \\ \\[-3mm] P(RY\!R) &=& \frac{4}{10}\cdot\frac{6}{9}\cdot\frac{3}{8} &=& \frac{1}{10} \\ \\[-3mm] P(Y\!RR) &=& \frac{6}{10}\cdot\frac{4}{9}\cdot\frac{3}{8} &=& \frac{1}{10}\end{array}$ Therefore: . $P(\text{at least 2 R}) \;=\;\frac{1}{30} + \frac{1}{10} + \frac{1}{10} + \frac{1}{10} \;=\;\frac{10}{30} \;=\;\frac{1}{3}$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 7, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8355732560157776, "perplexity": 278.1819510410066}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-36/segments/1471982292181.27/warc/CC-MAIN-20160823195812-00102-ip-10-153-172-175.ec2.internal.warc.gz"}
https://download.atlantis-press.com/proceedings/iccia-12/4270
# The Application of Electronic Map in Wireless Positioning System Authors Qinghui Wang, Wenzhou Wang, Dan Li, Lei Chang Corresponding Author Qinghui Wang Available Online May 2014. DOI https://doi.org/10.2991/iccia.2012.322How to use a DOI? Keywords MapXtreme, WebGIS, Map Matching, Electronic Map Abstract By using powerful secondary development tool MapXtreme and combining .NET development platform, Web Geographic Information System (WebGIS) used for the monitoring of the indoor personnel is designed. The system could display real-time location of the indoor personnel on the electronic map through getting the coordinates and other data measured by wireless positioning system combined with map matching. This system follows the browser/server mode (B/S mode), thus the visitors can easily access simply by the browser without the need of installing other application software. Copyright © 2013, the Authors. Published by Atlantis Press. Open Access This is an open access article distributed under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). Volume Title Proceedings of the 2012 2nd International Conference on Computer and Information Application (ICCIA 2012) Series Advances in Intelligent Systems Research Publication Date May 2014 ISBN 978-94-91216-41-1 ISSN 1951-6851 DOI https://doi.org/10.2991/iccia.2012.322How to use a DOI? Copyright © 2013, the Authors. Published by Atlantis Press. Open Access This is an open access article distributed under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). ### Cite this article TY - CONF AU - Qinghui Wang AU - Wenzhou Wang AU - Dan Li AU - Lei Chang PY - 2014/05 DA - 2014/05 TI - The Application of Electronic Map in Wireless Positioning System BT - Proceedings of the 2012 2nd International Conference on Computer and Information Application (ICCIA 2012) PB - Atlantis Press SP - 1298 EP - 1301 SN - 1951-6851 UR - https://doi.org/10.2991/iccia.2012.322 DO - https://doi.org/10.2991/iccia.2012.322 ID - Wang2014/05 ER -
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4252360463142395, "perplexity": 16999.106783090858}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446708046.99/warc/CC-MAIN-20221126180719-20221126210719-00481.warc.gz"}
https://worldwidescience.org/topicpages/c/complication+probability+models.html
#### Sample records for complication probability models 1. Statistical validation of normal tissue complication probability models NARCIS (Netherlands) Xu, Cheng-Jian; van der Schaaf, Arjen; van t Veld, Aart; Langendijk, Johannes A.; Schilstra, Cornelis 2012-01-01 PURPOSE: To investigate the applicability and value of double cross-validation and permutation tests as established statistical approaches in the validation of normal tissue complication probability (NTCP) models. METHODS AND MATERIALS: A penalized regression method, LASSO (least absolute shrinkage 2. Statistical Validation of Normal Tissue Complication Probability Models Energy Technology Data Exchange (ETDEWEB) Xu Chengjian, E-mail: c.j.xu@umcg.nl [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Schaaf, Arjen van der; Veld, Aart A. van' t; Langendijk, Johannes A. [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Schilstra, Cornelis [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Radiotherapy Institute Friesland, Leeuwarden (Netherlands) 2012-09-01 Purpose: To investigate the applicability and value of double cross-validation and permutation tests as established statistical approaches in the validation of normal tissue complication probability (NTCP) models. Methods and Materials: A penalized regression method, LASSO (least absolute shrinkage and selection operator), was used to build NTCP models for xerostomia after radiation therapy treatment of head-and-neck cancer. Model assessment was based on the likelihood function and the area under the receiver operating characteristic curve. Results: Repeated double cross-validation showed the uncertainty and instability of the NTCP models and indicated that the statistical significance of model performance can be obtained by permutation testing. Conclusion: Repeated double cross-validation and permutation tests are recommended to validate NTCP models before clinical use. 3. Statistical validation of normal tissue complication probability models. Science.gov (United States) Xu, Cheng-Jian; van der Schaaf, Arjen; Van't Veld, Aart A; Langendijk, Johannes A; Schilstra, Cornelis 2012-09-01 To investigate the applicability and value of double cross-validation and permutation tests as established statistical approaches in the validation of normal tissue complication probability (NTCP) models. A penalized regression method, LASSO (least absolute shrinkage and selection operator), was used to build NTCP models for xerostomia after radiation therapy treatment of head-and-neck cancer. Model assessment was based on the likelihood function and the area under the receiver operating characteristic curve. Repeated double cross-validation showed the uncertainty and instability of the NTCP models and indicated that the statistical significance of model performance can be obtained by permutation testing. Repeated double cross-validation and permutation tests are recommended to validate NTCP models before clinical use. Copyright © 2012 Elsevier Inc. All rights reserved. 4. Options and pitfalls of normal tissues complication probability models International Nuclear Information System (INIS) Dorr, Wolfgang 2011-01-01 5. Impact of statistical learning methods on the predictive power of multivariate normal tissue complication probability models NARCIS (Netherlands) Xu, Cheng-Jian; van der Schaaf, Arjen; Schilstra, Cornelis; Langendijk, Johannes A.; van t Veld, Aart A. 2012-01-01 PURPOSE: To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. METHODS AND MATERIALS: In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator 6. Impact of Statistical Learning Methods on the Predictive Power of Multivariate Normal Tissue Complication Probability Models Energy Technology Data Exchange (ETDEWEB) Xu Chengjian, E-mail: c.j.xu@umcg.nl [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Schaaf, Arjen van der; Schilstra, Cornelis; Langendijk, Johannes A.; Veld, Aart A. van' t [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands) 2012-03-15 Purpose: To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. Methods and Materials: In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator (LASSO), and Bayesian model averaging (BMA), were used to build NTCP models of xerostomia following radiotherapy treatment for head and neck cancer. Performance of each learning method was evaluated by a repeated cross-validation scheme in order to obtain a fair comparison among methods. Results: It was found that the LASSO and BMA methods produced models with significantly better predictive power than that of the stepwise selection method. Furthermore, the LASSO method yields an easily interpretable model as the stepwise method does, in contrast to the less intuitive BMA method. Conclusions: The commonly used stepwise selection method, which is simple to execute, may be insufficient for NTCP modeling. The LASSO method is recommended. 7. Impact of statistical learning methods on the predictive power of multivariate normal tissue complication probability models. Science.gov (United States) Xu, Cheng-Jian; van der Schaaf, Arjen; Schilstra, Cornelis; Langendijk, Johannes A; van't Veld, Aart A 2012-03-15 To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator (LASSO), and Bayesian model averaging (BMA), were used to build NTCP models of xerostomia following radiotherapy treatment for head and neck cancer. Performance of each learning method was evaluated by a repeated cross-validation scheme in order to obtain a fair comparison among methods. It was found that the LASSO and BMA methods produced models with significantly better predictive power than that of the stepwise selection method. Furthermore, the LASSO method yields an easily interpretable model as the stepwise method does, in contrast to the less intuitive BMA method. The commonly used stepwise selection method, which is simple to execute, may be insufficient for NTCP modeling. The LASSO method is recommended. Copyright © 2012 Elsevier Inc. All rights reserved. 8. Impact of Statistical Learning Methods on the Predictive Power of Multivariate Normal Tissue Complication Probability Models International Nuclear Information System (INIS) Xu Chengjian; Schaaf, Arjen van der; Schilstra, Cornelis; Langendijk, Johannes A.; Veld, Aart A. van’t 2012-01-01 Purpose: To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. Methods and Materials: In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator (LASSO), and Bayesian model averaging (BMA), were used to build NTCP models of xerostomia following radiotherapy treatment for head and neck cancer. Performance of each learning method was evaluated by a repeated cross-validation scheme in order to obtain a fair comparison among methods. Results: It was found that the LASSO and BMA methods produced models with significantly better predictive power than that of the stepwise selection method. Furthermore, the LASSO method yields an easily interpretable model as the stepwise method does, in contrast to the less intuitive BMA method. Conclusions: The commonly used stepwise selection method, which is simple to execute, may be insufficient for NTCP modeling. The LASSO method is recommended. 9. Normal Tissue Complication Probability Modeling of Acute Hematologic Toxicity in Cervical Cancer Patients Treated With Chemoradiotherapy International Nuclear Information System (INIS) Rose, Brent S.; Aydogan, Bulent; Liang, Yun; Yeginer, Mete; Hasselle, Michael D.; Dandekar, Virag; Bafana, Rounak; Yashar, Catheryn M.; Mundt, Arno J.; Roeske, John C.; Mell, Loren K. 2011-01-01 10. Improving normal tissue complication probability models: the need to adopt a "data-pooling" culture. Science.gov (United States) Deasy, Joseph O; Bentzen, Søren M; Jackson, Andrew; Ten Haken, Randall K; Yorke, Ellen D; Constine, Louis S; Sharma, Ashish; Marks, Lawrence B 2010-03-01 Clinical studies of the dependence of normal tissue response on dose-volume factors are often confusingly inconsistent, as the QUANTEC reviews demonstrate. A key opportunity to accelerate progress is to begin storing high-quality datasets in repositories. Using available technology, multiple repositories could be conveniently queried, without divulging protected health information, to identify relevant sources of data for further analysis. After obtaining institutional approvals, data could then be pooled, greatly enhancing the capability to construct predictive models that are more widely applicable and better powered to accurately identify key predictive factors (whether dosimetric, image-based, clinical, socioeconomic, or biological). Data pooling has already been carried out effectively in a few normal tissue complication probability studies and should become a common strategy. Copyright 2010 Elsevier Inc. All rights reserved. 11. Hypothyroidism after primary radiotherapy for head and neck squamous cell carcinoma: Normal tissue complication probability modeling with latent time correction DEFF Research Database (Denmark) Rønjom, Marianne Feen; Brink, Carsten; Bentzen, Søren 2013-01-01 To develop a normal tissue complication probability (NTCP) model of radiation-induced biochemical hypothyroidism (HT) after primary radiotherapy for head and neck squamous cell carcinoma (HNSCC) with adjustment for latency and clinical risk factors.......To develop a normal tissue complication probability (NTCP) model of radiation-induced biochemical hypothyroidism (HT) after primary radiotherapy for head and neck squamous cell carcinoma (HNSCC) with adjustment for latency and clinical risk factors.... 12. Method for Automatic Selection of Parameters in Normal Tissue Complication Probability Modeling. Science.gov (United States) Christophides, Damianos; Appelt, Ane L; Gusnanto, Arief; Lilley, John; Sebag-Montefiore, David 2018-07-01 To present a fully automatic method to generate multiparameter normal tissue complication probability (NTCP) models and compare its results with those of a published model, using the same patient cohort. Data were analyzed from 345 rectal cancer patients treated with external radiation therapy to predict the risk of patients developing grade 1 or ≥2 cystitis. In total, 23 clinical factors were included in the analysis as candidate predictors of cystitis. Principal component analysis was used to decompose the bladder dose-volume histogram into 8 principal components, explaining more than 95% of the variance. The data set of clinical factors and principal components was divided into training (70%) and test (30%) data sets, with the training data set used by the algorithm to compute an NTCP model. The first step of the algorithm was to obtain a bootstrap sample, followed by multicollinearity reduction using the variance inflation factor and genetic algorithm optimization to determine an ordinal logistic regression model that minimizes the Bayesian information criterion. The process was repeated 100 times, and the model with the minimum Bayesian information criterion was recorded on each iteration. The most frequent model was selected as the final "automatically generated model" (AGM). The published model and AGM were fitted on the training data sets, and the risk of cystitis was calculated. The 2 models had no significant differences in predictive performance, both for the training and test data sets (P value > .05) and found similar clinical and dosimetric factors as predictors. Both models exhibited good explanatory performance on the training data set (P values > .44), which was reduced on the test data sets (P values < .05). The predictive value of the AGM is equivalent to that of the expert-derived published model. It demonstrates potential in saving time, tackling problems with a large number of parameters, and standardizing variable selection in NTCP 13. Multivariate Normal Tissue Complication Probability Modeling of Heart Valve Dysfunction in Hodgkin Lymphoma Survivors International Nuclear Information System (INIS) Cella, Laura; Liuzzi, Raffaele; Conson, Manuel; D’Avino, Vittoria; Salvatore, Marco; Pacelli, Roberto 2013-01-01 Purpose: To establish a multivariate normal tissue complication probability (NTCP) model for radiation-induced asymptomatic heart valvular defects (RVD). Methods and Materials: Fifty-six patients treated with sequential chemoradiation therapy for Hodgkin lymphoma (HL) were retrospectively reviewed for RVD events. Clinical information along with whole heart, cardiac chambers, and lung dose distribution parameters was collected, and the correlations to RVD were analyzed by means of Spearman's rank correlation coefficient (Rs). For the selection of the model order and parameters for NTCP modeling, a multivariate logistic regression method using resampling techniques (bootstrapping) was applied. Model performance was evaluated using the area under the receiver operating characteristic curve (AUC). Results: When we analyzed the whole heart, a 3-variable NTCP model including the maximum dose, whole heart volume, and lung volume was shown to be the optimal predictive model for RVD (Rs = 0.573, P<.001, AUC = 0.83). When we analyzed the cardiac chambers individually, for the left atrium and for the left ventricle, an NTCP model based on 3 variables including the percentage volume exceeding 30 Gy (V30), cardiac chamber volume, and lung volume was selected as the most predictive model (Rs = 0.539, P<.001, AUC = 0.83; and Rs = 0.557, P<.001, AUC = 0.82, respectively). The NTCP values increase as heart maximum dose or cardiac chambers V30 increase. They also increase with larger volumes of the heart or cardiac chambers and decrease when lung volume is larger. Conclusions: We propose logistic NTCP models for RVD considering not only heart irradiation dose but also the combined effects of lung and heart volumes. Our study establishes the statistical evidence of the indirect effect of lung size on radio-induced heart toxicity 14. Normal tissue complication probabilities: dependence on choice of biological model and dose-volume histogram reduction scheme International Nuclear Information System (INIS) Moiseenko, Vitali; Battista, Jerry; Van Dyk, Jake 2000-01-01 Purpose: To evaluate the impact of dose-volume histogram (DVH) reduction schemes and models of normal tissue complication probability (NTCP) on ranking of radiation treatment plans. Methods and Materials: Data for liver complications in humans and for spinal cord in rats were used to derive input parameters of four different NTCP models. DVH reduction was performed using two schemes: 'effective volume' and 'preferred Lyman'. DVHs for competing treatment plans were derived from a sample DVH by varying dose uniformity in a high dose region so that the obtained cumulative DVHs intersected. Treatment plans were ranked according to the calculated NTCP values. Results: Whenever the preferred Lyman scheme was used to reduce the DVH, competing plans were indistinguishable as long as the mean dose was constant. The effective volume DVH reduction scheme did allow us to distinguish between these competing treatment plans. However, plan ranking depended on the radiobiological model used and its input parameters. Conclusions: Dose escalation will be a significant part of radiation treatment planning using new technologies, such as 3-D conformal radiotherapy and tomotherapy. Such dose escalation will depend on how the dose distributions in organs at risk are interpreted in terms of expected complication probabilities. The present study indicates considerable variability in predicted NTCP values because of the methods used for DVH reduction and radiobiological models and their input parameters. Animal studies and collection of standardized clinical data are needed to ascertain the effects of non-uniform dose distributions and to test the validity of the models currently in use Science.gov (United States) Bakhshandeh, Mohsen; Hashemi, Bijan; Mahdavi, Seied Rabi Mehdi; Nikoofar, Alireza; Vasheghani, Maryam; Kazemnejad, Anoshirvan 2013-02-01 To determine the dose-response relationship of the thyroid for radiation-induced hypothyroidism in head-and-neck radiation therapy, according to 6 normal tissue complication probability models, and to find the best-fit parameters of the models. Sixty-five patients treated with primary or postoperative radiation therapy for various cancers in the head-and-neck region were prospectively evaluated. Patient serum samples (tri-iodothyronine, thyroxine, thyroid-stimulating hormone [TSH], free tri-iodothyronine, and free thyroxine) were measured before and at regular time intervals until 1 year after the completion of radiation therapy. Dose-volume histograms (DVHs) of the patients' thyroid gland were derived from their computed tomography (CT)-based treatment planning data. Hypothyroidism was defined as increased TSH (subclinical hypothyroidism) or increased TSH in combination with decreased free thyroxine and thyroxine (clinical hypothyroidism). Thyroid DVHs were converted to 2 Gy/fraction equivalent doses using the linear-quadratic formula with α/β = 3 Gy. The evaluated models included the following: Lyman with the DVH reduced to the equivalent uniform dose (EUD), known as LEUD; Logit-EUD; mean dose; relative seriality; individual critical volume; and population critical volume models. The parameters of the models were obtained by fitting the patients' data using a maximum likelihood analysis method. The goodness of fit of the models was determined by the 2-sample Kolmogorov-Smirnov test. Ranking of the models was made according to Akaike's information criterion. Twenty-nine patients (44.6%) experienced hypothyroidism. None of the models was rejected according to the evaluation of the goodness of fit. The mean dose model was ranked as the best model on the basis of its Akaike's information criterion value. The D(50) estimated from the models was approximately 44 Gy. The implemented normal tissue complication probability models showed a parallel architecture for the Energy Technology Data Exchange (ETDEWEB) Bakhshandeh, Mohsen [Department of Medical Physics, Faculty of Medical Sciences, Tarbiat Modares University, Tehran (Iran, Islamic Republic of); Hashemi, Bijan, E-mail: bhashemi@modares.ac.ir [Department of Medical Physics, Faculty of Medical Sciences, Tarbiat Modares University, Tehran (Iran, Islamic Republic of); Mahdavi, Seied Rabi Mehdi [Department of Medical Physics, Faculty of Medical Sciences, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Nikoofar, Alireza; Vasheghani, Maryam [Department of Radiation Oncology, Hafte-Tir Hospital, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Kazemnejad, Anoshirvan [Department of Biostatistics, Faculty of Medical Sciences, Tarbiat Modares University, Tehran (Iran, Islamic Republic of) 2013-02-01 Purpose: To determine the dose-response relationship of the thyroid for radiation-induced hypothyroidism in head-and-neck radiation therapy, according to 6 normal tissue complication probability models, and to find the best-fit parameters of the models. Methods and Materials: Sixty-five patients treated with primary or postoperative radiation therapy for various cancers in the head-and-neck region were prospectively evaluated. Patient serum samples (tri-iodothyronine, thyroxine, thyroid-stimulating hormone [TSH], free tri-iodothyronine, and free thyroxine) were measured before and at regular time intervals until 1 year after the completion of radiation therapy. Dose-volume histograms (DVHs) of the patients' thyroid gland were derived from their computed tomography (CT)-based treatment planning data. Hypothyroidism was defined as increased TSH (subclinical hypothyroidism) or increased TSH in combination with decreased free thyroxine and thyroxine (clinical hypothyroidism). Thyroid DVHs were converted to 2 Gy/fraction equivalent doses using the linear-quadratic formula with {alpha}/{beta} = 3 Gy. The evaluated models included the following: Lyman with the DVH reduced to the equivalent uniform dose (EUD), known as LEUD; Logit-EUD; mean dose; relative seriality; individual critical volume; and population critical volume models. The parameters of the models were obtained by fitting the patients' data using a maximum likelihood analysis method. The goodness of fit of the models was determined by the 2-sample Kolmogorov-Smirnov test. Ranking of the models was made according to Akaike's information criterion. Results: Twenty-nine patients (44.6%) experienced hypothyroidism. None of the models was rejected according to the evaluation of the goodness of fit. The mean dose model was ranked as the best model on the basis of its Akaike's information criterion value. The D{sub 50} estimated from the models was approximately 44 Gy. Conclusions: The implemented 17. Method for Automatic Selection of Parameters in Normal Tissue Complication Probability Modeling DEFF Research Database (Denmark) Christophides, Damianos; Appelt, Ane L; Gusnanto, Arief 2018-01-01 by multicollinearity reduction using the variance inflation factor and genetic algorithm optimization to determine an ordinal logistic regression model that minimizes the Bayesian information criterion. The process was repeated 100 times, and the model with the minimum Bayesian information criterion was recorded... 18. Multivariate normal tissue complication probability modeling of gastrointestinal toxicity after external beam radiotherapy for localized prostate cancer International Nuclear Information System (INIS) Cella, Laura; D’Avino, Vittoria; Liuzzi, Raffaele; Conson, Manuel; Doria, Francesca; Faiella, Adriana; Loffredo, Filomena; Salvatore, Marco; Pacelli, Roberto 2013-01-01 The risk of radio-induced gastrointestinal (GI) complications is affected by several factors other than the dose to the rectum such as patient characteristics, hormonal or antihypertensive therapy, and acute rectal toxicity. Purpose of this work is to study clinical and dosimetric parameters impacting on late GI toxicity after prostate external beam radiotherapy (RT) and to establish multivariate normal tissue complication probability (NTCP) model for radiation-induced GI complications. A total of 57 men who had undergone definitive RT for prostate cancer were evaluated for GI events classified using the RTOG/EORTC scoring system. Their median age was 73 years (range 53–85). The patients were assessed for GI toxicity before, during, and periodically after RT completion. Several clinical variables along with rectum dose-volume parameters (Vx) were collected and their correlation to GI toxicity was analyzed by Spearman’s rank correlation coefficient (Rs). Multivariate logistic regression method using resampling techniques was applied to select model order and parameters for NTCP modeling. Model performance was evaluated through the area under the receiver operating characteristic curve (AUC). At a median follow-up of 30 months, 37% (21/57) patients developed G1-2 acute GI events while 33% (19/57) were diagnosed with G1-2 late GI events. An NTCP model for late mild/moderate GI toxicity based on three variables including V65 (OR = 1.03), antihypertensive and/or anticoagulant (AH/AC) drugs (OR = 0.24), and acute GI toxicity (OR = 4.3) was selected as the most predictive model (Rs = 0.47, p < 0.001; AUC = 0.79). This three-variable model outperforms the logistic model based on V65 only (Rs = 0.28, p < 0.001; AUC = 0.69). We propose a logistic NTCP model for late GI toxicity considering not only rectal irradiation dose but also clinical patient-specific factors. Accordingly, the risk of G1-2 late GI increases as V65 increases, it is higher for patients experiencing 19. Hypothyroidism after primary radiotherapy for head and neck squamous cell carcinoma: Normal tissue complication probability modeling with latent time correction International Nuclear Information System (INIS) Rønjom, Marianne Feen; Brink, Carsten; Bentzen, Søren M.; Hegedüs, Laszlo; Overgaard, Jens; Johansen, Jørgen 2013-01-01 Background and purpose: To develop a normal tissue complication probability (NTCP) model of radiation-induced biochemical hypothyroidism (HT) after primary radiotherapy for head and neck squamous cell carcinoma (HNSCC) with adjustment for latency and clinical risk factors. Patients and methods: Patients with HNSCC receiving definitive radiotherapy with 66–68 Gy without surgery were followed up with serial post-treatment thyrotropin (TSH) assessment. HT was defined as TSH >4.0 mU/l. Data were analyzed with both a logistic and a mixture model (correcting for latency) to determine risk factors for HT and develop an NTCP model based on mean thyroid dose (MTD) and thyroid volume. Results: 203 patients were included. Median follow-up: 25.1 months. Five-year estimated risk of HT was 25.6%. In the mixture model, the only independent risk factors for HT were thyroid volume (cm 3 ) (OR = 0.75 [95% CI: 0.64–0.85], p 3 , respectively. Conclusions: Comparing the logistic and mixture models demonstrates the importance of latent-time correction in NTCP-modeling. Thyroid dose constraints in treatment planning should be individualized based on thyroid volume 20. Calculation of the uncertainty in complication probability for various dose-response models, applied to the parotid gland International Nuclear Information System (INIS) Schilstra, C.; Meertens, H. 2001-01-01 Purpose: Usually, models that predict normal tissue complication probability (NTCP) are fitted to clinical data with the maximum likelihood (ML) method. This method inevitably causes a loss of information contained in the data. In this study, an alternative method is investigated that calculates the parameter probability distribution (PD), and, thus, conserves all information. The PD method also allows the calculation of the uncertainty in the NTCP, which is an (often-neglected) prerequisite for the intercomparison of both treatment plans and NTCP models. The PD and ML methods are applied to parotid gland data, and the results are compared. Methods and Materials: The drop in salivary flow due to radiotherapy was measured in 25 parotid glands of 15 patients. Together with the parotid gland dose-volume histograms (DVH), this enabled the calculation of the parameter PDs for three different NTCP models (Lyman, relative seriality, and critical volume). From these PDs, the NTCP and its uncertainty could be calculated for arbitrary parotid gland DVHs. ML parameters and resulting NTCP values were calculated also. Results: All models fitted equally well. The parameter PDs turned out to have nonnormal shapes and long tails. The NTCP predictions of the ML and PD method usually differed considerably, depending on the NTCP model and the nature of irradiation. NTCP curves and ML parameters suggested a highly parallel organization of the parotid gland. Conclusions: Considering the substantial differences between the NTCP predictions of the ML and PD method, the use of the PD method is preferred, because this is the only method that takes all information contained in the clinical data into account. Furthermore, PD method gives a true measure of the uncertainty in the NTCP 1. The Benefits of Including Clinical Factors in Rectal Normal Tissue Complication Probability Modeling After Radiotherapy for Prostate Cancer International Nuclear Information System (INIS) Defraene, Gilles; Van den Bergh, Laura; Al-Mamgani, Abrahim; Haustermans, Karin; Heemsbergen, Wilma; Van den Heuvel, Frank; Lebesque, Joos V. 2012-01-01 Purpose: To study the impact of clinical predisposing factors on rectal normal tissue complication probability modeling using the updated results of the Dutch prostate dose-escalation trial. Methods and Materials: Toxicity data of 512 patients (conformally treated to 68 Gy [n = 284] and 78 Gy [n = 228]) with complete follow-up at 3 years after radiotherapy were studied. Scored end points were rectal bleeding, high stool frequency, and fecal incontinence. Two traditional dose-based models (Lyman-Kutcher-Burman (LKB) and Relative Seriality (RS) and a logistic model were fitted using a maximum likelihood approach. Furthermore, these model fits were improved by including the most significant clinical factors. The area under the receiver operating characteristic curve (AUC) was used to compare the discriminating ability of all fits. Results: Including clinical factors significantly increased the predictive power of the models for all end points. In the optimal LKB, RS, and logistic models for rectal bleeding and fecal incontinence, the first significant (p = 0.011–0.013) clinical factor was “previous abdominal surgery.” As second significant (p = 0.012–0.016) factor, “cardiac history” was included in all three rectal bleeding fits, whereas including “diabetes” was significant (p = 0.039–0.048) in fecal incontinence modeling but only in the LKB and logistic models. High stool frequency fits only benefitted significantly (p = 0.003–0.006) from the inclusion of the baseline toxicity score. For all models rectal bleeding fits had the highest AUC (0.77) where it was 0.63 and 0.68 for high stool frequency and fecal incontinence, respectively. LKB and logistic model fits resulted in similar values for the volume parameter. The steepness parameter was somewhat higher in the logistic model, also resulting in a slightly lower D 50 . Anal wall DVHs were used for fecal incontinence, whereas anorectal wall dose best described the other two endpoints. Conclusions 2. The Benefits of Including Clinical Factors in Rectal Normal Tissue Complication Probability Modeling After Radiotherapy for Prostate Cancer Energy Technology Data Exchange (ETDEWEB) Defraene, Gilles, E-mail: gilles.defraene@uzleuven.be [Radiation Oncology Department, University Hospitals Leuven, Leuven (Belgium); Van den Bergh, Laura [Radiation Oncology Department, University Hospitals Leuven, Leuven (Belgium); Al-Mamgani, Abrahim [Department of Radiation Oncology, Erasmus Medical Center - Daniel den Hoed Cancer Center, Rotterdam (Netherlands); Haustermans, Karin [Radiation Oncology Department, University Hospitals Leuven, Leuven (Belgium); Heemsbergen, Wilma [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital, Amsterdam (Netherlands); Van den Heuvel, Frank [Radiation Oncology Department, University Hospitals Leuven, Leuven (Belgium); Lebesque, Joos V. [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital, Amsterdam (Netherlands) 2012-03-01 Purpose: To study the impact of clinical predisposing factors on rectal normal tissue complication probability modeling using the updated results of the Dutch prostate dose-escalation trial. Methods and Materials: Toxicity data of 512 patients (conformally treated to 68 Gy [n = 284] and 78 Gy [n = 228]) with complete follow-up at 3 years after radiotherapy were studied. Scored end points were rectal bleeding, high stool frequency, and fecal incontinence. Two traditional dose-based models (Lyman-Kutcher-Burman (LKB) and Relative Seriality (RS) and a logistic model were fitted using a maximum likelihood approach. Furthermore, these model fits were improved by including the most significant clinical factors. The area under the receiver operating characteristic curve (AUC) was used to compare the discriminating ability of all fits. Results: Including clinical factors significantly increased the predictive power of the models for all end points. In the optimal LKB, RS, and logistic models for rectal bleeding and fecal incontinence, the first significant (p = 0.011-0.013) clinical factor was 'previous abdominal surgery.' As second significant (p = 0.012-0.016) factor, 'cardiac history' was included in all three rectal bleeding fits, whereas including 'diabetes' was significant (p = 0.039-0.048) in fecal incontinence modeling but only in the LKB and logistic models. High stool frequency fits only benefitted significantly (p = 0.003-0.006) from the inclusion of the baseline toxicity score. For all models rectal bleeding fits had the highest AUC (0.77) where it was 0.63 and 0.68 for high stool frequency and fecal incontinence, respectively. LKB and logistic model fits resulted in similar values for the volume parameter. The steepness parameter was somewhat higher in the logistic model, also resulting in a slightly lower D{sub 50}. Anal wall DVHs were used for fecal incontinence, whereas anorectal wall dose best described the other two endpoints 3. Design and Selection of Machine Learning Methods Using Radiomics and Dosiomics for Normal Tissue Complication Probability Modeling of Xerostomia. Science.gov (United States) Gabryś, Hubert S; Buettner, Florian; Sterzing, Florian; Hauswald, Henrik; Bangert, Mark 2018-01-01 The purpose of this study is to investigate whether machine learning with dosiomic, radiomic, and demographic features allows for xerostomia risk assessment more precise than normal tissue complication probability (NTCP) models based on the mean radiation dose to parotid glands. A cohort of 153 head-and-neck cancer patients was used to model xerostomia at 0-6 months (early), 6-15 months (late), 15-24 months (long-term), and at any time (a longitudinal model) after radiotherapy. Predictive power of the features was evaluated by the area under the receiver operating characteristic curve (AUC) of univariate logistic regression models. The multivariate NTCP models were tuned and tested with single and nested cross-validation, respectively. We compared predictive performance of seven classification algorithms, six feature selection methods, and ten data cleaning/class balancing techniques using the Friedman test and the Nemenyi post hoc analysis. NTCP models based on the parotid mean dose failed to predict xerostomia (AUCs  0.85), dose gradients in the right-left (AUCs > 0.78), and the anterior-posterior (AUCs > 0.72) direction. Multivariate models of long-term xerostomia were typically based on the parotid volume, the parotid eccentricity, and the dose-volume histogram (DVH) spread with the generalization AUCs ranging from 0.74 to 0.88. On average, support vector machines and extra-trees were the top performing classifiers, whereas the algorithms based on logistic regression were the best choice for feature selection. We found no advantage in using data cleaning or class balancing methods. We demonstrated that incorporation of organ- and dose-shape descriptors is beneficial for xerostomia prediction in highly conformal radiotherapy treatments. Due to strong reliance on patient-specific, dose-independent factors, our results underscore the need for development of personalized data-driven risk profiles for NTCP models of xerostomia. The facilitated 4. Design and Selection of Machine Learning Methods Using Radiomics and Dosiomics for Normal Tissue Complication Probability Modeling of Xerostomia Directory of Open Access Journals (Sweden) Hubert S. Gabryś 2018-03-01 Full Text Available PurposeThe purpose of this study is to investigate whether machine learning with dosiomic, radiomic, and demographic features allows for xerostomia risk assessment more precise than normal tissue complication probability (NTCP models based on the mean radiation dose to parotid glands.Material and methodsA cohort of 153 head-and-neck cancer patients was used to model xerostomia at 0–6 months (early, 6–15 months (late, 15–24 months (long-term, and at any time (a longitudinal model after radiotherapy. Predictive power of the features was evaluated by the area under the receiver operating characteristic curve (AUC of univariate logistic regression models. The multivariate NTCP models were tuned and tested with single and nested cross-validation, respectively. We compared predictive performance of seven classification algorithms, six feature selection methods, and ten data cleaning/class balancing techniques using the Friedman test and the Nemenyi post hoc analysis.ResultsNTCP models based on the parotid mean dose failed to predict xerostomia (AUCs < 0.60. The most informative predictors were found for late and long-term xerostomia. Late xerostomia correlated with the contralateral dose gradient in the anterior–posterior (AUC = 0.72 and the right–left (AUC = 0.68 direction, whereas long-term xerostomia was associated with parotid volumes (AUCs > 0.85, dose gradients in the right–left (AUCs > 0.78, and the anterior–posterior (AUCs > 0.72 direction. Multivariate models of long-term xerostomia were typically based on the parotid volume, the parotid eccentricity, and the dose–volume histogram (DVH spread with the generalization AUCs ranging from 0.74 to 0.88. On average, support vector machines and extra-trees were the top performing classifiers, whereas the algorithms based on logistic regression were the best choice for feature selection. We found no advantage in using data cleaning or class balancing 5. Normal tissue complication probability for salivary glands International Nuclear Information System (INIS) Rana, B.S. 2008-01-01 The purpose of radiotherapy is to make a profitable balance between the morbidity (due to side effects of radiation) and cure of malignancy. To achieve this, one needs to know the relation between NTCP (normal tissue complication probability) and various treatment variables of a schedule viz. daily dose, duration of treatment, total dose and fractionation along with tissue conditions. Prospective studies require that a large number of patients be treated with varied schedule parameters and a statistically acceptable number of patients develop complications so that a true relation between NTCP and a particular variable is established. In this study Salivary Glands Complications have been considered. The cases treated in 60 Co teletherapy machine during the period 1994 to 2002 were analyzed and the clinicians judgement in ascertaining the end points was the only means of observations. The only end points were early and late xerestomia which were considered for NTCP evaluations for a period of 5 years 6. Prediction of radiation-induced liver disease by Lyman normal-tissue complication probability model in three-dimensional conformal radiation therapy for primary liver carcinoma International Nuclear Information System (INIS) Xu ZhiYong; Liang Shixiong; Zhu Ji; Zhu Xiaodong; Zhao Jiandong; Lu Haijie; Yang Yunli; Chen Long; Wang Anyu; Fu Xiaolong; Jiang Guoliang 2006-01-01 Purpose: To describe the probability of RILD by application of the Lyman-Kutcher-Burman normal-tissue complication (NTCP) model for primary liver carcinoma (PLC) treated with hypofractionated three-dimensional conformal radiotherapy (3D-CRT). Methods and Materials: A total of 109 PLC patients treated by 3D-CRT were followed for RILD. Of these patients, 93 were in liver cirrhosis of Child-Pugh Grade A, and 16 were in Child-Pugh Grade B. The Michigan NTCP model was used to predict the probability of RILD, and then the modified Lyman NTCP model was generated for Child-Pugh A and Child-Pugh B patients by maximum-likelihood analysis. Results: Of all patients, 17 developed RILD in which 8 were of Child-Pugh Grade A, and 9 were of Child-Pugh Grade B. The prediction of RILD by the Michigan model was underestimated for PLC patients. The modified n, m, TD 5 (1) were 1.1, 0.28, and 40.5 Gy and 0.7, 0.43, and 23 Gy for patients with Child-Pugh A and B, respectively, which yielded better estimations of RILD probability. The hepatic tolerable doses (TD 5 ) would be MDTNL of 21 Gy and 6 Gy, respectively, for Child-Pugh A and B patients. Conclusions: The Michigan model was probably not fit to predict RILD in PLC patients. A modified Lyman NTCP model for RILD was recommended 7. External validation of a normal tissue complication probability model for radiation-induced hypothyroidism in an independent cohort DEFF Research Database (Denmark) Rønjom, Marianne F; Brink, Carsten; Bentzen, Søren M 2015-01-01 blood tests in the validation cohort relative to the original cohort. However, Pearson's correlation coefficients between model and clinical outcome were high: r = 0.97 estimated by the original model versus the original cohort, and r = 0.97 estimated by the original model versus the new cohort....... CONCLUSION: Dmean and Vthyroid were significant predictors of RIHT in both cohorts. The original NTCP model demonstrated external validity owing to high Pearson's correlation coefficients between estimated and observed incidence rates of RIHT in the original as well as in the validation cohort. This model... 8. Normal tissue complication probability model parameter estimation for xerostomia in head and neck cancer patients based on scintigraphy and quality of life assessments International Nuclear Information System (INIS) Lee, Tsair-Fwu; Chao, Pei-Ju; Wang, Hung-Yu; Hsu, Hsuan-Chih; Chang, PaoShu; Chen, Wen-Cheng 2012-01-01 With advances in modern radiotherapy (RT), many patients with head and neck (HN) cancer can be effectively cured. However, xerostomia is a common complication in patients after RT for HN cancer. The purpose of this study was to use the Lyman–Kutcher–Burman (LKB) model to derive parameters for the normal tissue complication probability (NTCP) for xerostomia based on scintigraphy assessments and quality of life (QoL) questionnaires. We performed validation tests of the Quantitative Analysis of Normal Tissue Effects in the Clinic (QUANTEC) guidelines against prospectively collected QoL and salivary scintigraphic data. Thirty-one patients with HN cancer were enrolled. Salivary excretion factors (SEFs) measured by scintigraphy and QoL data from self-reported questionnaires were used for NTCP modeling to describe the incidence of grade 3 + xerostomia. The NTCP parameters estimated from the QoL and SEF datasets were compared. Model performance was assessed using Pearson’s chi-squared test, Nagelkerke’s R 2 , the area under the receiver operating characteristic curve, and the Hosmer–Lemeshow test. The negative predictive value (NPV) was checked for the rate of correctly predicting the lack of incidence. Pearson’s chi-squared test was used to test the goodness of fit and association. Using the LKB NTCP model and assuming n=1, the dose for uniform irradiation of the whole or partial volume of the parotid gland that results in 50% probability of a complication (TD 50 ) and the slope of the dose–response curve (m) were determined from the QoL and SEF datasets, respectively. The NTCP-fitted parameters for local disease were TD 50 =43.6 Gy and m=0.18 with the SEF data, and TD 50 =44.1 Gy and m=0.11 with the QoL data. The rate of grade 3 + xerostomia for treatment plans meeting the QUANTEC guidelines was specifically predicted, with a NPV of 100%, using either the QoL or SEF dataset. Our study shows the agreement between the NTCP parameter modeling based on SEF and 9. Normal tissue complication probability model parameter estimation for xerostomia in head and neck cancer patients based on scintigraphy and quality of life assessments Science.gov (United States) 2012-01-01 Background With advances in modern radiotherapy (RT), many patients with head and neck (HN) cancer can be effectively cured. However, xerostomia is a common complication in patients after RT for HN cancer. The purpose of this study was to use the Lyman–Kutcher–Burman (LKB) model to derive parameters for the normal tissue complication probability (NTCP) for xerostomia based on scintigraphy assessments and quality of life (QoL) questionnaires. We performed validation tests of the Quantitative Analysis of Normal Tissue Effects in the Clinic (QUANTEC) guidelines against prospectively collected QoL and salivary scintigraphic data. Methods Thirty-one patients with HN cancer were enrolled. Salivary excretion factors (SEFs) measured by scintigraphy and QoL data from self-reported questionnaires were used for NTCP modeling to describe the incidence of grade 3+ xerostomia. The NTCP parameters estimated from the QoL and SEF datasets were compared. Model performance was assessed using Pearson’s chi-squared test, Nagelkerke’s R2, the area under the receiver operating characteristic curve, and the Hosmer–Lemeshow test. The negative predictive value (NPV) was checked for the rate of correctly predicting the lack of incidence. Pearson’s chi-squared test was used to test the goodness of fit and association. Results Using the LKB NTCP model and assuming n=1, the dose for uniform irradiation of the whole or partial volume of the parotid gland that results in 50% probability of a complication (TD50) and the slope of the dose–response curve (m) were determined from the QoL and SEF datasets, respectively. The NTCP-fitted parameters for local disease were TD50=43.6 Gy and m=0.18 with the SEF data, and TD50=44.1 Gy and m=0.11 with the QoL data. The rate of grade 3+ xerostomia for treatment plans meeting the QUANTEC guidelines was specifically predicted, with a NPV of 100%, using either the QoL or SEF dataset. Conclusions Our study shows the agreement between the NTCP 10. Normal tissue complication probability model parameter estimation for xerostomia in head and neck cancer patients based on scintigraphy and quality of life assessments Directory of Open Access Journals (Sweden) Lee Tsair-Fwu 2012-12-01 Full Text Available Abstract Background With advances in modern radiotherapy (RT, many patients with head and neck (HN cancer can be effectively cured. However, xerostomia is a common complication in patients after RT for HN cancer. The purpose of this study was to use the Lyman–Kutcher–Burman (LKB model to derive parameters for the normal tissue complication probability (NTCP for xerostomia based on scintigraphy assessments and quality of life (QoL questionnaires. We performed validation tests of the Quantitative Analysis of Normal Tissue Effects in the Clinic (QUANTEC guidelines against prospectively collected QoL and salivary scintigraphic data. Methods Thirty-one patients with HN cancer were enrolled. Salivary excretion factors (SEFs measured by scintigraphy and QoL data from self-reported questionnaires were used for NTCP modeling to describe the incidence of grade 3+ xerostomia. The NTCP parameters estimated from the QoL and SEF datasets were compared. Model performance was assessed using Pearson’s chi-squared test, Nagelkerke’s R2, the area under the receiver operating characteristic curve, and the Hosmer–Lemeshow test. The negative predictive value (NPV was checked for the rate of correctly predicting the lack of incidence. Pearson’s chi-squared test was used to test the goodness of fit and association. Results Using the LKB NTCP model and assuming n=1, the dose for uniform irradiation of the whole or partial volume of the parotid gland that results in 50% probability of a complication (TD50 and the slope of the dose–response curve (m were determined from the QoL and SEF datasets, respectively. The NTCP-fitted parameters for local disease were TD50=43.6 Gy and m=0.18 with the SEF data, and TD50=44.1 Gy and m=0.11 with the QoL data. The rate of grade 3+ xerostomia for treatment plans meeting the QUANTEC guidelines was specifically predicted, with a NPV of 100%, using either the QoL or SEF dataset. Conclusions Our study shows the agreement 11. Development of a multivariable normal tissue complication probability (NTCP) model for tube feeding dependence after curative radiotherapy/chemo-radiotherapy in head and neck cancer International Nuclear Information System (INIS) Wopken, Kim; Bijl, Hendrik P.; Schaaf, Arjen van der; Laan, Hans Paul van der; Chouvalova, Olga; Steenbakkers, Roel J.H.M.; Doornaert, Patricia; Slotman, Ben J.; Oosting, Sjoukje F.; Christianen, Miranda E.M.C.; Laan, Bernard F.A.M. van der; Roodenburg, Jan L.N.; René Leemans, C.; Verdonck-de Leeuw, Irma M.; Langendijk, Johannes A. 2014-01-01 12. Normal tissue complication probability modeling for cochlea constraints to avoid causing tinnitus after head-and-neck intensity-modulated radiation therapy International Nuclear Information System (INIS) Lee, Tsair-Fwu; Yeh, Shyh-An; Chao, Pei-Ju; Chang, Liyun; Chiu, Chien-Liang; Ting, Hui-Min; Wang, Hung-Yu; Huang, Yu-Jie 2015-01-01 Radiation-induced tinnitus is a side effect of radiotherapy in the inner ear for cancers of the head and neck. Effective dose constraints for protecting the cochlea are under-reported. The aim of this study is to determine the cochlea dose limitation to avoid causing tinnitus after head-and-neck cancer (HNC) intensity-modulated radiation therapy (IMRT). In total 211 patients with HNC were included; the side effects of radiotherapy were investigated for 422 inner ears in the cohort. Forty-nine of the four hundred and twenty-two samples (11.6 %) developed grade 2+ tinnitus symptoms after IMRT, as diagnosed by a clinician. The Late Effects of Normal Tissues–Subjective, Objective, Management, Analytic (LENT-SOMA) criteria were used for tinnitus evaluation. The logistic and Lyman-Kutcher-Burman (LKB) normal tissue complication probability (NTCP) models were used for the analyses. The NTCP-fitted parameters were TD 50 = 46.31 Gy (95 % CI, 41.46–52.50), γ 50 = 1.27 (95 % CI, 1.02–1.55), and TD 50 = 46.52 Gy (95 % CI, 41.91–53.43), m = 0.35 (95 % CI, 0.30–0.42) for the logistic and LKB models, respectively. The suggested guideline TD 20 for the tolerance dose to produce a 20 % complication rate within a specific period of time was TD 20 = 33.62 Gy (95 % CI, 30.15–38.27) (logistic) and TD 20 = 32.82 Gy (95 % CI, 29.58–37.69) (LKB). To maintain the incidence of grade 2+ tinnitus toxicity <20 % in IMRT, we suggest that the mean dose to the cochlea should be <32 Gy. However, models should not be extrapolated to other patient populations without further verification and should first be confirmed before clinical implementation 13. Model uncertainty and probability International Nuclear Information System (INIS) Parry, G.W. 1994-01-01 This paper discusses the issue of model uncertainty. The use of probability as a measure of an analyst's uncertainty as well as a means of describing random processes has caused some confusion, even though the two uses are representing different types of uncertainty with respect to modeling a system. The importance of maintaining the distinction between the two types is illustrated with a simple example 14. The estimation of collision probabilities in complicated geometries International Nuclear Information System (INIS) Roth, M.J. 1969-04-01 This paper demonstrates how collision probabilities in complicated geometries may be estimated. It is assumed that the reactor core may be divided into a number of cells each with simple geometry so that a collision probability matrix can be calculated for each cell by standard methods. It is then shown how these may be joined together. (author) 15. Impact of Chemotherapy on Normal Tissue Complication Probability Models of Acute Hematologic Toxicity in Patients Receiving Pelvic Intensity Modulated Radiation Therapy Energy Technology Data Exchange (ETDEWEB) Bazan, Jose G.; Luxton, Gary; Kozak, Margaret M.; Anderson, Eric M.; Hancock, Steven L.; Kapp, Daniel S.; Kidd, Elizabeth A.; Koong, Albert C.; Chang, Daniel T., E-mail: dtchang@stanford.edu 2013-12-01 Purpose: To determine how chemotherapy agents affect radiation dose parameters that correlate with acute hematologic toxicity (HT) in patients treated with pelvic intensity modulated radiation therapy (P-IMRT) and concurrent chemotherapy. Methods and Materials: We assessed HT in 141 patients who received P-IMRT for anal, gynecologic, rectal, or prostate cancers, 95 of whom received concurrent chemotherapy. Patients were separated into 4 groups: mitomycin (MMC) + 5-fluorouracil (5FU, 37 of 141), platinum ± 5FU (Cis, 32 of 141), 5FU (26 of 141), and P-IMRT alone (46 of 141). The pelvic bone was contoured as a surrogate for pelvic bone marrow (PBM) and divided into subsites: ilium, lower pelvis, and lumbosacral spine (LSS). The volumes of each region receiving 5-40 Gy were calculated. The endpoint for HT was grade ≥3 (HT3+) leukopenia, neutropenia or thrombocytopenia. Normal tissue complication probability was calculated using the Lyman-Kutcher-Burman model. Logistic regression was used to analyze association between HT3+ and dosimetric parameters. Results: Twenty-six patients experienced HT3+: 10 of 37 (27%) MMC, 14 of 32 (44%) Cis, 2 of 26 (8%) 5FU, and 0 of 46 P-IMRT. PBM dosimetric parameters were correlated with HT3+ in the MMC group but not in the Cis group. LSS dosimetric parameters were well correlated with HT3+ in both the MMC and Cis groups. Constrained optimization (0model resulted in n=1, m = 0.11, TD{sub 50} = 31 Gy for LSS in the MMC group and n=1, m = 0.27, TD{sub 50} = 35 Gy for LSS in the Cis group. Conclusions: The incidence of HT3+ depends on type of chemotherapy received. Patients receiving P-IMRT ± 5FU have better bone marrow tolerance than those receiving irradiation concurrent with either Cis or MMC. Treatment with MMC has a lower TD{sub 50} and more steeply rising normal tissue complication probability curve compared with treatment with Cis. Dose tolerance of PBM and the LSS subsite may be lower for 16. Calculation of cranial nerve complication probability for acoustic neuroma radiosurgery International Nuclear Information System (INIS) Meeks, Sanford L.; Buatti, John M.; Foote, Kelly D.; Friedman, William A.; Bova, Francis J. 2000-01-01 Purpose: Estimations of complications from stereotactic radiosurgery usually rely simply on dose-volume or dose-diameter isoeffect curves. Due to the sparse clinical data available, these curves have typically not considered the target location in the brain, target histology, or treatment plan conformality as parameters in the calculation. In this study, a predictive model was generated to estimate the probability of cranial neuropathies as a result of acoustic schwannoma radiosurgery. Methods and Materials: The dose-volume histogram reduction scheme was used to calculate the normal tissue complication probability (NTCP) from brainstem dose-volume histograms. The model's fitting parameters were optimized to provide the best fit to the observed complication data for acoustic neuroma patients treated with stereotactic radiosurgery at the University of Florida. The calculation was then applied to the remainder of the patients in the database. Results: The best fit to our clinical data was obtained using n = 0.04, m = 0.15, and no. alphano. /no. betano. = 2.1 Gy -1 . Although the fitting parameter m is relatively consistent with ranges found in the literature, both the volume parameter, n, and no. alphano. /no. betano. are much smaller than the values quoted in the literature. The fit to our clinical data indicates that brainstem, or possibly a specific portion of the brainstem, is more radiosensitive than the parameters in the literature indicate, and that there is very little volume effect; in other words, irradiation of a small fraction of the brainstem yields NTCPs that are nearly as high as those calculated for entire volume irradiation. These new fitting parameters are specific to acoustic neuroma radiosurgery, and the small volume effect that we observe may be an artifact of the fixed relationship of acoustic tumors to specific regions of the brainstem. Applying the model to our patient database, we calculate an average NTCP of 7.2% for patients who had no 17. Information-theoretic methods for estimating of complicated probability distributions CERN Document Server Zong, Zhi 2006-01-01 Mixing up various disciplines frequently produces something that are profound and far-reaching. Cybernetics is such an often-quoted example. Mix of information theory, statistics and computing technology proves to be very useful, which leads to the recent development of information-theory based methods for estimating complicated probability distributions. Estimating probability distribution of a random variable is the fundamental task for quite some fields besides statistics, such as reliability, probabilistic risk analysis (PSA), machine learning, pattern recognization, image processing, neur 18. Model uncertainty: Probabilities for models? International Nuclear Information System (INIS) Winkler, R.L. 1994-01-01 Like any other type of uncertainty, model uncertainty should be treated in terms of probabilities. The question is how to do this. The most commonly-used approach has a drawback related to the interpretation of the probabilities assigned to the models. If we step back and look at the big picture, asking what the appropriate focus of the model uncertainty question should be in the context of risk and decision analysis, we see that a different probabilistic approach makes more sense, although it raise some implementation questions. Current work that is underway to address these questions looks very promising 19. Probability and stochastic modeling CERN Document Server 2012-01-01 Basic NotionsSample Space and EventsProbabilitiesCounting TechniquesIndependence and Conditional ProbabilityIndependenceConditioningThe Borel-Cantelli TheoremDiscrete Random VariablesRandom Variables and VectorsExpected ValueVariance and Other Moments. Inequalities for DeviationsSome Basic DistributionsConvergence of Random Variables. The Law of Large NumbersConditional ExpectationGenerating Functions. Branching Processes. Random Walk RevisitedBranching Processes Generating Functions Branching Processes Revisited More on Random WalkMarkov ChainsDefinitions and Examples. Probability Distributions of Markov ChainsThe First Step Analysis. Passage TimesVariables Defined on a Markov ChainErgodicity and Stationary DistributionsA Classification of States and ErgodicityContinuous Random VariablesContinuous DistributionsSome Basic Distributions Continuous Multivariate Distributions Sums of Independent Random Variables Conditional Distributions and ExpectationsDistributions in the General Case. SimulationDistribution F... 20. Incidence of late rectal bleeding in high-dose conformal radiotherapy of prostate cancer using equivalent uniform dose-based and dose-volume-based normal tissue complication probability models International Nuclear Information System (INIS) Soehn, Matthias; Yan Di; Liang Jian; Meldolesi, Elisa; Vargas, Carlos; Alber, Markus 2007-01-01 Purpose: Accurate modeling of rectal complications based on dose-volume histogram (DVH) data are necessary to allow safe dose escalation in radiotherapy of prostate cancer. We applied different equivalent uniform dose (EUD)-based and dose-volume-based normal tissue complication probability (NTCP) models to rectal wall DVHs and follow-up data for 319 prostate cancer patients to identify the dosimetric factors most predictive for Grade ≥ 2 rectal bleeding. Methods and Materials: Data for 319 patients treated at the William Beaumont Hospital with three-dimensional conformal radiotherapy (3D-CRT) under an adaptive radiotherapy protocol were used for this study. The following models were considered: (1) Lyman model and (2) logit-formula with DVH reduced to generalized EUD (3) serial reconstruction unit (RU) model (4) Poisson-EUD model, and (5) mean dose- and (6) cutoff dose-logistic regression model. The parameters and their confidence intervals were determined using maximum likelihood estimation. Results: Of the patients, 51 (16.0%) showed Grade 2 or higher bleeding. As assessed qualitatively and quantitatively, the Lyman- and Logit-EUD, serial RU, and Poisson-EUD model fitted the data very well. Rectal wall mean dose did not correlate to Grade 2 or higher bleeding. For the cutoff dose model, the volume receiving > 73.7 Gy showed most significant correlation to bleeding. However, this model fitted the data more poorly than the EUD-based models. Conclusions: Our study clearly confirms a volume effect for late rectal bleeding. This can be described very well by the EUD-like models, of which the serial RU- and Poisson-EUD model can describe the data with only two parameters. Dose-volume-based cutoff-dose models performed worse 1. Neutron transport by collision probability method in complicated geometries International Nuclear Information System (INIS) Constantin, Marin 2000-01-01 For the first flight collision probability (FFCP) method a rapidly increasing of the memory requirements and execution time with the number of discrete regions occurs. Generally, the use of the method is restricted at cell/supercell level. However, the amazing developments both in computer hardware and computer architecture allow a real extending of the problems' domain and a more detailed treatment of the geometry. Two ways are discussed into the paper: the direct design of new codes and the improving of the mainframe old versions. The author's experience is focused on the performances' improving of the 3D integral transport code PIJXYZ (from an old version to a modern one) and on the design and developing of the 2D transport code CP 2 D in the last years. In the first case an optimization process have been performed before the parallelization. In the second a modular design and the newest techniques (factorization of the geometry, the macrobands method, the mobile set of chords, the automatic calculation of the integration error, optimal algorithms for the innermost programming level, the mixed method for tracking process and CPs calculation, etc.) were adopted. In both cases the parallelization uses a PCs network system. Some short examples for CP 2 D and PIJXYZ calculation are presented: reactivity void effect in typical CANDU cells using a multistratified coolant model, a problem of some adjacent fuel assemblies, CANDU reactivity devices 3D simulation. (author) 2. Predictive modeling of complications. Science.gov (United States) Osorio, Joseph A; Scheer, Justin K; Ames, Christopher P 2016-09-01 Predictive analytic algorithms are designed to identify patterns in the data that allow for accurate predictions without the need for a hypothesis. Therefore, predictive modeling can provide detailed and patient-specific information that can be readily applied when discussing the risks of surgery with a patient. There are few studies using predictive modeling techniques in the adult spine surgery literature. These types of studies represent the beginning of the use of predictive analytics in spine surgery outcomes. We will discuss the advancements in the field of spine surgery with respect to predictive analytics, the controversies surrounding the technique, and the future directions. 3. A COMPLICATED GRIEF INTERVENTION MODEL African Journals Online (AJOL) 2010-07-29 Jul 29, 2010 ... work approach to therapy) in an attempt to develop a model for complicated grief intervention ..... way feels right for them, thus making them equal partners with .... This question points out the benefits of success and affords the. 4. A simple method to calculate the influence of dose inhomogeneity and fractionation in normal tissue complication probability evaluation International Nuclear Information System (INIS) Begnozzi, L.; Gentile, F.P.; Di Nallo, A.M.; Chiatti, L.; Zicari, C.; Consorti, R.; Benassi, M. 1994-01-01 Since volumetric dose distributions are available with 3-dimensional radiotherapy treatment planning they can be used in statistical evaluation of response to radiation. This report presents a method to calculate the influence of dose inhomogeneity and fractionation in normal tissue complication probability evaluation. The mathematical expression for the calculation of normal tissue complication probability has been derived combining the Lyman model with the histogram reduction method of Kutcher et al. and using the normalized total dose (NTD) instead of the total dose. The fitting of published tolerance data, in case of homogeneous or partial brain irradiation, has been considered. For the same total or partial volume homogeneous irradiation of the brain, curves of normal tissue complication probability have been calculated with fraction size of 1.5 Gy and of 3 Gy instead of 2 Gy, to show the influence of fraction size. The influence of dose distribution inhomogeneity and α/β value has also been simulated: Considering α/β=1.6 Gy or α/β=4.1 Gy for kidney clinical nephritis, the calculated curves of normal tissue complication probability are shown. Combining NTD calculations and histogram reduction techniques, normal tissue complication probability can be estimated taking into account the most relevant contributing factors, including the volume effect. (orig.) [de 5. Optimization of radiation therapy, III: a method of assessing complication probabilities from dose-volume histograms International Nuclear Information System (INIS) Lyman, J.T.; Wolbarst, A.B. 1987-01-01 To predict the likelihood of success of a therapeutic strategy, one must be able to assess the effects of the treatment upon both diseased and healthy tissues. This paper proposes a method for determining the probability that a healthy organ that receives a non-uniform distribution of X-irradiation, heat, chemotherapy, or other agent will escape complications. Starting with any given dose distribution, a dose-cumulative-volume histogram for the organ is generated. This is then reduced by an interpolation scheme (involving the volume-weighting of complication probabilities) to a slightly different histogram that corresponds to the same overall likelihood of complications, but which contains one less step. The procedure is repeated, one step at a time, until there remains a final, single-step histogram, for which the complication probability can be determined. The formalism makes use of a complication response function C(D, V) which, for the given treatment schedule, represents the probability of complications arising when the fraction V of the organ receives dose D and the rest of the organ gets none. Although the data required to generate this function are sparse at present, it should be possible to obtain the necessary information from in vivo and clinical studies. Volume effects are taken explicitly into account in two ways: the precise shape of the patient's histogram is employed in the calculation, and the complication response function is a function of the volume 6. Calculation of normal tissue complication probability and dose-volume histogram reduction schemes for tissues with a critical element architecture International Nuclear Information System (INIS) Niemierko, Andrzej; Goitein, Michael 1991-01-01 The authors investigate a model of normal tissue complication probability for tissues that may be represented by a critical element architecture. They derive formulas for complication probability that apply to both a partial volume irradiation and to an arbitrary inhomogeneous dose distribution. The dose-volume isoeffect relationship which is a consequence of a critical element architecture is discussed and compared to the empirical power law relationship. A dose-volume histogram reduction scheme for a 'pure' critical element model is derived. In addition, a point-based algorithm which does not require precomputation of a dose-volume histogram is derived. The existing published dose-volume histogram reduction algorithms are analyzed. The authors show that the existing algorithms, developed empirically without an explicit biophysical model, have a close relationship to the critical element model at low levels of complication probability. However, it is also showed that they have aspects which are not compatible with a critical element model and the authors propose a modification to one of them to circumvent its restriction to low complication probabilities. (author). 26 refs.; 7 figs 7. Probabilities of Pulmonary and Cardiac Complications and Radiographic Parameters in Breast Cancer Radiotherapy Energy Technology Data Exchange (ETDEWEB) Noh, O Kyu; Paek, Sung Ho; Ahn, Seung Do; Choi, Eun Kyung; Lee, Sang Wook; Song, Si Yeol; Yoon, Sang Min; Kim, Jong Hoon [Dept. of Radiation Oncology, Asan Medical Center, University of Ulsan College of Medicine, Seoul (Korea, Republic of) 2010-11-15 8. Probabilities of Pulmonary and Cardiac Complications and Radiographic Parameters in Breast Cancer Radiotherapy International Nuclear Information System (INIS) Noh, O Kyu; Paek, Sung Ho; Ahn, Seung Do; Choi, Eun Kyung; Lee, Sang Wook; Song, Si Yeol; Yoon, Sang Min; Kim, Jong Hoon 2010-01-01 9. The transition probabilities of the reciprocity model NARCIS (Netherlands) Snijders, T.A.B. 1999-01-01 The reciprocity model is a continuous-time Markov chain model used for modeling longitudinal network data. A new explicit expression is derived for its transition probability matrix. This expression can be checked relatively easily. Some properties of the transition probabilities are given, as well 10. Comparing linear probability model coefficients across groups DEFF Research Database (Denmark) Holm, Anders; Ejrnæs, Mette; Karlson, Kristian Bernt 2015-01-01 of the following three components: outcome truncation, scale parameters and distributional shape of the predictor variable. These results point to limitations in using linear probability model coefficients for group comparisons. We also provide Monte Carlo simulations and real examples to illustrate......This article offers a formal identification analysis of the problem in comparing coefficients from linear probability models between groups. We show that differences in coefficients from these models can result not only from genuine differences in effects, but also from differences in one or more...... these limitations, and we suggest a restricted approach to using linear probability model coefficients in group comparisons.... 11. The use of normal tissue complication probability to predict radiation hepatitis International Nuclear Information System (INIS) Keum, Ki Chang; Seong, Jin Sil; Suh, Chang Ok; Lee, Sang Wook; Chung, Eun Ji; Shin, Hyun Soo; Kim, Gwi Eon 2000-01-01 12. Dependency models and probability of joint events International Nuclear Information System (INIS) Oerjasaeter, O. 1982-08-01 Probabilistic dependencies between components/systems are discussed with reference to a broad classification of potential failure mechanisms. Further, a generalized time-dependency model, based on conditional probabilities for estimation of the probability of joint events and event sequences is described. The applicability of this model is clarified/demonstrated by various examples. It is concluded that the described model of dependency is a useful tool for solving a variety of practical problems concerning the probability of joint events and event sequences where common cause and time-dependent failure mechanisms are involved. (Auth.) 13. A probability space for quantum models Science.gov (United States) Lemmens, L. F. 2017-06-01 A probability space contains a set of outcomes, a collection of events formed by subsets of the set of outcomes and probabilities defined for all events. A reformulation in terms of propositions allows to use the maximum entropy method to assign the probabilities taking some constraints into account. The construction of a probability space for quantum models is determined by the choice of propositions, choosing the constraints and making the probability assignment by the maximum entropy method. This approach shows, how typical quantum distributions such as Maxwell-Boltzmann, Fermi-Dirac and Bose-Einstein are partly related with well-known classical distributions. The relation between the conditional probability density, given some averages as constraints and the appropriate ensemble is elucidated. 14. Modelling the probability of building fires Directory of Open Access Journals (Sweden) Vojtěch Barták 2014-12-01 Full Text Available Systematic spatial risk analysis plays a crucial role in preventing emergencies.In the Czech Republic, risk mapping is currently based on the risk accumulationprinciple, area vulnerability, and preparedness levels of Integrated Rescue Systemcomponents. Expert estimates are used to determine risk levels for individualhazard types, while statistical modelling based on data from actual incidents andtheir possible causes is not used. Our model study, conducted in cooperation withthe Fire Rescue Service of the Czech Republic as a model within the Liberec andHradec Králové regions, presents an analytical procedure leading to the creation ofbuilding fire probability maps based on recent incidents in the studied areas andon building parameters. In order to estimate the probability of building fires, aprediction model based on logistic regression was used. Probability of fire calculatedby means of model parameters and attributes of specific buildings can subsequentlybe visualized in probability maps. 15. Sampling, Probability Models and Statistical Reasoning Statistical Home; Journals; Resonance – Journal of Science Education; Volume 1; Issue 5. Sampling, Probability Models and Statistical Reasoning Statistical Inference. Mohan Delampady V R Padmawar. General Article Volume 1 Issue 5 May 1996 pp 49-58 ... 16. Correlations and Non-Linear Probability Models DEFF Research Database (Denmark) Breen, Richard; Holm, Anders; Karlson, Kristian Bernt 2014-01-01 the dependent variable of the latent variable model and its predictor variables. We show how this correlation can be derived from the parameters of non-linear probability models, develop tests for the statistical significance of the derived correlation, and illustrate its usefulness in two applications. Under......Although the parameters of logit and probit and other non-linear probability models are often explained and interpreted in relation to the regression coefficients of an underlying linear latent variable model, we argue that they may also be usefully interpreted in terms of the correlations between...... certain circumstances, which we explain, the derived correlation provides a way of overcoming the problems inherent in cross-sample comparisons of the parameters of non-linear probability models.... 17. Modeling experiments using quantum and Kolmogorov probability International Nuclear Information System (INIS) Hess, Karl 2008-01-01 Criteria are presented that permit a straightforward partition of experiments into sets that can be modeled using both quantum probability and the classical probability framework of Kolmogorov. These new criteria concentrate on the operational aspects of the experiments and lead beyond the commonly appreciated partition by relating experiments to commuting and non-commuting quantum operators as well as non-entangled and entangled wavefunctions. In other words the space of experiments that can be understood using classical probability is larger than usually assumed. This knowledge provides advantages for areas such as nanoscience and engineering or quantum computation. 18. Comparing coefficients of nested nonlinear probability models DEFF Research Database (Denmark) Kohler, Ulrich; Karlson, Kristian Bernt; Holm, Anders 2011-01-01 In a series of recent articles, Karlson, Holm and Breen have developed a method for comparing the estimated coeffcients of two nested nonlinear probability models. This article describes this method and the user-written program khb that implements the method. The KHB-method is a general decomposi......In a series of recent articles, Karlson, Holm and Breen have developed a method for comparing the estimated coeffcients of two nested nonlinear probability models. This article describes this method and the user-written program khb that implements the method. The KHB-method is a general...... decomposition method that is unaffected by the rescaling or attenuation bias that arise in cross-model comparisons in nonlinear models. It recovers the degree to which a control variable, Z, mediates or explains the relationship between X and a latent outcome variable, Y*, underlying the nonlinear probability... 19. Investigation of normal tissue complication probabilities in prostate and partial breast irradiation radiotherapy techniques International Nuclear Information System (INIS) Bezak, E.; Takam, R.; Bensaleh, S.; Yeoh, E.; Marcu, L. 2011-01-01 Full text: Normal- Tissue-Complication Probabilities of rectum, bladder and urethra following various radiation techniques for prostate cancer were evaluated using the relative-seriality and Lyman models. NTCPs of lungs, heart and skin, their dependence on sourceposition, balloon-deformation were also investigated for HDR mammosite brachytherapy. The prostate treatment techniques included external three dimentional conformal-radiotherapy, Low-Dose-Rate brachytherapy (1-125), High-Dose-Rate brachytherapy (Ir-I92). Dose- Volume-Histograms of critical structures for prostate and breast radiotherapy, retrieved from corresponding treatment planning systems, were converted to Biological Effective Dose (BEffD)-based and Equivalent Dose(Deq)-based DVHs to account for differences in radiation delivery and fractionation schedule. Literature-based model parameters were used to calculate NTCPs. Hypofractionated 3D-CRT (2.75 Gy/fraction, total dose 55 Gy) NTCPs of rectum, bladder and urethra were less than those for standard fractionated 4-field 3D-CRT (2-Gy/fraction, 64 Gy) and dose-escalated 4- and 5-field 3D-CRT (74 Gy). Rectal and bladder NTCPs (5.2% and 6.6%) following the dose-escalated 4-field 3D-CRT (74 Gy) were the highest among analyzed techniques. The average NTCP for rectum and urethra were 0.6% and 24.7% for LDRBT and 0.5% and 11.2% for HDR-BT. For Mammosite, NTCP was estimated to be 0.1 %, 0.1 %, 1.2% and 3.5% for skin desquamation, erythema, telangiectasia and fibrosis respectively (the source positioned at the balloon centre). A 4 mm Mammosite-balloon deformation leads to overdosing of PTV regions by ∼40%, resulting in excessive skin dose and increased NTCP. Conclusions Prostate brachytherapy resulted in NTCPs lower compared to external beam techniques. Mammosite-brachytherapy resulted in no heart/lung complications regardless of balloon deformation. However, 4 mm deformation caused 0.6% increase in tissue fibrosis NTCP. 20. Uncertainty the soul of modeling, probability & statistics CERN Document Server Briggs, William 2016-01-01 This book presents a philosophical approach to probability and probabilistic thinking, considering the underpinnings of probabilistic reasoning and modeling, which effectively underlie everything in data science. The ultimate goal is to call into question many standard tenets and lay the philosophical and probabilistic groundwork and infrastructure for statistical modeling. It is the first book devoted to the philosophy of data aimed at working scientists and calls for a new consideration in the practice of probability and statistics to eliminate what has been referred to as the "Cult of Statistical Significance". The book explains the philosophy of these ideas and not the mathematics, though there are a handful of mathematical examples. The topics are logically laid out, starting with basic philosophy as related to probability, statistics, and science, and stepping through the key probabilistic ideas and concepts, and ending with statistical models. Its jargon-free approach asserts that standard methods, suc... 1. The Probability Heuristics Model of Syllogistic Reasoning. Science.gov (United States) Chater, Nick; Oaksford, Mike 1999-01-01 Proposes a probability heuristic model for syllogistic reasoning and confirms the rationality of this heuristic by an analysis of the probabilistic validity of syllogistic reasoning that treats logical inference as a limiting case of probabilistic inference. Meta-analysis and two experiments involving 40 adult participants and using generalized… 2. Sampling, Probability Models and Statistical Reasoning -RE ... random sampling allows data to be modelled with the help of probability ... g based on different trials to get an estimate of the experimental error. ... research interests lie in the .... if e is indeed the true value of the proportion of defectives in the. 3. Applied probability models with optimization applications CERN Document Server Ross, Sheldon M 1992-01-01 Concise advanced-level introduction to stochastic processes that frequently arise in applied probability. Largely self-contained text covers Poisson process, renewal theory, Markov chains, inventory theory, Brownian motion and continuous time optimization models, much more. Problems and references at chapter ends. ""Excellent introduction."" - Journal of the American Statistical Association. Bibliography. 1970 edition. 4. A COMPLICATED GRIEF INTERVENTION MODEL African Journals Online (AJOL) 2010-07-29 Jul 29, 2010 ... of complicated grief as a contributing factor to impaired social functioning. This can ... includes a diagnosis of bereavement-related major depression if symptoms ..... networking, Social Sciences Citation Index, Social Sciences. 5. Normal tissue complication probabilities correlated with late effects in the rectum after prostate conformal radiotherapy International Nuclear Information System (INIS) Dale, Einar; Olsen, Dag R.; Fossa, Sophie D. 1999-01-01 Purpose: Radiation therapy of deep-sited tumours will always result in normal tissue doses to some extent. The aim of this study was to calculate different risk estimates of late effects in the rectum for a group of cancer prostate patients treated with conformal radiation therapy (CRT) and correlate these estimates with the occurrences of late effects. Since the rectum is a hollow organ, several ways of generating dose-volume distributions over the organ are possible, and we wanted to investigate two of them. Methods and Materials: A mathematical model, known as the Lyman-Kutcher model, conventionally used to estimate normal tissue complication probabilities (NTCP) associated with radiation therapy, was applied to a material of 52 cancer prostate patients. The patients were treated with a four field box technique, with the rectum as organ at risk. Dose-volume histograms (DVH) were generated for the whole rectum (including the cavity) and of the rectum wall. One to two years after the treatment, the patients completed a questionnaire concerning bowel (rectum) related morbidity quantifying the extent of late effects. Results: A correlation analysis using Spearman's rank correlation coefficient, for NTCP values calculated from the DVHs and the patients' scores, gave correlation coefficients which were not statistically significant at the p max , of the whole rectum, correlated better to observed late toxicity than D max derived from histograms of the rectum wall. Correlation coefficients from 'high-dose' measures were larger than those calculated from the NTCP values. Accordingly, as the volume parameter of the Lyman-Kutcher model was reduced, raising the impact of small high-dose volumes on the NTCP values, the correlation between observed effects and NTCP values became significant at p < 0.01 level. Conclusions: 1) High-dose levels corresponding to small volume fractions of the cumulative dose-volume histograms were best correlated with the occurrences of late 6. Multiple model cardinalized probability hypothesis density filter Science.gov (United States) Georgescu, Ramona; Willett, Peter 2011-09-01 The Probability Hypothesis Density (PHD) filter propagates the first-moment approximation to the multi-target Bayesian posterior distribution while the Cardinalized PHD (CPHD) filter propagates both the posterior likelihood of (an unlabeled) target state and the posterior probability mass function of the number of targets. Extensions of the PHD filter to the multiple model (MM) framework have been published and were implemented either with a Sequential Monte Carlo or a Gaussian Mixture approach. In this work, we introduce the multiple model version of the more elaborate CPHD filter. We present the derivation of the prediction and update steps of the MMCPHD particularized for the case of two target motion models and proceed to show that in the case of a single model, the new MMCPHD equations reduce to the original CPHD equations. 7. Discovering Diabetes Complications: an Ontology Based Model. Science.gov (United States) 2015-12-01 Diabetes is a serious disease that spread in the world dramatically. The diabetes patient has an average of risk to experience complications. Take advantage of recorded information to build ontology as information technology solution will help to predict patients who have average of risk level with certain complication. It is helpful to search and present patient's history regarding different risk factors. Discovering diabetes complications could be useful to prevent or delay the complications. We designed ontology based model, using adult diabetes patients' data, to discover the rules of diabetes with its complications in disease to disease relationship. Various rules between different risk factors of diabetes Patients and certain complications generated. Furthermore, new complications (diseases) might be discovered as new finding of this study, discovering diabetes complications could be useful to prevent or delay the complications. The system can identify the patients who are suffering from certain risk factors such as high body mass index (obesity) and starting controlling and maintaining plan. 8. Statistical physics of pairwise probability models DEFF Research Database (Denmark) Roudi, Yasser; Aurell, Erik; Hertz, John 2009-01-01 (dansk abstrakt findes ikke) Statistical models for describing the probability distribution over the states of biological systems are commonly used for dimensional reduction. Among these models, pairwise models are very attractive in part because they can be fit using a reasonable amount of  data......: knowledge of the means and correlations between pairs of elements in the system is sufficient. Not surprisingly, then, using pairwise models for studying neural data has been the focus of many studies in recent years. In this paper, we describe how tools from statistical physics can be employed for studying... 9. The Probability Model of Expectation Disconfirmation Process Directory of Open Access Journals (Sweden) Hui-Hsin HUANG 2015-06-01 Full Text Available This paper proposes a probability model to explore the dynamic process of customer’s satisfaction. Bases on expectation disconfirmation theory, the satisfaction is constructed with customer’s expectation before buying behavior and the perceived performance after purchase. The experiment method is designed to measure expectation disconfirmation effects and we also use the collection data to estimate the overall satisfaction and model calibration. The results show good fitness between the model and the real data. This model has application for business marketing areas in order to manage relationship satisfaction. 10. A quantum probability model of causal reasoning Directory of Open Access Journals (Sweden) Jennifer S Trueblood 2012-05-01 Full Text Available People can often outperform statistical methods and machine learning algorithms in situations that involve making inferences about the relationship between causes and effects. While people are remarkably good at causal reasoning in many situations, there are several instances where they deviate from expected responses. This paper examines three situations where judgments related to causal inference problems produce unexpected results and describes a quantum inference model based on the axiomatic principles of quantum probability theory that can explain these effects. Two of the three phenomena arise from the comparison of predictive judgments (i.e., the conditional probability of an effect given a cause with diagnostic judgments (i.e., the conditional probability of a cause given an effect. The third phenomenon is a new finding examining order effects in predictive causal judgments. The quantum inference model uses the notion of incompatibility among different causes to account for all three phenomena. Psychologically, the model assumes that individuals adopt different points of view when thinking about different causes. The model provides good fits to the data and offers a coherent account for all three causal reasoning effects thus proving to be a viable new candidate for modeling human judgment. 11. A new formula for normal tissue complication probability (NTCP) as a function of equivalent uniform dose (EUD). Science.gov (United States) Luxton, Gary; Keall, Paul J; King, Christopher R 2008-01-07 To facilitate the use of biological outcome modeling for treatment planning, an exponential function is introduced as a simpler equivalent to the Lyman formula for calculating normal tissue complication probability (NTCP). The single parameter of the exponential function is chosen to reproduce the Lyman calculation to within approximately 0.3%, and thus enable easy conversion of data contained in empirical fits of Lyman parameters for organs at risk (OARs). Organ parameters for the new formula are given in terms of Lyman model m and TD(50), and conversely m and TD(50) are expressed in terms of the parameters of the new equation. The role of the Lyman volume-effect parameter n is unchanged from its role in the Lyman model. For a non-homogeneously irradiated OAR, an equation relates d(ref), n, v(eff) and the Niemierko equivalent uniform dose (EUD), where d(ref) and v(eff) are the reference dose and effective fractional volume of the Kutcher-Burman reduction algorithm (i.e. the LKB model). It follows in the LKB model that uniform EUD irradiation of an OAR results in the same NTCP as the original non-homogeneous distribution. The NTCP equation is therefore represented as a function of EUD. The inverse equation expresses EUD as a function of NTCP and is used to generate a table of EUD versus normal tissue complication probability for the Emami-Burman parameter fits as well as for OAR parameter sets from more recent data. 12. A new formula for normal tissue complication probability (NTCP) as a function of equivalent uniform dose (EUD) International Nuclear Information System (INIS) Luxton, Gary; Keall, Paul J; King, Christopher R 2008-01-01 To facilitate the use of biological outcome modeling for treatment planning, an exponential function is introduced as a simpler equivalent to the Lyman formula for calculating normal tissue complication probability (NTCP). The single parameter of the exponential function is chosen to reproduce the Lyman calculation to within ∼0.3%, and thus enable easy conversion of data contained in empirical fits of Lyman parameters for organs at risk (OARs). Organ parameters for the new formula are given in terms of Lyman model m and TD 50 , and conversely m and TD 50 are expressed in terms of the parameters of the new equation. The role of the Lyman volume-effect parameter n is unchanged from its role in the Lyman model. For a non-homogeneously irradiated OAR, an equation relates d ref , n, v eff and the Niemierko equivalent uniform dose (EUD), where d ref and v eff are the reference dose and effective fractional volume of the Kutcher-Burman reduction algorithm (i.e. the LKB model). It follows in the LKB model that uniform EUD irradiation of an OAR results in the same NTCP as the original non-homogeneous distribution. The NTCP equation is therefore represented as a function of EUD. The inverse equation expresses EUD as a function of NTCP and is used to generate a table of EUD versus normal tissue complication probability for the Emami-Burman parameter fits as well as for OAR parameter sets from more recent data 13. Statistical models based on conditional probability distributions International Nuclear Information System (INIS) Narayanan, R.S. 1991-10-01 We present a formulation of statistical mechanics models based on conditional probability distribution rather than a Hamiltonian. We show that it is possible to realize critical phenomena through this procedure. Closely linked with this formulation is a Monte Carlo algorithm, in which a configuration generated is guaranteed to be statistically independent from any other configuration for all values of the parameters, in particular near the critical point. (orig.) 14. Calculation of complication probability of pion treatment at PSI using dose-volume histograms International Nuclear Information System (INIS) Nakagawa, Keiichi; Akanuma, Atsuo; Aoki, Yukimasa 1991-01-01 15. Probability CERN Document Server Shiryaev, A N 1996-01-01 This book contains a systematic treatment of probability from the ground up, starting with intuitive ideas and gradually developing more sophisticated subjects, such as random walks, martingales, Markov chains, ergodic theory, weak convergence of probability measures, stationary stochastic processes, and the Kalman-Bucy filter Many examples are discussed in detail, and there are a large number of exercises The book is accessible to advanced undergraduates and can be used as a text for self-study This new edition contains substantial revisions and updated references The reader will find a deeper study of topics such as the distance between probability measures, metrization of weak convergence, and contiguity of probability measures Proofs for a number of some important results which were merely stated in the first edition have been added The author included new material on the probability of large deviations, and on the central limit theorem for sums of dependent random variables 16. Predicting Cumulative Incidence Probability: Marginal and Cause-Specific Modelling DEFF Research Database (Denmark) Scheike, Thomas H.; Zhang, Mei-Jie 2005-01-01 cumulative incidence probability; cause-specific hazards; subdistribution hazard; binomial modelling......cumulative incidence probability; cause-specific hazards; subdistribution hazard; binomial modelling... 17. Statistical physics of pairwise probability models Directory of Open Access Journals (Sweden) Yasser Roudi 2009-11-01 Full Text Available Statistical models for describing the probability distribution over the states of biological systems are commonly used for dimensional reduction. Among these models, pairwise models are very attractive in part because they can be fit using a reasonable amount of data: knowledge of the means and correlations between pairs of elements in the system is sufficient. Not surprisingly, then, using pairwise models for studying neural data has been the focus of many studies in recent years. In this paper, we describe how tools from statistical physics can be employed for studying and using pairwise models. We build on our previous work on the subject and study the relation between different methods for fitting these models and evaluating their quality. In particular, using data from simulated cortical networks we study how the quality of various approximate methods for inferring the parameters in a pairwise model depends on the time bin chosen for binning the data. We also study the effect of the size of the time bin on the model quality itself, again using simulated data. We show that using finer time bins increases the quality of the pairwise model. We offer new ways of deriving the expressions reported in our previous work for assessing the quality of pairwise models. 18. Traffic simulation based ship collision probability modeling Energy Technology Data Exchange (ETDEWEB) Goerlandt, Floris, E-mail: floris.goerlandt@tkk.f [Aalto University, School of Science and Technology, Department of Applied Mechanics, Marine Technology, P.O. Box 15300, FI-00076 AALTO, Espoo (Finland); Kujala, Pentti [Aalto University, School of Science and Technology, Department of Applied Mechanics, Marine Technology, P.O. Box 15300, FI-00076 AALTO, Espoo (Finland) 2011-01-15 Maritime traffic poses various risks in terms of human, environmental and economic loss. In a risk analysis of ship collisions, it is important to get a reasonable estimate for the probability of such accidents and the consequences they lead to. In this paper, a method is proposed to assess the probability of vessels colliding with each other. The method is capable of determining the expected number of accidents, the locations where and the time when they are most likely to occur, while providing input for models concerned with the expected consequences. At the basis of the collision detection algorithm lays an extensive time domain micro-simulation of vessel traffic in the given area. The Monte Carlo simulation technique is applied to obtain a meaningful prediction of the relevant factors of the collision events. Data obtained through the Automatic Identification System is analyzed in detail to obtain realistic input data for the traffic simulation: traffic routes, the number of vessels on each route, the ship departure times, main dimensions and sailing speed. The results obtained by the proposed method for the studied case of the Gulf of Finland are presented, showing reasonable agreement with registered accident and near-miss data. 19. Evaluation of nuclear power plant component failure probability and core damage probability using simplified PSA model International Nuclear Information System (INIS) 2000-01-01 It is anticipated that the change of frequency of surveillance tests, preventive maintenance or parts replacement of safety related components may cause the change of component failure probability and result in the change of core damage probability. It is also anticipated that the change is different depending on the initiating event frequency or the component types. This study assessed the change of core damage probability using simplified PSA model capable of calculating core damage probability in a short time period, which is developed by the US NRC to process accident sequence precursors, when various component's failure probability is changed between 0 and 1, or Japanese or American initiating event frequency data are used. As a result of the analysis, (1) It was clarified that frequency of surveillance test, preventive maintenance or parts replacement of motor driven pumps (high pressure injection pumps, residual heat removal pumps, auxiliary feedwater pumps) should be carefully changed, since the core damage probability's change is large, when the base failure probability changes toward increasing direction. (2) Core damage probability change is insensitive to surveillance test frequency change, since the core damage probability change is small, when motor operated valves and turbine driven auxiliary feed water pump failure probability changes around one figure. (3) Core damage probability change is small, when Japanese failure probability data are applied to emergency diesel generator, even if failure probability changes one figure from the base value. On the other hand, when American failure probability data is applied, core damage probability increase is large, even if failure probability changes toward increasing direction. Therefore, when Japanese failure probability data is applied, core damage probability change is insensitive to surveillance tests frequency change etc. (author) 20. Geometric modeling in probability and statistics CERN Document Server Calin, Ovidiu 2014-01-01 This book covers topics of Informational Geometry, a field which deals with the differential geometric study of the manifold probability density functions. This is a field that is increasingly attracting the interest of researchers from many different areas of science, including mathematics, statistics, geometry, computer science, signal processing, physics and neuroscience. It is the authors’ hope that the present book will be a valuable reference for researchers and graduate students in one of the aforementioned fields. This textbook is a unified presentation of differential geometry and probability theory, and constitutes a text for a course directed at graduate or advanced undergraduate students interested in applications of differential geometry in probability and statistics. The book contains over 100 proposed exercises meant to help students deepen their understanding, and it is accompanied by software that is able to provide numerical computations of several information geometric objects. The reader... 1. APPROXIMATION OF PROBABILITY DISTRIBUTIONS IN QUEUEING MODELS Directory of Open Access Journals (Sweden) T. I. Aliev 2013-03-01 Full Text Available For probability distributions with variation coefficient, not equal to unity, mathematical dependences for approximating distributions on the basis of first two moments are derived by making use of multi exponential distributions. It is proposed to approximate distributions with coefficient of variation less than unity by using hypoexponential distribution, which makes it possible to generate random variables with coefficient of variation, taking any value in a range (0; 1, as opposed to Erlang distribution, having only discrete values of coefficient of variation. 2. Calculating the Probability of Returning a Loan with Binary Probability Models Directory of Open Access Journals (Sweden) Julian Vasilev 2014-12-01 Full Text Available The purpose of this article is to give a new approach in calculating the probability of returning a loan. A lot of factors affect the value of the probability. In this article by using statistical and econometric models some influencing factors are proved. The main approach is concerned with applying probit and logit models in loan management institutions. A new aspect of the credit risk analysis is given. Calculating the probability of returning a loan is a difficult task. We assume that specific data fields concerning the contract (month of signing, year of signing, given sum and data fields concerning the borrower of the loan (month of birth, year of birth (age, gender, region, where he/she lives may be independent variables in a binary logistics model with a dependent variable “the probability of returning a loan”. It is proved that the month of signing a contract, the year of signing a contract, the gender and the age of the loan owner do not affect the probability of returning a loan. It is proved that the probability of returning a loan depends on the sum of contract, the remoteness of the loan owner and the month of birth. The probability of returning a loan increases with the increase of the given sum, decreases with the proximity of the customer, increases for people born in the beginning of the year and decreases for people born at the end of the year. 3. Classical probability model for Bell inequality International Nuclear Information System (INIS) Khrennikov, Andrei 2014-01-01 We show that by taking into account randomness of realization of experimental contexts it is possible to construct common Kolmogorov space for data collected for these contexts, although they can be incompatible. We call such a construction 'Kolmogorovization' of contextuality. This construction of common probability space is applied to Bell's inequality. It is well known that its violation is a consequence of collecting statistical data in a few incompatible experiments. In experiments performed in quantum optics contexts are determined by selections of pairs of angles (θ i ,θ ' j ) fixing orientations of polarization beam splitters. Opposite to the common opinion, we show that statistical data corresponding to measurements of polarizations of photons in the singlet state, e.g., in the form of correlations, can be described in the classical probabilistic framework. The crucial point is that in constructing the common probability space one has to take into account not only randomness of the source (as Bell did), but also randomness of context-realizations (in particular, realizations of pairs of angles (θ i , θ ' j )). One may (but need not) say that randomness of 'free will' has to be accounted for. 4. Stochastic population dynamic models as probability networks Science.gov (United States) M.E. and D.C. Lee. Borsuk 2009-01-01 The dynamics of a population and its response to environmental change depend on the balance of birth, death and age-at-maturity, and there have been many attempts to mathematically model populations based on these characteristics. Historically, most of these models were deterministic, meaning that the results were strictly determined by the equations of the model and... 5. Probability Modeling and Thinking: What Can We Learn from Practice? Science.gov (United States) Pfannkuch, Maxine; Budgett, Stephanie; Fewster, Rachel; Fitch, Marie; Pattenwise, Simeon; Wild, Chris; Ziedins, Ilze 2016-01-01 Because new learning technologies are enabling students to build and explore probability models, we believe that there is a need to determine the big enduring ideas that underpin probabilistic thinking and modeling. By uncovering the elements of the thinking modes of expert users of probability models we aim to provide a base for the setting of… 6. Normal Tissue Complication Probability Estimation by the Lyman-Kutcher-Burman Method Does Not Accurately Predict Spinal Cord Tolerance to Stereotactic Radiosurgery International Nuclear Information System (INIS) Daly, Megan E.; Luxton, Gary; Choi, Clara Y.H.; Gibbs, Iris C.; Chang, Steven D.; Adler, John R.; Soltys, Scott G. 2012-01-01 Purpose: To determine whether normal tissue complication probability (NTCP) analyses of the human spinal cord by use of the Lyman-Kutcher-Burman (LKB) model, supplemented by linear–quadratic modeling to account for the effect of fractionation, predict the risk of myelopathy from stereotactic radiosurgery (SRS). Methods and Materials: From November 2001 to July 2008, 24 spinal hemangioblastomas in 17 patients were treated with SRS. Of the tumors, 17 received 1 fraction with a median dose of 20 Gy (range, 18–30 Gy) and 7 received 20 to 25 Gy in 2 or 3 sessions, with cord maximum doses of 22.7 Gy (range, 17.8–30.9 Gy) and 22.0 Gy (range, 20.2–26.6 Gy), respectively. By use of conventional values for α/β, volume parameter n, 50% complication probability dose TD 50 , and inverse slope parameter m, a computationally simplified implementation of the LKB model was used to calculate the biologically equivalent uniform dose and NTCP for each treatment. Exploratory calculations were performed with alternate values of α/β and n. Results: In this study 1 case (4%) of myelopathy occurred. The LKB model using radiobiological parameters from Emami and the logistic model with parameters from Schultheiss overestimated complication rates, predicting 13 complications (54%) and 18 complications (75%), respectively. An increase in the volume parameter (n), to assume greater parallel organization, improved the predictive value of the models. Maximum-likelihood LKB fitting of α/β and n yielded better predictions (0.7 complications), with n = 0.023 and α/β = 17.8 Gy. Conclusions: The spinal cord tolerance to the dosimetry of SRS is higher than predicted by the LKB model using any set of accepted parameters. Only a high α/β value in the LKB model and only a large volume effect in the logistic model with Schultheiss data could explain the low number of complications observed. This finding emphasizes that radiobiological models traditionally used to estimate spinal cord NTCP 7. Model checking meets probability: a gentle introduction NARCIS (Netherlands) Katoen, Joost P. 2013-01-01 This paper considers fully probabilistic system models. Each transition is quantified with a probability—its likelihood of occurrence. Properties are expressed as automata that either accept or reject system runs. The central question is to determine the fraction of accepted system runs. We also 8. Idealized models of the joint probability distribution of wind speeds Science.gov (United States) 2018-05-01 The joint probability distribution of wind speeds at two separate locations in space or points in time completely characterizes the statistical dependence of these two quantities, providing more information than linear measures such as correlation. In this study, we consider two models of the joint distribution of wind speeds obtained from idealized models of the dependence structure of the horizontal wind velocity components. The bivariate Rice distribution follows from assuming that the wind components have Gaussian and isotropic fluctuations. The bivariate Weibull distribution arises from power law transformations of wind speeds corresponding to vector components with Gaussian, isotropic, mean-zero variability. Maximum likelihood estimates of these distributions are compared using wind speed data from the mid-troposphere, from different altitudes at the Cabauw tower in the Netherlands, and from scatterometer observations over the sea surface. While the bivariate Rice distribution is more flexible and can represent a broader class of dependence structures, the bivariate Weibull distribution is mathematically simpler and may be more convenient in many applications. The complexity of the mathematical expressions obtained for the joint distributions suggests that the development of explicit functional forms for multivariate speed distributions from distributions of the components will not be practical for more complicated dependence structure or more than two speed variables. 9. Some simple applications of probability models to birth intervals International Nuclear Information System (INIS) Shrestha, G. 1987-07-01 An attempt has been made in this paper to apply some simple probability models to birth intervals under the assumption of constant fecundability and varying fecundability among women. The parameters of the probability models are estimated by using the method of moments and the method of maximum likelihood. (author). 9 refs, 2 tabs 10. Impact of target point deviations on control and complication probabilities in stereotactic radiosurgery of AVMs and metastases International Nuclear Information System (INIS) Treuer, Harald; Kocher, Martin; Hoevels, Moritz; Hunsche, Stefan; Luyken, Klaus; Maarouf, Mohammad; Voges, Juergen; Mueller, Rolf-Peter; Sturm, Volker 2006-01-01 Objective: Determination of the impact of inaccuracies in the determination and setup of the target point in stereotactic radiosurgery (SRS) on the expectable complication and control probabilities. Methods: Two randomized samples of patients with arteriovenous malformation (AVM) (n = 20) and with brain metastases (n = 20) treated with SRS were formed, and the probability for complete obliteration (COP) or complete remission (CRP), the size of the 10 Gy-volume in the brain tissue (VOI10), and the probability for radiation necrosis (NTCP) were calculated. The dose-effect relations for COP and CRP were fitted to clinical data. Target point deviations were simulated through random vectors and the resulting probabilities and volumes were calculated and compared with the values of the treatment plan. Results: The decrease of the relative value of the control probabilities at 1 mm target point deviation was up to 4% for AVMs and up to 10% for metastases. At 2 mm the median decrease was 5% for AVMs and 9% for metastases. The value for the target point deviation, at which COP and CRP decreased about 0.05 in 90% of the cases, was 1.3 mm. The increase of NTCP was maximally 0.0025 per mm target point deviation for AVMs and 0.0035/mm for metastases. The maximal increase of VOI10 was 0.7 cm 3 /mm target point deviation in both patient groups. Conclusions: The upper limit for tolerable target point deviations is at 1.3 mm. If this value cannot be achieved during the system test, a supplementary safety margin should be applied for the definition of the target volume. A better accuracy level is desirable, in order to ensure optimal chances for the success of the treatment. The target point precision is less important for the minimization of the probability of radiation necroses 11. Probability model for analyzing fire management alternatives: theory and structure Science.gov (United States) Frederick W. Bratten 1982-01-01 A theoretical probability model has been developed for analyzing program alternatives in fire management. It includes submodels or modules for predicting probabilities of fire behavior, fire occurrence, fire suppression, effects of fire on land resources, and financial effects of fire. Generalized "fire management situations" are used to represent actual fire... 12. Convergence of Transition Probability Matrix in CLVMarkov Models Science.gov (United States) Permana, D.; Pasaribu, U. S.; Indratno, S. W.; Suprayogi, S. 2018-04-01 A transition probability matrix is an arrangement of transition probability from one states to another in a Markov chain model (MCM). One of interesting study on the MCM is its behavior for a long time in the future. The behavior is derived from one property of transition probabilty matrix for n steps. This term is called the convergence of the n-step transition matrix for n move to infinity. Mathematically, the convergence of the transition probability matrix is finding the limit of the transition matrix which is powered by n where n moves to infinity. The convergence form of the transition probability matrix is very interesting as it will bring the matrix to its stationary form. This form is useful for predicting the probability of transitions between states in the future. The method usually used to find the convergence of transition probability matrix is through the process of limiting the distribution. In this paper, the convergence of the transition probability matrix is searched using a simple concept of linear algebra that is by diagonalizing the matrix.This method has a higher level of complexity because it has to perform the process of diagonalization in its matrix. But this way has the advantage of obtaining a common form of power n of the transition probability matrix. This form is useful to see transition matrix before stationary. For example cases are taken from CLV model using MCM called Model of CLV-Markov. There are several models taken by its transition probability matrix to find its convergence form. The result is that the convergence of the matrix of transition probability through diagonalization has similarity with convergence with commonly used distribution of probability limiting method. 13. On the structure of the quantum-mechanical probability models International Nuclear Information System (INIS) Cufaro-Petroni, N. 1992-01-01 In this paper the role of the mathematical probability models in the classical and quantum physics in shortly analyzed. In particular the formal structure of the quantum probability spaces (QPS) is contrasted with the usual Kolmogorovian models of probability by putting in evidence the connections between this structure and the fundamental principles of the quantum mechanics. The fact that there is no unique Kolmogorovian model reproducing a QPS is recognized as one of the main reasons of the paradoxical behaviors pointed out in the quantum theory from its early days. 8 refs 14. Modeling the probability distribution of peak discharge for infiltrating hillslopes Science.gov (United States) Baiamonte, Giorgio; Singh, Vijay P. 2017-07-01 Hillslope response plays a fundamental role in the prediction of peak discharge at the basin outlet. The peak discharge for the critical duration of rainfall and its probability distribution are needed for designing urban infrastructure facilities. This study derives the probability distribution, denoted as GABS model, by coupling three models: (1) the Green-Ampt model for computing infiltration, (2) the kinematic wave model for computing discharge hydrograph from the hillslope, and (3) the intensity-duration-frequency (IDF) model for computing design rainfall intensity. The Hortonian mechanism for runoff generation is employed for computing the surface runoff hydrograph. Since the antecedent soil moisture condition (ASMC) significantly affects the rate of infiltration, its effect on the probability distribution of peak discharge is investigated. Application to a watershed in Sicily, Italy, shows that with the increase of probability, the expected effect of ASMC to increase the maximum discharge diminishes. Only for low values of probability, the critical duration of rainfall is influenced by ASMC, whereas its effect on the peak discharge seems to be less for any probability. For a set of parameters, the derived probability distribution of peak discharge seems to be fitted by the gamma distribution well. Finally, an application to a small watershed, with the aim to test the possibility to arrange in advance the rational runoff coefficient tables to be used for the rational method, and a comparison between peak discharges obtained by the GABS model with those measured in an experimental flume for a loamy-sand soil were carried out. 15. A new plan-scoring method using normal tissue complication probability for personalized treatment plan decisions in prostate cancer Science.gov (United States) Kim, Kwang Hyeon; Lee, Suk; Shim, Jang Bo; Yang, Dae Sik; Yoon, Won Sup; Park, Young Je; Kim, Chul Yong; Cao, Yuan Jie; Chang, Kyung Hwan 2018-01-01 The aim of this study was to derive a new plan-scoring index using normal tissue complication probabilities to verify different plans in the selection of personalized treatment. Plans for 12 patients treated with tomotherapy were used to compare scoring for ranking. Dosimetric and biological indexes were analyzed for the plans for a clearly distinguishable group ( n = 7) and a similar group ( n = 12), using treatment plan verification software that we developed. The quality factor ( QF) of our support software for treatment decisions was consistent with the final treatment plan for the clearly distinguishable group (average QF = 1.202, 100% match rate, n = 7) and the similar group (average QF = 1.058, 33% match rate, n = 12). Therefore, we propose a normal tissue complication probability (NTCP) based on the plan scoring index for verification of different plans for personalized treatment-plan selection. Scoring using the new QF showed a 100% match rate (average NTCP QF = 1.0420). The NTCP-based new QF scoring method was adequate for obtaining biological verification quality and organ risk saving using the treatment-planning decision-support software we developed for prostate cancer. 16. Review of Literature for Model Assisted Probability of Detection Energy Technology Data Exchange (ETDEWEB) Meyer, Ryan M. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Crawford, Susan L. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Lareau, John P. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Anderson, Michael T. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States) 2014-09-30 This is a draft technical letter report for NRC client documenting a literature review of model assisted probability of detection (MAPOD) for potential application to nuclear power plant components for improvement of field NDE performance estimations. 17. Chemo-IMRT of Oropharyngeal Cancer Aiming to Reduce Dysphagia: Swallowing Organs Late Complication Probabilities and Dosimetric Correlates International Nuclear Information System (INIS) Eisbruch, Avraham; Kim, Hyungjin M.; Feng, Felix Y.; Lyden, Teresa H.; Haxer, Marc J.; Feng, Mary; Worden, Frank P.; Bradford, Carol R.; Prince, Mark E.; Moyer, Jeffrey S.; Wolf, Gregory T.; Chepeha, Douglas B.; Ten Haken, Randall K. 2011-01-01 Purpose: Assess dosimetric correlates of long-term dysphagia after chemo-intensity-modulated radiotherapy (IMRT) of oropharyngeal cancer (OPC) sparing parts of the swallowing organs. Patients and Methods: Prospective longitudinal study: weekly chemotherapy concurrent with IMRT for Stages III/IV OPC, aiming to reduce dysphagia by sparing noninvolved parts of swallowing-related organs: pharyngeal constrictors (PC), glottic and supraglottic larynx (GSL), and esophagus, as well as oral cavity and major salivary glands. Dysphagia outcomes included patient-reported Swallowing and Eating Domain scores, Observer-based (CTCAEv.2) dysphagia, and videofluoroscopy (VF), before and periodically after therapy through 2 years. Relationships between dosimetric factors and worsening (from baseline) of dysphagia through 2 years were assessed by linear mixed-effects model. Results: Seventy-three patients participated. Observer-based dysphagia was not modeled because at >6 months there were only four Grade ≥2 cases (one of whom was feeding-tube dependent). PC, GSL, and esophagus mean doses, as well as their partial volume doses (V D s), were each significantly correlated with all dysphagia outcomes. However, the V D s for each organ intercorrelated and also highly correlated with the mean doses, leaving only mean doses significant. Mean doses to each of the parts of the PCs (superior, middle, and inferior) were also significantly correlated with all dysphagia measures, with superior PCs demonstrating highest correlations. For VF-based strictures, most significant predictor was esophageal mean doses (48±17 Gy in patients with, vs 27±12 in patients without strictures, p = 0.004). Normal tissue complication probabilities (NTCPs) increased moderately with mean doses without any threshold. For increased VF-based aspirations or worsened VF summary scores, toxic doses (TDs) 50 and TD 25 were 63 Gy and 56 Gy for PC, and 56 Gy and 39 Gy for GSL, respectively. For both PC and GSL, patient 18. Models for probability and statistical inference theory and applications CERN Document Server Stapleton, James H 2007-01-01 This concise, yet thorough, book is enhanced with simulations and graphs to build the intuition of readersModels for Probability and Statistical Inference was written over a five-year period and serves as a comprehensive treatment of the fundamentals of probability and statistical inference. With detailed theoretical coverage found throughout the book, readers acquire the fundamentals needed to advance to more specialized topics, such as sampling, linear models, design of experiments, statistical computing, survival analysis, and bootstrapping.Ideal as a textbook for a two-semester sequence on probability and statistical inference, early chapters provide coverage on probability and include discussions of: discrete models and random variables; discrete distributions including binomial, hypergeometric, geometric, and Poisson; continuous, normal, gamma, and conditional distributions; and limit theory. Since limit theory is usually the most difficult topic for readers to master, the author thoroughly discusses mo... 19. Methods for Reducing Normal Tissue Complication Probabilities in Oropharyngeal Cancer: Dose Reduction or Planning Target Volume Elimination Energy Technology Data Exchange (ETDEWEB) Samuels, Stuart E.; Eisbruch, Avraham; Vineberg, Karen; Lee, Jae; Lee, Choonik; Matuszak, Martha M.; Ten Haken, Randall K.; Brock, Kristy K., E-mail: kbrock@med.umich.edu 2016-11-01 Purpose: Strategies to reduce the toxicities of head and neck radiation (ie, dysphagia [difficulty swallowing] and xerostomia [dry mouth]) are currently underway. However, the predicted benefit of dose and planning target volume (PTV) reduction strategies is unknown. The purpose of the present study was to compare the normal tissue complication probabilities (NTCP) for swallowing and salivary structures in standard plans (70 Gy [P70]), dose-reduced plans (60 Gy [P60]), and plans eliminating the PTV margin. Methods and Materials: A total of 38 oropharyngeal cancer (OPC) plans were analyzed. Standard organ-sparing volumetric modulated arc therapy plans (P70) were created and then modified by eliminating the PTVs and treating the clinical tumor volumes (CTVs) only (C70) or maintaining the PTV but reducing the dose to 60 Gy (P60). NTCP dose models for the pharyngeal constrictors, glottis/supraglottic larynx, parotid glands (PGs), and submandibular glands (SMGs) were analyzed. The minimal clinically important benefit was defined as a mean change in NTCP of >5%. The P70 NTCP thresholds and overlap percentages of the organs at risk with the PTVs (56-59 Gy, vPTV{sub 56}) were evaluated to identify the predictors for NTCP improvement. Results: With the P60 plans, only the ipsilateral PG (iPG) benefited (23.9% vs 16.2%; P<.01). With the C70 plans, only the iPG (23.9% vs 17.5%; P<.01) and contralateral SMG (cSMG) (NTCP 32.1% vs 22.9%; P<.01) benefited. An iPG NTCP threshold of 20% and 30% predicted NTCP benefits for the P60 and C70 plans, respectively (P<.001). A cSMG NTCP threshold of 30% predicted for an NTCP benefit with the C70 plans (P<.001). Furthermore, for the iPG, a vPTV{sub 56} >13% predicted benefit with P60 (P<.001) and C70 (P=.002). For the cSMG, a vPTV{sub 56} >22% predicted benefit with C70 (P<.01). Conclusions: PTV elimination and dose-reduction lowered the NTCP of the iPG, and PTV elimination lowered the NTCP of the cSMG. NTCP thresholds and the 20. Continuum shell-model with complicated configurations International Nuclear Information System (INIS) Barz, H.W.; Hoehn, J. 1977-05-01 The traditional shell model has been combined with the coupled channels method in order to describe resonance reactions. For that purpose the configuration space is divided into two subspaces (Feshbach projection method). Complicated shell-model configurations can be included into the subspace of discrete states which contains the single particle resonance states too. In the subspace of scattering states the equation of motion is solved by using the coupled channels method. Thereby the orthogonality between scattering states and discrete states is ensured. Resonance states are defined with outgoing waves in all channels. By means of simple model calculations the special role of the continuum is investigated. In this connection the energy dependence of the resonance parameters, the isospin mixture via the continuum, threshold effect, as well as the influence of the number of channels taken into account on the widths, positions and dipole strengths of the resonance are discussed. The model is mainly applied to the description of giant resonances excited by the scattering of nucleons and photo-nucleus processes (source term method) found in reactions on light nuclei. The giant resonance observed in the 15 N(p,n) reaction is explained by the inclusion of 2p-2h states. The same is true for the giant resonance in 13 C(J = 1/2, 3/2) as well as for the giant resonance built on the first 3 - state in 16 O. By means of a correlation analysis for the reduced widths amplitudes an access to the doorway conception is found. (author) 1. A probability model for the failure of pressure containing parts International Nuclear Information System (INIS) Thomas, H.M. 1978-01-01 The model provides a method of estimating the order of magnitude of the leakage failure probability of pressure containing parts. It is a fatigue based model which makes use of the statistics available for both specimens and vessels. Some novel concepts are introduced but essentially the model simply quantifies the obvious i.e. that failure probability increases with increases in stress levels, number of cycles, volume of material and volume of weld metal. A further model based on fracture mechanics estimates the catastrophic fraction of leakage failures. (author) 2. Maximum parsimony, substitution model, and probability phylogenetic trees. Science.gov (United States) Weng, J F; Thomas, D A; Mareels, I 2011-01-01 The problem of inferring phylogenies (phylogenetic trees) is one of the main problems in computational biology. There are three main methods for inferring phylogenies-Maximum Parsimony (MP), Distance Matrix (DM) and Maximum Likelihood (ML), of which the MP method is the most well-studied and popular method. In the MP method the optimization criterion is the number of substitutions of the nucleotides computed by the differences in the investigated nucleotide sequences. However, the MP method is often criticized as it only counts the substitutions observable at the current time and all the unobservable substitutions that really occur in the evolutionary history are omitted. In order to take into account the unobservable substitutions, some substitution models have been established and they are now widely used in the DM and ML methods but these substitution models cannot be used within the classical MP method. Recently the authors proposed a probability representation model for phylogenetic trees and the reconstructed trees in this model are called probability phylogenetic trees. One of the advantages of the probability representation model is that it can include a substitution model to infer phylogenetic trees based on the MP principle. In this paper we explain how to use a substitution model in the reconstruction of probability phylogenetic trees and show the advantage of this approach with examples. 3. Gap probability - Measurements and models of a pecan orchard Science.gov (United States) Strahler, Alan H.; Li, Xiaowen; Moody, Aaron; Liu, YI 1992-01-01 Measurements and models are compared for gap probability in a pecan orchard. Measurements are based on panoramic photographs of 50* by 135 view angle made under the canopy looking upwards at regular positions along transects between orchard trees. The gap probability model is driven by geometric parameters at two levels-crown and leaf. Crown level parameters include the shape of the crown envelope and spacing of crowns; leaf level parameters include leaf size and shape, leaf area index, and leaf angle, all as functions of canopy position. 4. Illustrating Probability through Roulette: A Spreadsheet Simulation Model Directory of Open Access Journals (Sweden) Kala Chand Seal 2005-11-01 Full Text Available Teaching probability can be challenging because the mathematical formulas often are too abstract and complex for the students to fully grasp the underlying meaning and effect of the concepts. Games can provide a way to address this issue. For example, the game of roulette can be an exciting application for teaching probability concepts. In this paper, we implement a model of roulette in a spreadsheet that can simulate outcomes of various betting strategies. The simulations can be analyzed to gain better insights into the corresponding probability structures. We use the model to simulate a particular betting strategy known as the bet-doubling, or Martingale, strategy. This strategy is quite popular and is often erroneously perceived as a winning strategy even though the probability analysis shows that such a perception is incorrect. The simulation allows us to present the true implications of such a strategy for a player with a limited betting budget and relate the results to the underlying theoretical probability structure. The overall validation of the model, its use for teaching, including its application to analyze other types of betting strategies are discussed. 5. Reach/frequency for printed media: Personal probabilities or models DEFF Research Database (Denmark) Mortensen, Peter Stendahl 2000-01-01 The author evaluates two different ways of estimating reach and frequency of plans for printed media. The first assigns reading probabilities to groups of respondents and calculates reach and frequency by simulation. the second estimates parameters to a model for reach/frequency. It is concluded ...... and estiamtes from such models are shown to be closer to panel data. the problem, however, is to get valid input for such models from readership surveys. Means for this are discussed.... 6. Camera-Model Identification Using Markovian Transition Probability Matrix Science.gov (United States) Xu, Guanshuo; Gao, Shang; Shi, Yun Qing; Hu, Ruimin; Su, Wei Detecting the (brands and) models of digital cameras from given digital images has become a popular research topic in the field of digital forensics. As most of images are JPEG compressed before they are output from cameras, we propose to use an effective image statistical model to characterize the difference JPEG 2-D arrays of Y and Cb components from the JPEG images taken by various camera models. Specifically, the transition probability matrices derived from four different directional Markov processes applied to the image difference JPEG 2-D arrays are used to identify statistical difference caused by image formation pipelines inside different camera models. All elements of the transition probability matrices, after a thresholding technique, are directly used as features for classification purpose. Multi-class support vector machines (SVM) are used as the classification tool. The effectiveness of our proposed statistical model is demonstrated by large-scale experimental results. 7. Multiple-event probability in general-relativistic quantum mechanics. II. A discrete model International Nuclear Information System (INIS) Mondragon, Mauricio; Perez, Alejandro; Rovelli, Carlo 2007-01-01 We introduce a simple quantum mechanical model in which time and space are discrete and periodic. These features avoid the complications related to continuous-spectrum operators and infinite-norm states. The model provides a tool for discussing the probabilistic interpretation of generally covariant quantum systems, without the confusion generated by spurious infinities. We use the model to illustrate the formalism of general-relativistic quantum mechanics, and to test the definition of multiple-event probability introduced in a companion paper [Phys. Rev. D 75, 084033 (2007)]. We consider a version of the model with unitary time evolution and a version without unitary time evolution 8. Modeling highway travel time distribution with conditional probability models Energy Technology Data Exchange (ETDEWEB) Oliveira Neto, Francisco Moraes [ORNL; Chin, Shih-Miao [ORNL; Hwang, Ho-Ling [ORNL; Han, Lee [University of Tennessee, Knoxville (UTK) 2014-01-01 ABSTRACT Under the sponsorship of the Federal Highway Administration's Office of Freight Management and Operations, the American Transportation Research Institute (ATRI) has developed performance measures through the Freight Performance Measures (FPM) initiative. Under this program, travel speed information is derived from data collected using wireless based global positioning systems. These telemetric data systems are subscribed and used by trucking industry as an operations management tool. More than one telemetric operator submits their data dumps to ATRI on a regular basis. Each data transmission contains truck location, its travel time, and a clock time/date stamp. Data from the FPM program provides a unique opportunity for studying the upstream-downstream speed distributions at different locations, as well as different time of the day and day of the week. This research is focused on the stochastic nature of successive link travel speed data on the continental United States Interstates network. Specifically, a method to estimate route probability distributions of travel time is proposed. This method uses the concepts of convolution of probability distributions and bivariate, link-to-link, conditional probability to estimate the expected distributions for the route travel time. Major contribution of this study is the consideration of speed correlation between upstream and downstream contiguous Interstate segments through conditional probability. The established conditional probability distributions, between successive segments, can be used to provide travel time reliability measures. This study also suggests an adaptive method for calculating and updating route travel time distribution as new data or information is added. This methodology can be useful to estimate performance measures as required by the recent Moving Ahead for Progress in the 21st Century Act (MAP 21). 9. Computation of Probabilities in Causal Models of History of Science Directory of Open Access Journals (Sweden) Osvaldo Pessoa Jr. 2006-12-01 Full Text Available : The aim of this paper is to investigate the ascription of probabilities in a causal model of an episode in the history of science. The aim of such a quantitative approach is to allow the implementation of the causal model in a computer, to run simulations. As an example, we look at the beginning of the science of magnetism, “explaining” — in a probabilistic way, in terms of a single causal model — why the field advanced in China but not in Europe (the difference is due to different prior probabilities of certain cultural manifestations. Given the number of years between the occurrences of two causally connected advances X and Y, one proposes a criterion for stipulating the value pY=X of the conditional probability of an advance Y occurring, given X. Next, one must assume a specific form for the cumulative probability function pY=X(t, which we take to be the time integral of an exponential distribution function, as is done in physics of radioactive decay. Rules for calculating the cumulative functions for more than two events are mentioned, involving composition, disjunction and conjunction of causes. We also consider the problems involved in supposing that the appearance of events in time follows an exponential distribution, which are a consequence of the fact that a composition of causes does not follow an exponential distribution, but a “hypoexponential” one. We suggest that a gamma distribution function might more adequately represent the appearance of advances. 10. Fixation probability in a two-locus intersexual selection model. Science.gov (United States) Durand, Guillermo; Lessard, Sabin 2016-06-01 We study a two-locus model of intersexual selection in a finite haploid population reproducing according to a discrete-time Moran model with a trait locus expressed in males and a preference locus expressed in females. We show that the probability of ultimate fixation of a single mutant allele for a male ornament introduced at random at the trait locus given any initial frequency state at the preference locus is increased by weak intersexual selection and recombination, weak or strong. Moreover, this probability exceeds the initial frequency of the mutant allele even in the case of a costly male ornament if intersexual selection is not too weak. On the other hand, the probability of ultimate fixation of a single mutant allele for a female preference towards a male ornament introduced at random at the preference locus is increased by weak intersexual selection and weak recombination if the female preference is not costly, and is strong enough in the case of a costly male ornament. The analysis relies on an extension of the ancestral recombination-selection graph for samples of haplotypes to take into account events of intersexual selection, while the symbolic calculation of the fixation probabilities is made possible in a reasonable time by an optimizing algorithm. Copyright © 2016 Elsevier Inc. All rights reserved. 11. Estimating Model Probabilities using Thermodynamic Markov Chain Monte Carlo Methods Science.gov (United States) Ye, M.; Liu, P.; Beerli, P.; Lu, D.; Hill, M. C. 2014-12-01 Markov chain Monte Carlo (MCMC) methods are widely used to evaluate model probability for quantifying model uncertainty. In a general procedure, MCMC simulations are first conducted for each individual model, and MCMC parameter samples are then used to approximate marginal likelihood of the model by calculating the geometric mean of the joint likelihood of the model and its parameters. It has been found the method of evaluating geometric mean suffers from the numerical problem of low convergence rate. A simple test case shows that even millions of MCMC samples are insufficient to yield accurate estimation of the marginal likelihood. To resolve this problem, a thermodynamic method is used to have multiple MCMC runs with different values of a heating coefficient between zero and one. When the heating coefficient is zero, the MCMC run is equivalent to a random walk MC in the prior parameter space; when the heating coefficient is one, the MCMC run is the conventional one. For a simple case with analytical form of the marginal likelihood, the thermodynamic method yields more accurate estimate than the method of using geometric mean. This is also demonstrated for a case of groundwater modeling with consideration of four alternative models postulated based on different conceptualization of a confining layer. This groundwater example shows that model probabilities estimated using the thermodynamic method are more reasonable than those obtained using the geometric method. The thermodynamic method is general, and can be used for a wide range of environmental problem for model uncertainty quantification. 12. Analysis of clinical complication data for radiation hepatitis using a parallel architecture model International Nuclear Information System (INIS) Jackson, A.; Haken, R.K. ten; Robertson, J.M.; Kessler, M.L.; Kutcher, G.J.; Lawrence, T.S. 1995-01-01 Purpose: The detailed knowledge of dose volume distributions available from the three-dimensional (3D) conformal radiation treatment of tumors in the liver (reported elsewhere) offers new opportunities to quantify the effect of volume on the probability of producing radiation hepatitis. We aim to test a new parallel architecture model of normal tissue complication probability (NTCP) with these data. Methods and Materials: Complication data and dose volume histograms from a total of 93 patients with normal liver function, treated on a prospective protocol with 3D conformal radiation therapy and intraarterial hepatic fluorodeoxyuridine, were analyzed with a new parallel architecture model. Patient treatment fell into six categories differing in doses delivered and volumes irradiated. By modeling the radiosensitivity of liver subunits, we are able to use dose volume histograms to calculate the fraction of the liver damaged in each patient. A complication results if this fraction exceeds the patient's functional reserve. To determine the patient distribution of functional reserves and the subunit radiosensitivity, the maximum likelihood method was used to fit the observed complication data. Results: The parallel model fit the complication data well, although uncertainties on the functional reserve distribution and subunit radiosensitivy are highly correlated. Conclusion: The observed radiation hepatitis complications show a threshold effect that can be described well with a parallel architecture model. However, additional independent studies are required to better determine the parameters defining the functional reserve distribution and subunit radiosensitivity 13. Assigning probability distributions to input parameters of performance assessment models Energy Technology Data Exchange (ETDEWEB) Mishra, Srikanta [INTERA Inc., Austin, TX (United States) 2002-02-01 This study presents an overview of various approaches for assigning probability distributions to input parameters and/or future states of performance assessment models. Specifically,three broad approaches are discussed for developing input distributions: (a) fitting continuous distributions to data, (b) subjective assessment of probabilities, and (c) Bayesian updating of prior knowledge based on new information. The report begins with a summary of the nature of data and distributions, followed by a discussion of several common theoretical parametric models for characterizing distributions. Next, various techniques are presented for fitting continuous distributions to data. These include probability plotting, method of moments, maximum likelihood estimation and nonlinear least squares analysis. The techniques are demonstrated using data from a recent performance assessment study for the Yucca Mountain project. Goodness of fit techniques are also discussed, followed by an overview of how distribution fitting is accomplished in commercial software packages. The issue of subjective assessment of probabilities is dealt with in terms of the maximum entropy distribution selection approach, as well as some common rules for codifying informal expert judgment. Formal expert elicitation protocols are discussed next, and are based primarily on the guidance provided by the US NRC. The Bayesian framework for updating prior distributions (beliefs) when new information becomes available is discussed. A simple numerical approach is presented for facilitating practical applications of the Bayes theorem. Finally, a systematic framework for assigning distributions is presented: (a) for the situation where enough data are available to define an empirical CDF or fit a parametric model to the data, and (b) to deal with the situation where only a limited amount of information is available. 14. Assigning probability distributions to input parameters of performance assessment models International Nuclear Information System (INIS) Mishra, Srikanta 2002-02-01 This study presents an overview of various approaches for assigning probability distributions to input parameters and/or future states of performance assessment models. Specifically,three broad approaches are discussed for developing input distributions: (a) fitting continuous distributions to data, (b) subjective assessment of probabilities, and (c) Bayesian updating of prior knowledge based on new information. The report begins with a summary of the nature of data and distributions, followed by a discussion of several common theoretical parametric models for characterizing distributions. Next, various techniques are presented for fitting continuous distributions to data. These include probability plotting, method of moments, maximum likelihood estimation and nonlinear least squares analysis. The techniques are demonstrated using data from a recent performance assessment study for the Yucca Mountain project. Goodness of fit techniques are also discussed, followed by an overview of how distribution fitting is accomplished in commercial software packages. The issue of subjective assessment of probabilities is dealt with in terms of the maximum entropy distribution selection approach, as well as some common rules for codifying informal expert judgment. Formal expert elicitation protocols are discussed next, and are based primarily on the guidance provided by the US NRC. The Bayesian framework for updating prior distributions (beliefs) when new information becomes available is discussed. A simple numerical approach is presented for facilitating practical applications of the Bayes theorem. Finally, a systematic framework for assigning distributions is presented: (a) for the situation where enough data are available to define an empirical CDF or fit a parametric model to the data, and (b) to deal with the situation where only a limited amount of information is available 15. Fitting the Probability Distribution Functions to Model Particulate Matter Concentrations International Nuclear Information System (INIS) El-Shanshoury, Gh.I. 2017-01-01 The main objective of this study is to identify the best probability distribution and the plotting position formula for modeling the concentrations of Total Suspended Particles (TSP) as well as the Particulate Matter with an aerodynamic diameter<10 μm (PM 10 ). The best distribution provides the estimated probabilities that exceed the threshold limit given by the Egyptian Air Quality Limit value (EAQLV) as well the number of exceedance days is estimated. The standard limits of the EAQLV for TSP and PM 10 concentrations are 24-h average of 230 μg/m 3 and 70 μg/m 3 , respectively. Five frequency distribution functions with seven formula of plotting positions (empirical cumulative distribution functions) are compared to fit the average of daily TSP and PM 10 concentrations in year 2014 for Ain Sokhna city. The Quantile-Quantile plot (Q-Q plot) is used as a method for assessing how closely a data set fits a particular distribution. A proper probability distribution that represents the TSP and PM 10 has been chosen based on the statistical performance indicator values. The results show that Hosking and Wallis plotting position combined with Frechet distribution gave the highest fit for TSP and PM 10 concentrations. Burr distribution with the same plotting position follows Frechet distribution. The exceedance probability and days over the EAQLV are predicted using Frechet distribution. In 2014, the exceedance probability and days for TSP concentrations are 0.052 and 19 days, respectively. Furthermore, the PM 10 concentration is found to exceed the threshold limit by 174 days 16. An empirical probability model of detecting species at low densities. Science.gov (United States) Delaney, David G; Leung, Brian 2010-06-01 False negatives, not detecting things that are actually present, are an important but understudied problem. False negatives are the result of our inability to perfectly detect species, especially those at low density such as endangered species or newly arriving introduced species. They reduce our ability to interpret presence-absence survey data and make sound management decisions (e.g., rapid response). To reduce the probability of false negatives, we need to compare the efficacy and sensitivity of different sampling approaches and quantify an unbiased estimate of the probability of detection. We conducted field experiments in the intertidal zone of New England and New York to test the sensitivity of two sampling approaches (quadrat vs. total area search, TAS), given different target characteristics (mobile vs. sessile). Using logistic regression we built detection curves for each sampling approach that related the sampling intensity and the density of targets to the probability of detection. The TAS approach reduced the probability of false negatives and detected targets faster than the quadrat approach. Mobility of targets increased the time to detection but did not affect detection success. Finally, we interpreted two years of presence-absence data on the distribution of the Asian shore crab (Hemigrapsus sanguineus) in New England and New York, using our probability model for false negatives. The type of experimental approach in this paper can help to reduce false negatives and increase our ability to detect species at low densities by refining sampling approaches, which can guide conservation strategies and management decisions in various areas of ecology such as conservation biology and invasion ecology. 17. Human Inferences about Sequences: A Minimal Transition Probability Model. Directory of Open Access Journals (Sweden) Florent Meyniel 2016-12-01 Full Text Available The brain constantly infers the causes of the inputs it receives and uses these inferences to generate statistical expectations about future observations. Experimental evidence for these expectations and their violations include explicit reports, sequential effects on reaction times, and mismatch or surprise signals recorded in electrophysiology and functional MRI. Here, we explore the hypothesis that the brain acts as a near-optimal inference device that constantly attempts to infer the time-varying matrix of transition probabilities between the stimuli it receives, even when those stimuli are in fact fully unpredictable. This parsimonious Bayesian model, with a single free parameter, accounts for a broad range of findings on surprise signals, sequential effects and the perception of randomness. Notably, it explains the pervasive asymmetry between repetitions and alternations encountered in those studies. Our analysis suggests that a neural machinery for inferring transition probabilities lies at the core of human sequence knowledge. 18. NASA Lewis Launch Collision Probability Model Developed and Analyzed Science.gov (United States) Bollenbacher, Gary; Guptill, James D 1999-01-01 There are nearly 10,000 tracked objects orbiting the earth. These objects encompass manned objects, active and decommissioned satellites, spent rocket bodies, and debris. They range from a few centimeters across to the size of the MIR space station. Anytime a new satellite is launched, the launch vehicle with its payload attached passes through an area of space in which these objects orbit. Although the population density of these objects is low, there always is a small but finite probability of collision between the launch vehicle and one or more of these space objects. Even though the probability of collision is very low, for some payloads even this small risk is unacceptable. To mitigate the small risk of collision associated with launching at an arbitrary time within the daily launch window, NASA performs a prelaunch mission assurance Collision Avoidance Analysis (or COLA). For the COLA of the Cassini spacecraft, the NASA Lewis Research Center conducted an in-house development and analysis of a model for launch collision probability. The model allows a minimum clearance criteria to be used with the COLA analysis to ensure an acceptably low probability of collision. If, for any given liftoff time, the nominal launch vehicle trajectory would pass a space object with less than the minimum required clearance, launch would not be attempted at that time. The model assumes that the nominal positions of the orbiting objects and of the launch vehicle can be predicted as a function of time, and therefore, that any tracked object that comes within close proximity of the launch vehicle can be identified. For any such pair, these nominal positions can be used to calculate a nominal miss distance. The actual miss distances may differ substantially from the nominal miss distance, due, in part, to the statistical uncertainty of the knowledge of the objects positions. The model further assumes that these position uncertainties can be described with position covariance matrices 19. Probability Model for Data Redundancy Detection in Sensor Networks Directory of Open Access Journals (Sweden) Suman Kumar 2009-01-01 Full Text Available Sensor networks are made of autonomous devices that are able to collect, store, process and share data with other devices. Large sensor networks are often redundant in the sense that the measurements of some nodes can be substituted by other nodes with a certain degree of confidence. This spatial correlation results in wastage of link bandwidth and energy. In this paper, a model for two associated Poisson processes, through which sensors are distributed in a plane, is derived. A probability condition is established for data redundancy among closely located sensor nodes. The model generates a spatial bivariate Poisson process whose parameters depend on the parameters of the two individual Poisson processes and on the distance between the associated points. The proposed model helps in building efficient algorithms for data dissemination in the sensor network. A numerical example is provided investigating the advantage of this model. 20. Discrete event simulation: Modeling simultaneous complications and outcomes NARCIS (Netherlands) Quik, E.H.; Feenstra, T.L.; Krabbe, P.F.M. 2012-01-01 OBJECTIVES: To present an effective and elegant model approach to deal with specific characteristics of complex modeling. METHODS: A discrete event simulation (DES) model with multiple complications and multiple outcomes that each can occur simultaneously was developed. In this DES model parameters, 1. A Probability-Based Hybrid User Model for Recommendation System Directory of Open Access Journals (Sweden) Jia Hao 2016-01-01 Full Text Available With the rapid development of information communication technology, the available information or knowledge is exponentially increased, and this causes the well-known information overload phenomenon. This problem is more serious in product design corporations because over half of the valuable design time is consumed in knowledge acquisition, which highly extends the design cycle and weakens the competitiveness. Therefore, the recommender systems become very important in the domain of product domain. This research presents a probability-based hybrid user model, which is a combination of collaborative filtering and content-based filtering. This hybrid model utilizes user ratings and item topics or classes, which are available in the domain of product design, to predict the knowledge requirement. The comprehensive analysis of the experimental results shows that the proposed method gains better performance in most of the parameter settings. This work contributes a probability-based method to the community for implement recommender system when only user ratings and item topics are available. 2. On the Inclusion of Short-distance Bystander Effects into a Logistic Tumor Control Probability Model. Science.gov (United States) Tempel, David G; Brodin, N Patrik; Tomé, Wolfgang A 2018-01-01 Currently, interactions between voxels are neglected in the tumor control probability (TCP) models used in biologically-driven intensity-modulated radiotherapy treatment planning. However, experimental data suggests that this may not always be justified when bystander effects are important. We propose a model inspired by the Ising model, a short-range interaction model, to investigate if and when it is important to include voxel to voxel interactions in biologically-driven treatment planning. This Ising-like model for TCP is derived by first showing that the logistic model of tumor control is mathematically equivalent to a non-interacting Ising model. Using this correspondence, the parameters of the logistic model are mapped to the parameters of an Ising-like model and bystander interactions are introduced as a short-range interaction as is the case for the Ising model. As an example, we apply the model to study the effect of bystander interactions in the case of radiation therapy for prostate cancer. The model shows that it is adequate to neglect bystander interactions for dose distributions that completely cover the treatment target and yield TCP estimates that lie in the shoulder of the dose response curve. However, for dose distributions that yield TCP estimates that lie on the steep part of the dose response curve or for inhomogeneous dose distributions having significant hot and/or cold regions, bystander effects may be important. Furthermore, the proposed model highlights a previously unexplored and potentially fruitful connection between the fields of statistical mechanics and tumor control probability/normal tissue complication probability modeling. 3. The effect of 6 and 15 MV on intensity-modulated radiation therapy prostate cancer treatment: plan evaluation, tumour control probability and normal tissue complication probability analysis, and the theoretical risk of secondary induced malignancies Science.gov (United States) Hussein, M; Aldridge, S; Guerrero Urbano, T; Nisbet, A 2012-01-01 Objective The aim of this study was to investigate the effect of 6 and 15-MV photon energies on intensity-modulated radiation therapy (IMRT) prostate cancer treatment plan outcome and to compare the theoretical risks of secondary induced malignancies. Methods Separate prostate cancer IMRT plans were prepared for 6 and 15-MV beams. Organ-equivalent doses were obtained through thermoluminescent dosemeter measurements in an anthropomorphic Aldersen radiation therapy human phantom. The neutron dose contribution at 15 MV was measured using polyallyl-diglycol-carbonate neutron track etch detectors. Risk coefficients from the International Commission on Radiological Protection Report 103 were used to compare the risk of fatal secondary induced malignancies in out-of-field organs and tissues for 6 and 15 MV. For the bladder and the rectum, a comparative evaluation of the risk using three separate models was carried out. Dose–volume parameters for the rectum, bladder and prostate planning target volume were evaluated, as well as normal tissue complication probability (NTCP) and tumour control probability calculations. Results There is a small increased theoretical risk of developing a fatal cancer from 6 MV compared with 15 MV, taking into account all the organs. Dose–volume parameters for the rectum and bladder show that 15 MV results in better volume sparing in the regions below 70 Gy, but the volume exposed increases slightly beyond this in comparison with 6 MV, resulting in a higher NTCP for the rectum of 3.6% vs 3.0% (p=0.166). Conclusion The choice to treat using IMRT at 15 MV should not be excluded, but should be based on risk vs benefit while considering the age and life expectancy of the patient together with the relative risk of radiation-induced cancer and NTCPs. PMID:22010028 4. Modelling the Probability of Landslides Impacting Road Networks Science.gov (United States) Taylor, F. E.; Malamud, B. D. 2012-04-01 5. Identification of Patient Benefit From Proton Therapy for Advanced Head and Neck Cancer Patients Based on Individual and Subgroup Normal Tissue Complication Probability Analysis Energy Technology Data Exchange (ETDEWEB) Jakobi, Annika, E-mail: Annika.Jakobi@OncoRay.de [OncoRay-National Center for Radiation Research in Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Helmholtz-Zentrum Dresden-Rossendorf, Dresden (Germany); Bandurska-Luque, Anna [OncoRay-National Center for Radiation Research in Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Helmholtz-Zentrum Dresden-Rossendorf, Dresden (Germany); Department of Radiation Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden (Germany); Stützer, Kristin; Haase, Robert; Löck, Steffen [OncoRay-National Center for Radiation Research in Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Helmholtz-Zentrum Dresden-Rossendorf, Dresden (Germany); Wack, Linda-Jacqueline [Section for Biomedical Physics, University Hospital for Radiation Oncology, Eberhard Karls Universät Tübingen (Germany); Mönnich, David [Section for Biomedical Physics, University Hospital for Radiation Oncology, Eberhard Karls Universät Tübingen (Germany); German Cancer Research Center, Heidelberg (Germany); German Cancer Consortium, Tübingen (Germany); Thorwarth, Daniela [Section for Biomedical Physics, University Hospital for Radiation Oncology, Eberhard Karls Universät Tübingen (Germany); and others 2015-08-01 Purpose: The purpose of this study was to determine, by treatment plan comparison along with normal tissue complication probability (NTCP) modeling, whether a subpopulation of patients with head and neck squamous cell carcinoma (HNSCC) could be identified that would gain substantial benefit from proton therapy in terms of NTCP. Methods and Materials: For 45 HNSCC patients, intensity modulated radiation therapy (IMRT) was compared to intensity modulated proton therapy (IMPT). Physical dose distributions were evaluated as well as the resulting NTCP values, using modern models for acute mucositis, xerostomia, aspiration, dysphagia, laryngeal edema, and trismus. Patient subgroups were defined based on primary tumor location. Results: Generally, IMPT reduced the NTCP values while keeping similar target coverage for all patients. Subgroup analyses revealed a higher individual reduction of swallowing-related side effects by IMPT for patients with tumors in the upper head and neck area, whereas the risk reduction of acute mucositis was more pronounced in patients with tumors in the larynx region. More patients with tumors in the upper head and neck area had a reduction in NTCP of more than 10%. Conclusions: Subgrouping can help to identify patients who may benefit more than others from the use of IMPT and, thus, can be a useful tool for a preselection of patients in the clinic where there are limited PT resources. Because the individual benefit differs within a subgroup, the relative merits should additionally be evaluated by individual treatment plan comparisons. 6. Identification of Patient Benefit From Proton Therapy for Advanced Head and Neck Cancer Patients Based on Individual and Subgroup Normal Tissue Complication Probability Analysis International Nuclear Information System (INIS) Jakobi, Annika; Bandurska-Luque, Anna; Stützer, Kristin; Haase, Robert; Löck, Steffen; Wack, Linda-Jacqueline; Mönnich, David; Thorwarth, Daniela 2015-01-01 Purpose: The purpose of this study was to determine, by treatment plan comparison along with normal tissue complication probability (NTCP) modeling, whether a subpopulation of patients with head and neck squamous cell carcinoma (HNSCC) could be identified that would gain substantial benefit from proton therapy in terms of NTCP. Methods and Materials: For 45 HNSCC patients, intensity modulated radiation therapy (IMRT) was compared to intensity modulated proton therapy (IMPT). Physical dose distributions were evaluated as well as the resulting NTCP values, using modern models for acute mucositis, xerostomia, aspiration, dysphagia, laryngeal edema, and trismus. Patient subgroups were defined based on primary tumor location. Results: Generally, IMPT reduced the NTCP values while keeping similar target coverage for all patients. Subgroup analyses revealed a higher individual reduction of swallowing-related side effects by IMPT for patients with tumors in the upper head and neck area, whereas the risk reduction of acute mucositis was more pronounced in patients with tumors in the larynx region. More patients with tumors in the upper head and neck area had a reduction in NTCP of more than 10%. Conclusions: Subgrouping can help to identify patients who may benefit more than others from the use of IMPT and, thus, can be a useful tool for a preselection of patients in the clinic where there are limited PT resources. Because the individual benefit differs within a subgroup, the relative merits should additionally be evaluated by individual treatment plan comparisons 7. Robust mislabel logistic regression without modeling mislabel probabilities. Science.gov (United States) Hung, Hung; Jou, Zhi-Yu; Huang, Su-Yun 2018-03-01 Logistic regression is among the most widely used statistical methods for linear discriminant analysis. In many applications, we only observe possibly mislabeled responses. Fitting a conventional logistic regression can then lead to biased estimation. One common resolution is to fit a mislabel logistic regression model, which takes into consideration of mislabeled responses. Another common method is to adopt a robust M-estimation by down-weighting suspected instances. In this work, we propose a new robust mislabel logistic regression based on γ-divergence. Our proposal possesses two advantageous features: (1) It does not need to model the mislabel probabilities. (2) The minimum γ-divergence estimation leads to a weighted estimating equation without the need to include any bias correction term, that is, it is automatically bias-corrected. These features make the proposed γ-logistic regression more robust in model fitting and more intuitive for model interpretation through a simple weighting scheme. Our method is also easy to implement, and two types of algorithms are included. Simulation studies and the Pima data application are presented to demonstrate the performance of γ-logistic regression. © 2017, The International Biometric Society. 8. Cladding failure probability modeling for risk evaluations of fast reactors International Nuclear Information System (INIS) Mueller, C.J.; Kramer, J.M. 1987-01-01 This paper develops the methodology to incorporate cladding failure data and associated modeling into risk evaluations of liquid metal-cooled fast reactors (LMRs). Current US innovative designs for metal-fueled pool-type LMRs take advantage of inherent reactivity feedback mechanisms to limit reactor temperature increases in response to classic anticipated-transient-without-scram (ATWS) initiators. Final shutdown without reliance on engineered safety features can then be accomplished if sufficient time is available for operator intervention to terminate fission power production and/or provide auxiliary cooling prior to significant core disruption. Coherent cladding failure under the sustained elevated temperatures of ATWS events serves as one indicator of core disruption. In this paper we combine uncertainties in cladding failure data with uncertainties in calculations of ATWS cladding temperature conditions to calculate probabilities of cladding failure as a function of the time for accident recovery 9. Cladding failure probability modeling for risk evaluations of fast reactors International Nuclear Information System (INIS) Mueller, C.J.; Kramer, J.M. 1987-01-01 This paper develops the methodology to incorporate cladding failure data and associated modeling into risk evaluations of liquid metal-cooled fast reactors (LMRs). Current U.S. innovative designs for metal-fueled pool-type LMRs take advantage of inherent reactivity feedback mechanisms to limit reactor temperature increases in response to classic anticipated-transient-without-scram (ATWS) initiators. Final shutdown without reliance on engineered safety features can then be accomplished if sufficient time is available for operator intervention to terminate fission power production and/or provide auxiliary cooling prior to significant core disruption. Coherent cladding failure under the sustained elevated temperatures of ATWS events serves as one indicator of core disruption. In this paper we combine uncertainties in cladding failure data with uncertainties in calculations of ATWS cladding temperature conditions to calculate probabilities of cladding failure as a function of the time for accident recovery. (orig.) 10. Increase in tumor control and normal tissue complication probabilities in advanced head-and-neck cancer for dose-escalated intensity-modulated photon and proton therapy Directory of Open Access Journals (Sweden) Annika eJakobi 2015-11-01 Full Text Available Introduction:Presently used radio-chemotherapy regimens result in moderate local control rates for patients with advanced head and neck squamous cell carcinoma (HNSCC. Dose escalation (DE may be an option to improve patient outcome, but may also increase the risk of toxicities in healthy tissue. The presented treatment planning study evaluated the feasibility of two DE levels for advanced HNSCC patients, planned with either intensity-modulated photon therapy (IMXT or proton therapy (IMPT.Materials and Methods:For 45 HNSCC patients, IMXT and IMPT treatment plans were created including DE via a simultaneous integrated boost (SIB in the high-risk volume, while maintaining standard fractionation with 2 Gy per fraction in the remaining target volume. Two DE levels for the SIB were compared: 2.3 Gy and 2.6 Gy. Treatment plan evaluation included assessment of tumor control probabilities (TCP and normal tissue complication probabilities (NTCP.Results:An increase of approximately 10% in TCP was estimated between the DE levels. A pronounced high-dose rim surrounding the SIB volume was identified in IMXT treatment. Compared to IMPT, this extra dose slightly increased the TCP values and to a larger extent the NTCP values. For both modalities, the higher DE level led only to a small increase in NTCP values (mean differences < 2% in all models, except for the risk of aspiration, which increased on average by 8% and 6% with IMXT and IMPT, respectively, but showed a considerable patient dependence. Conclusions:Both DE levels appear applicable to patients with IMXT and IMPT since all calculated NTCP values, except for one, increased only little for the higher DE level. The estimated TCP increase is of relevant magnitude. The higher DE schedule needs to be investigated carefully in the setting of a prospective clinical trial, especially regarding toxicities caused by high local doses that lack a sound dose response description, e.g., ulcers. 11. Normal Tissue Complication Probability Analysis of Acute Gastrointestinal Toxicity in Cervical Cancer Patients Undergoing Intensity Modulated Radiation Therapy and Concurrent Cisplatin International Nuclear Information System (INIS) Simpson, Daniel R.; Song, William Y.; Moiseenko, Vitali; Rose, Brent S.; Yashar, Catheryn M.; Mundt, Arno J.; Mell, Loren K. 2012-01-01 12. Dosimetric complication probability and acoustic analysis of vocal cord region in oropharyngeal carcinoma treated with voice-sparing intensity modulated radiotherapy International Nuclear Information System (INIS) Jain, S.; Gupta, T.; Agarwal, J.P.; Baccher, G.; Shrivastava, S.K.; Reenadevi; Master, J. 2008-01-01 Radiation to larynx has long been associated with speech and voice dysfunction. The objective is to study dosimetric parameters and complication probability of vocal cord region (VCR) and the effect of voice-sparing (VS) in the patients treated with intensity modulated radiotherapy (IMRT). The secondary objective is to describe the post-radiation acoustic voice characteristics and correlate them with the dosimetric parameters. (author) 13. High-resolution urban flood modelling - a joint probability approach Science.gov (United States) Hartnett, Michael; Olbert, Agnieszka; Nash, Stephen 2017-04-01 (Divoky et al., 2005). Nevertheless, such events occur and in Ireland alone there are several cases of serious damage due to flooding resulting from a combination of high sea water levels and river flows driven by the same meteorological conditions (e.g. Olbert et al. 2015). A November 2009 fluvial-coastal flooding of Cork City bringing €100m loss was one such incident. This event was used by Olbert et al. (2015) to determine processes controlling urban flooding and is further explored in this study to elaborate on coastal and fluvial flood mechanisms and their roles in controlling water levels. The objective of this research is to develop a methodology to assess combined effect of multiple source flooding on flood probability and severity in urban areas and to establish a set of conditions that dictate urban flooding due to extreme climatic events. These conditions broadly combine physical flood drivers (such as coastal and fluvial processes), their mechanisms and thresholds defining flood severity. The two main physical processes controlling urban flooding: high sea water levels (coastal flooding) and high river flows (fluvial flooding), and their threshold values for which flood is likely to occur, are considered in this study. Contribution of coastal and fluvial drivers to flooding and their impacts are assessed in a two-step process. The first step involves frequency analysis and extreme value statistical modelling of storm surges, tides and river flows and ultimately the application of joint probability method to estimate joint exceedence return periods for combination of surges, tide and river flows. In the second step, a numerical model of Cork Harbour MSN_Flood comprising a cascade of four nested high-resolution models is used to perform simulation of flood inundation under numerous hypothetical coastal and fluvial flood scenarios. The risk of flooding is quantified based on a range of physical aspects such as the extent and depth of inundation (Apel et al 14. Modelling soft error probability in firmware: A case study African Journals Online (AJOL) The purpose is to estimate the probability that external disruptive events (such as ..... also changed the 16-bit magic variable to its unique 'magic' value. .... is mutually independent, not only over registers but over spikes, such that the above. 15. Normal tissue complication probability: Does simultaneous integrated boost intensity-modulated radiotherapy score over other techniques in treatment of prostate adenocarcinoma Directory of Open Access Journals (Sweden) Jothy Basu K 2009-01-01 Full Text Available Aim: The main objective of this study was to analyze the radiobiological effect of different treatment strategies on high-risk prostate adenocarcinoma. Materials and Methods: Ten cases of high-risk prostate adenocarcinoma were selected for this dosimetric study. Four different treatment strategies used for treating prostate cancer were compared. Conventional four-field box technique covering prostate and nodal volumes followed by three-field conformal boost (3D + 3DCRT, four-field box technique followed by intensity-modulated radiotherapy (IMRT boost (3D + IMRT, IMRT followed by IMRT boost (IMRT + IMRT, and simultaneous integrated boost IMRT (SIBIMRT were compared in terms of tumor control probability (TCP and normal tissue complication probability (NTCP. The dose prescription except for SIBIMRT was 45 Gy in 25 fractions for the prostate and nodal volumes in the initial phase and 27 Gy in 15 fractions for the prostate in the boost phase. For SIBIMRT, equivalent doses were calculated using biologically equivalent dose assuming the α/β ratio of 1.5 Gy with a dose prescription of 60.75 Gy for the gross tumor volume (GTV and 45 Gy for the clinical target volume in 25 fractions. IMRT plans were made with 15-MV equispaced seven coplanar fields. NTCP was calculated using the Lyman-Kutcher-Burman (LKB model. Results: An NTCP of 10.7 ± 0.99%, 8.36 ± 0.66%, 6.72 ± 0.85%, and 1.45 ± 0.11% for the bladder and 14.9 ± 0.99%, 14.04 ± 0.66%, 11.38 ± 0.85%, 5.12 ± 0.11% for the rectum was seen with 3D + 3DCRT, 3D + IMRT, IMRT + IMRT, and SIBIMRT respectively. Conclusions: SIBIMRT had the least NTCP over all other strategies with a reduced treatment time (3 weeks less. It should be the technique of choice for dose escalation in prostate carcinoma. 16. Portomesenteric venous thrombosis: A rare but probably under-reported complication of laparoscopic surgery: A case series Directory of Open Access Journals (Sweden) Yan Mei Goh 2017-01-01 Full Text Available Portomesenteric venous thrombosis (PMVT is a rare but well-reported complication following laparoscopic surgery. We present three cases of PMVT following laparoscopic surgery. Our first case is a 71-year-old morbidly obese woman admitted for elective laparoscopic giant hiatus hernia (LGHH repair. Post-operatively, she developed multi-organ dysfunction and computed tomography scan revealed portal venous gas and extensive small bowel infarct. The second patient is a 51-year-old man with known previous deep venous thrombosis who also had elective LGHH repair. He presented 8 weeks post-operatively with severe abdominal pain and required major bowel resection. Our third case is an 86-year-old woman who developed worsening abdominal tenderness 3 days after laparoscopic right hemicolectomy for adenocarcinoma and was diagnosed with an incidental finding of thrombus in the portal vein. She did not require further surgical intervention. The current guidelines for thromboprophylaxis follow-up in this patient group may not be adequate for the patients at risk. Hence, we propose prolonged period of thromboprophylaxis in the patients undergoing major laparoscopic surgery. 17. Optimizing an objective function under a bivariate probability model NARCIS (Netherlands) X. Brusset; N.M. Temme (Nico) 2007-01-01 htmlabstractThe motivation of this paper is to obtain an analytical closed form of a quadratic objective function arising from a stochastic decision process with bivariate exponential probability distribution functions that may be dependent. This method is applicable when results need to be 18. Parametric modeling of probability of bank loan default in Kenya ... African Journals Online (AJOL) This makes the study on probability of a customer defaulting very useful while analyzing the credit risk policies. In this paper, we use a raw data set that contains demographic information about the borrowers. The data sets have been used to identify which risk factors associated with the borrowers contribute towards default. 19. Naive Probability: Model-based Estimates of Unique Events Science.gov (United States) 2014-05-04 of inference. Argument and Computation, 1–17, iFirst. Khemlani, S., & Johnson-Laird, P.N. (2012b). Theories of the syllogism: A meta -analysis...is the probability that… 1 space tourism will achieve widespread popularity in the next 50 years? advances in material science will lead to the... governments dedicate more resources to contacting extra-terrestrials? 8 the United States adopts an open border policy of universal acceptance? English is 20. Risk Prediction Model for Severe Postoperative Complication in Bariatric Surgery. Science.gov (United States) Stenberg, Erik; Cao, Yang; Szabo, Eva; Näslund, Erik; Näslund, Ingmar; Ottosson, Johan 2018-01-12 Factors associated with risk for adverse outcome are important considerations in the preoperative assessment of patients for bariatric surgery. As yet, prediction models based on preoperative risk factors have not been able to predict adverse outcome sufficiently. This study aimed to identify preoperative risk factors and to construct a risk prediction model based on these. Patients who underwent a bariatric surgical procedure in Sweden between 2010 and 2014 were identified from the Scandinavian Obesity Surgery Registry (SOReg). Associations between preoperative potential risk factors and severe postoperative complications were analysed using a logistic regression model. A multivariate model for risk prediction was created and validated in the SOReg for patients who underwent bariatric surgery in Sweden, 2015. Revision surgery (standardized OR 1.19, 95% confidence interval (CI) 1.14-0.24, p prediction model. Despite high specificity, the sensitivity of the model was low. Revision surgery, high age, low BMI, large waist circumference, and dyspepsia/GERD were associated with an increased risk for severe postoperative complication. The prediction model based on these factors, however, had a sensitivity that was too low to predict risk in the individual patient case. 1. Underground economy modelling: simple models with complicated dynamics OpenAIRE Albu, Lucian-Liviu 2003-01-01 The paper aims to model the underground economy using two different models: one based on the labor supply method and a generalized model for the allocation of time. The model based on the labor supply method is conceived as a simulating one in order to determine some reasonable thresholds of the underground sector extension based only on the available macroeconomic statistical data. The generalized model for the allocation of time is a model based on direct approach which estimates the underg... 2. Wind-wave modelling aspects within complicate topography Directory of Open Access Journals (Sweden) S. Christopoulos Full Text Available Wave forecasting aspects for basins with complicate geomorphology, such as the Aegean Sea, are investigated through an intercomparison study. The efficiency of the available wind models (ECMWF, UKMO to reproduce wind patterns over special basins, as well as three wave models incorporating different physics and characteristics (WAM, AUT, WACCAS, are tested for selected storm cases representing the typical wind situations over the basin. From the wave results, discussed in terms of time-series and statistical parameters, the crucial role is pointed out of the wind resolution and the reliability of the different wave models to estimate the wave climate in such a basin. The necessary grid resolution is also tested, while for a specific test case (December 1991 ERS-1 satellite data are compared with those of the model. 3. Canonical Probability Distributions for Model Building, Learning, and Inference National Research Council Canada - National Science Library Druzdzel, Marek J 2006-01-01 ...) improvements of stochastic sampling algorithms based on importance sampling, and (3) practical applications of our general purpose decision modeling environment to diagnosis of complex systems... 4. Estimation and asymptotic theory for transition probabilities in Markov Renewal Multi–state models NARCIS (Netherlands) Spitoni, C.; Verduijn, M.; Putter, H. 2012-01-01 In this paper we discuss estimation of transition probabilities for semi–Markov multi–state models. Non–parametric and semi–parametric estimators of the transition probabilities for a large class of models (forward going models) are proposed. Large sample theory is derived using the functional 5. Direct modeling of regression effects for transition probabilities in the progressive illness-death model DEFF Research Database (Denmark) Azarang, Leyla; Scheike, Thomas; de Uña-Álvarez, Jacobo 2017-01-01 In this work, we present direct regression analysis for the transition probabilities in the possibly non-Markov progressive illness–death model. The method is based on binomial regression, where the response is the indicator of the occupancy for the given state along time. Randomly weighted score... 6. Modeling the probability of giving birth at health institutions among ... African Journals Online (AJOL) 2014-06-02 Jun 2, 2014 ... deemed sufficient evidence of the utility of the logistic regression model. .... making decision by herself and gravid had significantly predicted the ..... Health Education: Theory, Research and Practice-3rd edition.2002. 7. Complications from arteriovenous malformation radiosurgery: multivariate analysis and risk modeling International Nuclear Information System (INIS) Flickinger, John C.; Kondziolka, Douglas; Pollock, Bruce E.; Maitz, Ann H.; Lunsford, L. Dade 1997-01-01 Purpose/Objective: To assess the relationships of radiosurgery treatment parameters to the development of complications from radiosurgery for arteriovenous malformations (AVM). Methods and Materials: We evaluated follow-up imaging and clinical data in 307 AVM patients who received gamma knife radiosurgery at the University of Pittsburgh between 1987 and 1993. All patients had regular clinical or imaging follow up for a minimum of 2 years (range: 24-96 months, median = 44 months). Results: Post-radiosurgical imaging (PRI) changes developed in 30.5% of patients with regular follow-up magnetic resonance imaging, and were symptomatic in 10.7% of all patients at 7 years. PRI changes resolved within 3 years developed significantly less often (p = 0.0274) in patients with symptoms (52.8%) compared to asymptomatic patients (94.8%). The 7-year actuarial rate for developing persistent symptomatic PRI changes was 5.05%. Multivariate logistic regression modeling found that the 12 Gy volume was the only independent variable that correlated significantly with PRI changes (p < 0.0001) while symptomatic PRI changes were correlated with both 12 Gy volume (p = 0.0013) and AVM location (p 0.0066). Conclusion: Complications from AVM radiosurgery can be predicted with a statistical model relating the risks of developing symptomatic post-radiosurgical imaging changes to 12 Gy treatment volume and location 8. Investigating the probability of detection of typical cavity shapes through modelling and comparison of geophysical techniques Science.gov (United States) James, P. 2011-12-01 With a growing need for housing in the U.K., the government has proposed increased development of brownfield sites. However, old mine workings and natural cavities represent a potential hazard before, during and after construction on such sites, and add further complication to subsurface parameters. Cavities are hence a limitation to certain redevelopment and their detection is an ever important consideration. The current standard technique for cavity detection is a borehole grid, which is intrusive, non-continuous, slow and expensive. A new robust investigation standard in the detection of cavities is sought and geophysical techniques offer an attractive alternative. Geophysical techniques have previously been utilised successfully in the detection of cavities in various geologies, but still has an uncertain reputation in the engineering industry. Engineers are unsure of the techniques and are inclined to rely on well known techniques than utilise new technologies. Bad experiences with geophysics are commonly due to the indiscriminate choice of particular techniques. It is imperative that a geophysical survey is designed with the specific site and target in mind at all times, and the ability and judgement to rule out some, or all, techniques. To this author's knowledge no comparative software exists to aid technique choice. Also, previous modelling software limit the shapes of bodies and hence typical cavity shapes are not represented. Here, we introduce 3D modelling software (Matlab) which computes and compares the response to various cavity targets from a range of techniques (gravity, gravity gradient, magnetic, magnetic gradient and GPR). Typical near surface cavity shapes are modelled including shafts, bellpits, various lining and capping materials, and migrating voids. The probability of cavity detection is assessed in typical subsurface and noise conditions across a range of survey parameters. Techniques can be compared and the limits of detection distance 9. Discrete probability models and methods probability on graphs and trees, Markov chains and random fields, entropy and coding CERN Document Server Brémaud, Pierre 2017-01-01 The emphasis in this book is placed on general models (Markov chains, random fields, random graphs), universal methods (the probabilistic method, the coupling method, the Stein-Chen method, martingale methods, the method of types) and versatile tools (Chernoff's bound, Hoeffding's inequality, Holley's inequality) whose domain of application extends far beyond the present text. Although the examples treated in the book relate to the possible applications, in the communication and computing sciences, in operations research and in physics, this book is in the first instance concerned with theory. The level of the book is that of a beginning graduate course. It is self-contained, the prerequisites consisting merely of basic calculus (series) and basic linear algebra (matrices). The reader is not assumed to be trained in probability since the first chapters give in considerable detail the background necessary to understand the rest of the book. . 10. Some aspects of statistical modeling of human-error probability International Nuclear Information System (INIS) Prairie, R.R. 1982-01-01 Human reliability analyses (HRA) are often performed as part of risk assessment and reliability projects. Recent events in nuclear power have shown the potential importance of the human element. There are several on-going efforts in the US and elsewhere with the purpose of modeling human error such that the human contribution can be incorporated into an overall risk assessment associated with one or more aspects of nuclear power. An effort that is described here uses the HRA (event tree) to quantify and model the human contribution to risk. As an example, risk analyses are being prepared on several nuclear power plants as part of the Interim Reliability Assessment Program (IREP). In this process the risk analyst selects the elements of his fault tree that could be contributed to by human error. He then solicits the HF analyst to do a HRA on this element 11. Probability model for worst case solar proton event fluences International Nuclear Information System (INIS) Xapsos, M.A.; Summers, G.P.; Barth, J.L.; Stassinopoulos, E.G.; Burke, E.A. 1999-01-01 The effects that solar proton events have on microelectronics and solar arrays are important considerations for spacecraft in geostationary orbits, polar orbits and on interplanetary missions. A predictive model of worst case solar proton event fluences is presented. It allows the expected worst case event fluence to be calculated for a given confidence level and for periods of time corresponding to space missions. The proton energy range is from >1 to >300 MeV, so that the model is useful for a variety of radiation effects applications. For each proton energy threshold, the maximum entropy principle is used to select the initial distribution of solar proton event fluences. This turns out to be a truncated power law, i.e., a power law for smaller event fluences that smoothly approaches zero at a maximum fluence. The strong agreement of the distribution with satellite data for the last three solar cycles indicates this description captures the essential features of a solar proton event fluence distribution. Extreme value theory is then applied to the initial distribution of events to obtain the model of worst case fluences 12. Linear-quadratic model predictions for tumor control probability International Nuclear Information System (INIS) Yaes, R.J. 1987-01-01 Sigmoid dose-response curves for tumor control are calculated from the linear-quadratic model parameters α and Β, obtained from human epidermoid carcinoma cell lines, and are much steeper than the clinical dose-response curves for head and neck cancers. One possible explanation is the presence of small radiation-resistant clones arising from mutations in an initially homogeneous tumor. Using the mutation theory of Delbruck and Luria and of Goldie and Coldman, the authors discuss the implications of such radiation-resistant clones for clinical radiation therapy 13. Interpretation of the results of statistical measurements. [search for basic probability model Science.gov (United States) Olshevskiy, V. V. 1973-01-01 For random processes, the calculated probability characteristic, and the measured statistical estimate are used in a quality functional, which defines the difference between the two functions. Based on the assumption that the statistical measurement procedure is organized so that the parameters for a selected model are optimized, it is shown that the interpretation of experimental research is a search for a basic probability model. 14. Ruin probability of the renewal model with risky investment and large claims Institute of Scientific and Technical Information of China (English) 2009-01-01 The ruin probability of the renewal risk model with investment strategy for a capital market index is investigated in this paper.For claim sizes with common distribution of extended regular variation,we study the asymptotic behaviour of the ruin probability.As a corollary,we establish a simple asymptotic formula for the ruin probability for the case of Pareto-like claims. 15. Using multinomial and imprecise probability for non-parametric modelling of rainfall in Manizales (Colombia Directory of Open Access Journals (Sweden) Ibsen Chivatá Cárdenas 2008-05-01 Full Text Available This article presents a rainfall model constructed by applying non-parametric modelling and imprecise probabilities; these tools were used because there was not enough homogeneous information in the study area. The area’s hydro-logical information regarding rainfall was scarce and existing hydrological time series were not uniform. A distributed extended rainfall model was constructed from so-called probability boxes (p-boxes, multinomial probability distribu-tion and confidence intervals (a friendly algorithm was constructed for non-parametric modelling by combining the last two tools. This model confirmed the high level of uncertainty involved in local rainfall modelling. Uncertainty en-compassed the whole range (domain of probability values thereby showing the severe limitations on information, leading to the conclusion that a detailed estimation of probability would lead to significant error. Nevertheless, rele-vant information was extracted; it was estimated that maximum daily rainfall threshold (70 mm would be surpassed at least once every three years and the magnitude of uncertainty affecting hydrological parameter estimation. This paper’s conclusions may be of interest to non-parametric modellers and decisions-makers as such modelling and imprecise probability represents an alternative for hydrological variable assessment and maybe an obligatory proce-dure in the future. Its potential lies in treating scarce information and represents a robust modelling strategy for non-seasonal stochastic modelling conditions 16. The ruin probability of a discrete time risk model under constant interest rate with heavy tails NARCIS (Netherlands) Tang, Q. 2004-01-01 This paper investigates the ultimate ruin probability of a discrete time risk model with a positive constant interest rate. Under the assumption that the gross loss of the company within one year is subexponentially distributed, a simple asymptotic relation for the ruin probability is derived and 17. Compact baby universe model in ten dimension and probability function of quantum gravity International Nuclear Information System (INIS) Yan Jun; Hu Shike 1991-01-01 The quantum probability functions are calculated for ten-dimensional compact baby universe model. The authors find that the probability for the Yang-Mills baby universe to undergo a spontaneous compactification down to a four-dimensional spacetime is greater than that to remain in the original homogeneous multidimensional state. Some questions about large-wormhole catastrophe are also discussed 18. Probabilistic Inference: Task Dependency and Individual Differences of Probability Weighting Revealed by Hierarchical Bayesian Modeling. Science.gov (United States) Boos, Moritz; Seer, Caroline; Lange, Florian; Kopp, Bruno 2016-01-01 Cognitive determinants of probabilistic inference were examined using hierarchical Bayesian modeling techniques. A classic urn-ball paradigm served as experimental strategy, involving a factorial two (prior probabilities) by two (likelihoods) design. Five computational models of cognitive processes were compared with the observed behavior. Parameter-free Bayesian posterior probabilities and parameter-free base rate neglect provided inadequate models of probabilistic inference. The introduction of distorted subjective probabilities yielded more robust and generalizable results. A general class of (inverted) S-shaped probability weighting functions had been proposed; however, the possibility of large differences in probability distortions not only across experimental conditions, but also across individuals, seems critical for the model's success. It also seems advantageous to consider individual differences in parameters of probability weighting as being sampled from weakly informative prior distributions of individual parameter values. Thus, the results from hierarchical Bayesian modeling converge with previous results in revealing that probability weighting parameters show considerable task dependency and individual differences. Methodologically, this work exemplifies the usefulness of hierarchical Bayesian modeling techniques for cognitive psychology. Theoretically, human probabilistic inference might be best described as the application of individualized strategic policies for Bayesian belief revision. 19. Probabilistic inference: Task dependency and individual differences of probability weighting revealed by hierarchical Bayesian modelling Directory of Open Access Journals (Sweden) Moritz eBoos 2016-05-01 Full Text Available Cognitive determinants of probabilistic inference were examined using hierarchical Bayesian modelling techniques. A classic urn-ball paradigm served as experimental strategy, involving a factorial two (prior probabilities by two (likelihoods design. Five computational models of cognitive processes were compared with the observed behaviour. Parameter-free Bayesian posterior probabilities and parameter-free base rate neglect provided inadequate models of probabilistic inference. The introduction of distorted subjective probabilities yielded more robust and generalizable results. A general class of (inverted S-shaped probability weighting functions had been proposed; however, the possibility of large differences in probability distortions not only across experimental conditions, but also across individuals, seems critical for the model’s success. It also seems advantageous to consider individual differences in parameters of probability weighting as being sampled from weakly informative prior distributions of individual parameter values. Thus, the results from hierarchical Bayesian modelling converge with previous results in revealing that probability weighting parameters show considerable task dependency and individual differences. Methodologically, this work exemplifies the usefulness of hierarchical Bayesian modelling techniques for cognitive psychology. Theoretically, human probabilistic inference might be best described as the application of individualized strategic policies for Bayesian belief revision. 20. An analytical calculation of neighbourhood order probabilities for high dimensional Poissonian processes and mean field models International Nuclear Information System (INIS) Tercariol, Cesar Augusto Sangaletti; Kiipper, Felipe de Moura; Martinez, Alexandre Souto 2007-01-01 Consider that the coordinates of N points are randomly generated along the edges of a d-dimensional hypercube (random point problem). The probability P (d,N) m,n that an arbitrary point is the mth nearest neighbour to its own nth nearest neighbour (Cox probabilities) plays an important role in spatial statistics. Also, it has been useful in the description of physical processes in disordered media. Here we propose a simpler derivation of Cox probabilities, where we stress the role played by the system dimensionality d. In the limit d → ∞, the distances between pair of points become independent (random link model) and closed analytical forms for the neighbourhood probabilities are obtained both for the thermodynamic limit and finite-size system. Breaking the distance symmetry constraint drives us to the random map model, for which the Cox probabilities are obtained for two cases: whether a point is its own nearest neighbour or not 1. Risk Probabilities DEFF Research Database (Denmark) Rojas-Nandayapa, Leonardo Tail probabilities of sums of heavy-tailed random variables are of a major importance in various branches of Applied Probability, such as Risk Theory, Queueing Theory, Financial Management, and are subject to intense research nowadays. To understand their relevance one just needs to think...... analytic expression for the distribution function of a sum of random variables. The presence of heavy-tailed random variables complicates the problem even more. The objective of this dissertation is to provide better approximations by means of sharp asymptotic expressions and Monte Carlo estimators... 2. Developing a probability-based model of aquifer vulnerability in an agricultural region Science.gov (United States) Chen, Shih-Kai; Jang, Cheng-Shin; Peng, Yi-Huei 2013-04-01 SummaryHydrogeological settings of aquifers strongly influence the regional groundwater movement and pollution processes. Establishing a map of aquifer vulnerability is considerably critical for planning a scheme of groundwater quality protection. This study developed a novel probability-based DRASTIC model of aquifer vulnerability in the Choushui River alluvial fan, Taiwan, using indicator kriging and to determine various risk categories of contamination potentials based on estimated vulnerability indexes. Categories and ratings of six parameters in the probability-based DRASTIC model were probabilistically characterized according to the parameter classification methods of selecting a maximum estimation probability and calculating an expected value. Moreover, the probability-based estimation and assessment gave us an excellent insight into propagating the uncertainty of parameters due to limited observation data. To examine the prediction capacity of pollutants for the developed probability-based DRASTIC model, medium, high, and very high risk categories of contamination potentials were compared with observed nitrate-N exceeding 0.5 mg/L indicating the anthropogenic groundwater pollution. The analyzed results reveal that the developed probability-based DRASTIC model is capable of predicting high nitrate-N groundwater pollution and characterizing the parameter uncertainty via the probability estimation processes. 3. Modeling the radiation transfer of discontinuous canopies: results for gap probability and single-scattering contribution Science.gov (United States) Zhao, Feng; Zou, Kai; Shang, Hong; Ji, Zheng; Zhao, Huijie; Huang, Wenjiang; Li, Cunjun 2010-10-01 In this paper we present an analytical model for the computation of radiation transfer of discontinuous vegetation canopies. Some initial results of gap probability and bidirectional gap probability of discontinuous vegetation canopies, which are important parameters determining the radiative environment of the canopies, are given and compared with a 3- D computer simulation model. In the model, negative exponential attenuation of light within individual plant canopies is assumed. Then the computation of gap probability is resolved by determining the entry points and exiting points of the ray with the individual plants via their equations in space. For the bidirectional gap probability, which determines the single-scattering contribution of the canopy, a gap statistical analysis based model was adopted to correct the dependence of gap probabilities for both solar and viewing directions. The model incorporates the structural characteristics, such as plant sizes, leaf size, row spacing, foliage density, planting density, leaf inclination distribution. Available experimental data are inadequate for a complete validation of the model. So it was evaluated with a three dimensional computer simulation model for 3D vegetative scenes, which shows good agreement between these two models' results. This model should be useful to the quantification of light interception and the modeling of bidirectional reflectance distributions of discontinuous canopies. 4. Quantification of a decision-making failure probability of the accident management using cognitive analysis model International Nuclear Information System (INIS) Yoshida, Yoshitaka; Ohtani, Masanori; Fujita, Yushi 2002-01-01 In the nuclear power plant, much knowledge is acquired through probabilistic safety assessment (PSA) of a severe accident, and accident management (AM) is prepared. It is necessary to evaluate the effectiveness of AM using the decision-making failure probability of an emergency organization, operation failure probability of operators, success criteria of AM and reliability of AM equipments in PSA. However, there has been no suitable qualification method for PSA so far to obtain the decision-making failure probability, because the decision-making failure of an emergency organization treats the knowledge based error. In this work, we developed a new method for quantification of the decision-making failure probability of an emergency organization using cognitive analysis model, which decided an AM strategy, in a nuclear power plant at the severe accident, and tried to apply it to a typical pressurized water reactor (PWR) plant. As a result: (1) It could quantify the decision-making failure probability adjusted to PSA for general analysts, who do not necessarily possess professional human factors knowledge, by choosing the suitable value of a basic failure probability and an error-factor. (2) The decision-making failure probabilities of six AMs were in the range of 0.23 to 0.41 using the screening evaluation method and in the range of 0.10 to 0.19 using the detailed evaluation method as the result of trial evaluation based on severe accident analysis of a typical PWR plant, and a result of sensitivity analysis of the conservative assumption, failure probability decreased about 50%. (3) The failure probability using the screening evaluation method exceeded that using detailed evaluation method by 99% of probability theoretically, and the failure probability of AM in this study exceeded 100%. From this result, it was shown that the decision-making failure probability was more conservative than the detailed evaluation method, and the screening evaluation method satisfied 5. Quantification of a decision-making failure probability of the accident management using cognitive analysis model Energy Technology Data Exchange (ETDEWEB) Yoshida, Yoshitaka; Ohtani, Masanori [Institute of Nuclear Safety System, Inc., Mihama, Fukui (Japan); Fujita, Yushi [TECNOVA Corp., Tokyo (Japan) 2002-09-01 In the nuclear power plant, much knowledge is acquired through probabilistic safety assessment (PSA) of a severe accident, and accident management (AM) is prepared. It is necessary to evaluate the effectiveness of AM using the decision-making failure probability of an emergency organization, operation failure probability of operators, success criteria of AM and reliability of AM equipments in PSA. However, there has been no suitable qualification method for PSA so far to obtain the decision-making failure probability, because the decision-making failure of an emergency organization treats the knowledge based error. In this work, we developed a new method for quantification of the decision-making failure probability of an emergency organization using cognitive analysis model, which decided an AM strategy, in a nuclear power plant at the severe accident, and tried to apply it to a typical pressurized water reactor (PWR) plant. As a result: (1) It could quantify the decision-making failure probability adjusted to PSA for general analysts, who do not necessarily possess professional human factors knowledge, by choosing the suitable value of a basic failure probability and an error-factor. (2) The decision-making failure probabilities of six AMs were in the range of 0.23 to 0.41 using the screening evaluation method and in the range of 0.10 to 0.19 using the detailed evaluation method as the result of trial evaluation based on severe accident analysis of a typical PWR plant, and a result of sensitivity analysis of the conservative assumption, failure probability decreased about 50%. (3) The failure probability using the screening evaluation method exceeded that using detailed evaluation method by 99% of probability theoretically, and the failure probability of AM in this study exceeded 100%. From this result, it was shown that the decision-making failure probability was more conservative than the detailed evaluation method, and the screening evaluation method satisfied 6. Probability theory CERN Document Server Dorogovtsev, A Ya; Skorokhod, A V; Silvestrov, D S; Skorokhod, A V 1997-01-01 This book of problems is intended for students in pure and applied mathematics. There are problems in traditional areas of probability theory and problems in the theory of stochastic processes, which has wide applications in the theory of automatic control, queuing and reliability theories, and in many other modern science and engineering fields. Answers to most of the problems are given, and the book provides hints and solutions for more complicated problems. 7. Knock probability estimation through an in-cylinder temperature model with exogenous noise Science.gov (United States) Bares, P.; Selmanaj, D.; Guardiola, C.; Onder, C. 2018-01-01 This paper presents a new knock model which combines a deterministic knock model based on the in-cylinder temperature and an exogenous noise disturbing this temperature. The autoignition of the end-gas is modelled by an Arrhenius-like function and the knock probability is estimated by propagating a virtual error probability distribution. Results show that the random nature of knock can be explained by uncertainties at the in-cylinder temperature estimation. The model only has one parameter for calibration and thus can be easily adapted online. In order to reduce the measurement uncertainties associated with the air mass flow sensor, the trapped mass is derived from the in-cylinder pressure resonance, which improves the knock probability estimation and reduces the number of sensors needed for the model. A four stroke SI engine was used for model validation. By varying the intake temperature, the engine speed, the injected fuel mass, and the spark advance, specific tests were conducted, which furnished data with various knock intensities and probabilities. The new model is able to predict the knock probability within a sufficient range at various operating conditions. The trapped mass obtained by the acoustical model was compared in steady conditions by using a fuel balance and a lambda sensor and differences below 1 % were found. 8. Probability theory for 3-layer remote sensing radiative transfer model: univariate case. Science.gov (United States) Ben-David, Avishai; Davidson, Charles E 2012-04-23 A probability model for a 3-layer radiative transfer model (foreground layer, cloud layer, background layer, and an external source at the end of line of sight) has been developed. The 3-layer model is fundamentally important as the primary physical model in passive infrared remote sensing. The probability model is described by the Johnson family of distributions that are used as a fit for theoretically computed moments of the radiative transfer model. From the Johnson family we use the SU distribution that can address a wide range of skewness and kurtosis values (in addition to addressing the first two moments, mean and variance). In the limit, SU can also describe lognormal and normal distributions. With the probability model one can evaluate the potential for detecting a target (vapor cloud layer), the probability of observing thermal contrast, and evaluate performance (receiver operating characteristics curves) in clutter-noise limited scenarios. This is (to our knowledge) the first probability model for the 3-layer remote sensing geometry that treats all parameters as random variables and includes higher-order statistics. © 2012 Optical Society of America 9. Success probability orientated optimization model for resource allocation of the technological innovation multi-project system Institute of Scientific and Technical Information of China (English) Weixu Dai; Weiwei Wu; Bo Yu; Yunhao Zhu 2016-01-01 A success probability orientated optimization model for resource al ocation of the technological innovation multi-project system is studied. Based on the definition of the technological in-novation multi-project system, the leveling optimization of cost and success probability is set as the objective of resource al ocation. The cost function and the probability function of the optimization model are constructed. Then the objective function of the model is constructed and the solving process is explained. The model is applied to the resource al ocation of an enterprise’s technological innovation multi-project system. The results show that the pro-posed model is more effective in rational resource al ocation, and is more applicable in maximizing the utility of the technological innovation multi-project system. 10. Trending in Probability of Collision Measurements via a Bayesian Zero-Inflated Beta Mixed Model Science.gov (United States) Vallejo, Jonathon; Hejduk, Matt; Stamey, James 2015-01-01 We investigate the performance of a generalized linear mixed model in predicting the Probabilities of Collision (Pc) for conjunction events. Specifically, we apply this model to the log(sub 10) transformation of these probabilities and argue that this transformation yields values that can be considered bounded in practice. Additionally, this bounded random variable, after scaling, is zero-inflated. Consequently, we model these values using the zero-inflated Beta distribution, and utilize the Bayesian paradigm and the mixed model framework to borrow information from past and current events. This provides a natural way to model the data and provides a basis for answering questions of interest, such as what is the likelihood of observing a probability of collision equal to the effective value of zero on a subsequent observation. 11. A stochastic model for the probability of malaria extinction by mass drug administration. Science.gov (United States) Pemberton-Ross, Peter; Chitnis, Nakul; Pothin, Emilie; Smith, Thomas A 2017-09-18 Mass drug administration (MDA) has been proposed as an intervention to achieve local extinction of malaria. Although its effect on the reproduction number is short lived, extinction may subsequently occur in a small population due to stochastic fluctuations. This paper examines how the probability of stochastic extinction depends on population size, MDA coverage and the reproduction number under control, R c . A simple compartmental model is developed which is used to compute the probability of extinction using probability generating functions. The expected time to extinction in small populations after MDA for various scenarios in this model is calculated analytically. The results indicate that mass drug administration (Firstly, R c must be sustained at R c  95% to have a non-negligible probability of successful elimination. Stochastic fluctuations only significantly affect the probability of extinction in populations of about 1000 individuals or less. The expected time to extinction via stochastic fluctuation is less than 10 years only in populations less than about 150 individuals. Clustering of secondary infections and of MDA distribution both contribute positively to the potential probability of success, indicating that MDA would most effectively be administered at the household level. There are very limited circumstances in which MDA will lead to local malaria elimination with a substantial probability. 12. Predicting critical transitions in dynamical systems from time series using nonstationary probability density modeling. Science.gov (United States) Kwasniok, Frank 2013-11-01 A time series analysis method for predicting the probability density of a dynamical system is proposed. A nonstationary parametric model of the probability density is estimated from data within a maximum likelihood framework and then extrapolated to forecast the future probability density and explore the system for critical transitions or tipping points. A full systematic account of parameter uncertainty is taken. The technique is generic, independent of the underlying dynamics of the system. The method is verified on simulated data and then applied to prediction of Arctic sea-ice extent. 13. Height probabilities in the Abelian sandpile model on the generalized finite Bethe lattice Science.gov (United States) Chen, Haiyan; Zhang, Fuji 2013-08-01 In this paper, we study the sandpile model on the generalized finite Bethe lattice with a particular boundary condition. Using a combinatorial method, we give the exact expressions for all single-site probabilities and some two-site joint probabilities. As a by-product, we prove that the height probabilities of bulk vertices are all the same for the Bethe lattice with certain given boundary condition, which was found from numerical evidence by Grassberger and Manna ["Some more sandpiles," J. Phys. (France) 51, 1077-1098 (1990)], 10.1051/jphys:0199000510110107700 but without a proof. 14. Modeling Stochastic Complexity in Complex Adaptive Systems: Non-Kolmogorov Probability and the Process Algebra Approach. Science.gov (United States) Sulis, William H 2017-10-01 Walter Freeman III pioneered the application of nonlinear dynamical systems theories and methodologies in his work on mesoscopic brain dynamics.Sadly, mainstream psychology and psychiatry still cling to linear correlation based data analysis techniques, which threaten to subvert the process of experimentation and theory building. In order to progress, it is necessary to develop tools capable of managing the stochastic complexity of complex biopsychosocial systems, which includes multilevel feedback relationships, nonlinear interactions, chaotic dynamics and adaptability. In addition, however, these systems exhibit intrinsic randomness, non-Gaussian probability distributions, non-stationarity, contextuality, and non-Kolmogorov probabilities, as well as the absence of mean and/or variance and conditional probabilities. These properties and their implications for statistical analysis are discussed. An alternative approach, the Process Algebra approach, is described. It is a generative model, capable of generating non-Kolmogorov probabilities. It has proven useful in addressing fundamental problems in quantum mechanics and in the modeling of developing psychosocial systems. 15. Probability Distribution and Deviation Information Fusion Driven Support Vector Regression Model and Its Application Directory of Open Access Journals (Sweden) Changhao Fan 2017-01-01 Full Text Available In modeling, only information from the deviation between the output of the support vector regression (SVR model and the training sample is considered, whereas the other prior information of the training sample, such as probability distribution information, is ignored. Probabilistic distribution information describes the overall distribution of sample data in a training sample that contains different degrees of noise and potential outliers, as well as helping develop a high-accuracy model. To mine and use the probability distribution information of a training sample, a new support vector regression model that incorporates probability distribution information weight SVR (PDISVR is proposed. In the PDISVR model, the probability distribution of each sample is considered as the weight and is then introduced into the error coefficient and slack variables of SVR. Thus, the deviation and probability distribution information of the training sample are both used in the PDISVR model to eliminate the influence of noise and outliers in the training sample and to improve predictive performance. Furthermore, examples with different degrees of noise were employed to demonstrate the performance of PDISVR, which was then compared with those of three SVR-based methods. The results showed that PDISVR performs better than the three other methods. 16. Estimation and prediction of maximum daily rainfall at Sagar Island using best fit probability models Science.gov (United States) Mandal, S.; Choudhury, B. U. 2015-07-01 Sagar Island, setting on the continental shelf of Bay of Bengal, is one of the most vulnerable deltas to the occurrence of extreme rainfall-driven climatic hazards. Information on probability of occurrence of maximum daily rainfall will be useful in devising risk management for sustaining rainfed agrarian economy vis-a-vis food and livelihood security. Using six probability distribution models and long-term (1982-2010) daily rainfall data, we studied the probability of occurrence of annual, seasonal and monthly maximum daily rainfall (MDR) in the island. To select the best fit distribution models for annual, seasonal and monthly time series based on maximum rank with minimum value of test statistics, three statistical goodness of fit tests, viz. Kolmogorove-Smirnov test (K-S), Anderson Darling test ( A 2 ) and Chi-Square test ( X 2) were employed. The fourth probability distribution was identified from the highest overall score obtained from the three goodness of fit tests. Results revealed that normal probability distribution was best fitted for annual, post-monsoon and summer seasons MDR, while Lognormal, Weibull and Pearson 5 were best fitted for pre-monsoon, monsoon and winter seasons, respectively. The estimated annual MDR were 50, 69, 86, 106 and 114 mm for return periods of 2, 5, 10, 20 and 25 years, respectively. The probability of getting an annual MDR of >50, >100, >150, >200 and >250 mm were estimated as 99, 85, 40, 12 and 03 % level of exceedance, respectively. The monsoon, summer and winter seasons exhibited comparatively higher probabilities (78 to 85 %) for MDR of >100 mm and moderate probabilities (37 to 46 %) for >150 mm. For different recurrence intervals, the percent probability of MDR varied widely across intra- and inter-annual periods. In the island, rainfall anomaly can pose a climatic threat to the sustainability of agricultural production and thus needs adequate adaptation and mitigation measures. 17. Quantum probability and cognitive modeling: some cautions and a promising direction in modeling physics learning. Science.gov (United States) Franceschetti, Donald R; Gire, Elizabeth 2013-06-01 Quantum probability theory offers a viable alternative to classical probability, although there are some ambiguities inherent in transferring the quantum formalism to a less determined realm. A number of physicists are now looking at the applicability of quantum ideas to the assessment of physics learning, an area particularly suited to quantum probability ideas. 18. Estimating the Probability of Vegetation to Be Groundwater Dependent Based on the Evaluation of Tree Models Directory of Open Access Journals (Sweden) Isabel C. Pérez Hoyos 2016-04-01 Full Text Available Groundwater Dependent Ecosystems (GDEs are increasingly threatened by humans’ rising demand for water resources. Consequently, it is imperative to identify the location of GDEs to protect them. This paper develops a methodology to identify the probability of an ecosystem to be groundwater dependent. Probabilities are obtained by modeling the relationship between the known locations of GDEs and factors influencing groundwater dependence, namely water table depth and climatic aridity index. Probabilities are derived for the state of Nevada, USA, using modeled water table depth and aridity index values obtained from the Global Aridity database. The model selected results from the performance comparison of classification trees (CT and random forests (RF. Based on a threshold-independent accuracy measure, RF has a better ability to generate probability estimates. Considering a threshold that minimizes the misclassification rate for each model, RF also proves to be more accurate. Regarding training accuracy, performance measures such as accuracy, sensitivity, and specificity are higher for RF. For the test set, higher values of accuracy and kappa for CT highlight the fact that these measures are greatly affected by low prevalence. As shown for RF, the choice of the cutoff probability value has important consequences on model accuracy and the overall proportion of locations where GDEs are found. 19. Uncovering the Best Skill Multimap by Constraining the Error Probabilities of the Gain-Loss Model Science.gov (United States) Anselmi, Pasquale; Robusto, Egidio; Stefanutti, Luca 2012-01-01 The Gain-Loss model is a probabilistic skill multimap model for assessing learning processes. In practical applications, more than one skill multimap could be plausible, while none corresponds to the true one. The article investigates whether constraining the error probabilities is a way of uncovering the best skill assignment among a number of… 20. Aggregate and Individual Replication Probability within an Explicit Model of the Research Process Science.gov (United States) Miller, Jeff; Schwarz, Wolf 2011-01-01 We study a model of the research process in which the true effect size, the replication jitter due to changes in experimental procedure, and the statistical error of effect size measurement are all normally distributed random variables. Within this model, we analyze the probability of successfully replicating an initial experimental result by… 1. A Taxonomy of Latent Structure Assumptions for Probability Matrix Decomposition Models. Science.gov (United States) Meulders, Michel; De Boeck, Paul; Van Mechelen, Iven 2003-01-01 Proposed a taxonomy of latent structure assumptions for probability matrix decomposition (PMD) that includes the original PMD model and a three-way extension of the multiple classification latent class model. Simulation study results show the usefulness of the taxonomy. (SLD) 2. Transition probabilities of health states for workers in Malaysia using a Markov chain model Science.gov (United States) Samsuddin, Shamshimah; Ismail, Noriszura 2017-04-01 The aim of our study is to estimate the transition probabilities of health states for workers in Malaysia who contribute to the Employment Injury Scheme under the Social Security Organization Malaysia using the Markov chain model. Our study uses four states of health (active, temporary disability, permanent disability and death) based on the data collected from the longitudinal studies of workers in Malaysia for 5 years. The transition probabilities vary by health state, age and gender. The results show that men employees are more likely to have higher transition probabilities to any health state compared to women employees. The transition probabilities can be used to predict the future health of workers in terms of a function of current age, gender and health state. 3. The probabilities of one- and multi-track events for modeling radiation-induced cell kill Energy Technology Data Exchange (ETDEWEB) Schneider, Uwe; Vasi, Fabiano; Besserer, Juergen [University of Zuerich, Department of Physics, Science Faculty, Zurich (Switzerland); Radiotherapy Hirslanden, Zurich (Switzerland) 2017-08-15 In view of the clinical importance of hypofractionated radiotherapy, track models which are based on multi-hit events are currently reinvestigated. These models are often criticized, because it is believed that the probability of multi-track hits is negligible. In this work, the probabilities for one- and multi-track events are determined for different biological targets. The obtained probabilities can be used with nano-dosimetric cluster size distributions to obtain the parameters of track models. We quantitatively determined the probabilities for one- and multi-track events for 100, 500 and 1000 keV electrons, respectively. It is assumed that the single tracks are statistically independent and follow a Poisson distribution. Three different biological targets were investigated: (1) a DNA strand (2 nm scale); (2) two adjacent chromatin fibers (60 nm); and (3) fiber loops (300 nm). It was shown that the probabilities for one- and multi-track events are increasing with energy, size of the sensitive target structure, and dose. For a 2 x 2 x 2 nm{sup 3} target, one-track events are around 10,000 times more frequent than multi-track events. If the size of the sensitive structure is increased to 100-300 nm, the probabilities for one- and multi-track events are of the same order of magnitude. It was shown that target theories can play a role for describing radiation-induced cell death if the targets are of the size of two adjacent chromatin fibers or fiber loops. The obtained probabilities can be used together with the nano-dosimetric cluster size distributions to determine model parameters for target theories. (orig.) 4. Dissolution Model Development: Formulation Effects and Filter Complications DEFF Research Database (Denmark) Berthelsen, Ragna; Holm, Rene; Jacobsen, Jette 2016-01-01 This study describes various complications related to sample preparation (filtration) during development of a dissolution method intended to discriminate among different fenofibrate immediate-release formulations. Several dissolution apparatus and sample preparation techniques were tested. The fl....... With the tested drug–formulation combination, the best in vivo–in vitro correlation was found after filtration of the dissolution samples through 0.45-μm hydrophobic PTFE membrane filters.... 5. The effect of coupling hydrologic and hydrodynamic models on probable maximum flood estimation Science.gov (United States) Felder, Guido; Zischg, Andreas; Weingartner, Rolf 2017-07-01 Deterministic rainfall-runoff modelling usually assumes stationary hydrological system, as model parameters are calibrated with and therefore dependant on observed data. However, runoff processes are probably not stationary in the case of a probable maximum flood (PMF) where discharge greatly exceeds observed flood peaks. Developing hydrodynamic models and using them to build coupled hydrologic-hydrodynamic models can potentially improve the plausibility of PMF estimations. This study aims to assess the potential benefits and constraints of coupled modelling compared to standard deterministic hydrologic modelling when it comes to PMF estimation. The two modelling approaches are applied using a set of 100 spatio-temporal probable maximum precipitation (PMP) distribution scenarios. The resulting hydrographs, the resulting peak discharges as well as the reliability and the plausibility of the estimates are evaluated. The discussion of the results shows that coupling hydrologic and hydrodynamic models substantially improves the physical plausibility of PMF modelling, although both modelling approaches lead to PMF estimations for the catchment outlet that fall within a similar range. Using a coupled model is particularly suggested in cases where considerable flood-prone areas are situated within a catchment. 6. On Probability Leakage OpenAIRE Briggs, William M. 2012-01-01 The probability leakage of model M with respect to evidence E is defined. Probability leakage is a kind of model error. It occurs when M implies that events $y$, which are impossible given E, have positive probability. Leakage does not imply model falsification. Models with probability leakage cannot be calibrated empirically. Regression models, which are ubiquitous in statistical practice, often evince probability leakage. 7. Exact results for survival probability in the multistate Landau-Zener model International Nuclear Information System (INIS) Volkov, M V; Ostrovsky, V N 2004-01-01 An exact formula is derived for survival probability in the multistate Landau-Zener model in the special case where the initially populated state corresponds to the extremal (maximum or minimum) slope of a linear diabatic potential curve. The formula was originally guessed by S Brundobler and V Elzer (1993 J. Phys. A: Math. Gen. 26 1211) based on numerical calculations. It is a simple generalization of the expression for the probability of diabatic passage in the famous two-state Landau-Zener model. Our result is obtained via analysis and summation of the entire perturbation theory series 8. Modelling the Probability Density Function of IPTV Traffic Packet Delay Variation Directory of Open Access Journals (Sweden) Michal Halas 2012-01-01 Full Text Available This article deals with modelling the Probability density function of IPTV traffic packet delay variation. The use of this modelling is in an efficient de-jitter buffer estimation. When an IP packet travels across a network, it experiences delay and its variation. This variation is caused by routing, queueing systems and other influences like the processing delay of the network nodes. When we try to separate these at least three types of delay variation, we need a way to measure these types separately. This work is aimed to the delay variation caused by queueing systems which has the main implications to the form of the Probability density function. 9. An Empirical Model for Estimating the Probability of Electrical Short Circuits from Tin Whiskers. Part 2 Science.gov (United States) Courey, Karim; Wright, Clara; Asfour, Shihab; Onar, Arzu; Bayliss, Jon; Ludwig, Larry 2009-01-01 In this experiment, an empirical model to quantify the probability of occurrence of an electrical short circuit from tin whiskers as a function of voltage was developed. This empirical model can be used to improve existing risk simulation models. FIB and TEM images of a tin whisker confirm the rare polycrystalline structure on one of the three whiskers studied. FIB cross-section of the card guides verified that the tin finish was bright tin. 10. Use of the AIC with the EM algorithm: A demonstration of a probability model selection technique Energy Technology Data Exchange (ETDEWEB) Glosup, J.G.; Axelrod M.C. [Lawrence Livermore National Lab., CA (United States) 1994-11-15 The problem of discriminating between two potential probability models, a Gaussian distribution and a mixture of Gaussian distributions, is considered. The focus of our interest is a case where the models are potentially non-nested and the parameters of the mixture model are estimated through the EM algorithm. The AIC, which is frequently used as a criterion for discriminating between non-nested models, is modified to work with the EM algorithm and is shown to provide a model selection tool for this situation. A particular problem involving an infinite mixture distribution known as Middletons Class A model is used to demonstrate the effectiveness and limitations of this method. 11. Application of damping mechanism model and stacking fault probability in Fe-Mn alloy International Nuclear Information System (INIS) Huang, S.K.; Wen, Y.H.; Li, N.; Teng, J.; Ding, S.; Xu, Y.G. 2008-01-01 In this paper, the damping mechanism model of Fe-Mn alloy was analyzed using dislocation theory. Moreover, as an important parameter in Fe-Mn based alloy, the effect of stacking fault probability on the damping capacity of Fe-19.35Mn alloy after deep-cooling or tensile deformation was also studied. The damping capacity was measured using reversal torsion pendulum. The stacking fault probability of γ-austenite and ε-martensite was determined by means of X-ray diffraction (XRD) profile analysis. The microstructure was observed using scanning electronic microscope (SEM). The results indicated that with the strain amplitude increasing above a critical value, the damping capacity of Fe-19.35Mn alloy increased rapidly which could be explained using the breakaway model of Shockley partial dislocations. Deep-cooling and suitable tensile deformation could improve the damping capacity owning to the increasing of stacking fault probability of Fe-19.35Mn alloy 12. An extended car-following model considering random safety distance with different probabilities Science.gov (United States) Wang, Jufeng; Sun, Fengxin; Cheng, Rongjun; Ge, Hongxia; Wei, Qi 2018-02-01 Because of the difference in vehicle type or driving skill, the driving strategy is not exactly the same. The driving speeds of the different vehicles may be different for the same headway. Since the optimal velocity function is just determined by the safety distance besides the maximum velocity and headway, an extended car-following model accounting for random safety distance with different probabilities is proposed in this paper. The linear stable condition for this extended traffic model is obtained by using linear stability theory. Numerical simulations are carried out to explore the complex phenomenon resulting from multiple safety distance in the optimal velocity function. The cases of multiple types of safety distances selected with different probabilities are presented. Numerical results show that the traffic flow with multiple safety distances with different probabilities will be more unstable than that with single type of safety distance, and will result in more stop-and-go phenomena. 13. Probability of atrial fibrillation after ablation: Using a parametric nonlinear temporal decomposition mixed effects model. Science.gov (United States) Rajeswaran, Jeevanantham; Blackstone, Eugene H; Ehrlinger, John; Li, Liang; Ishwaran, Hemant; Parides, Michael K 2018-01-01 Atrial fibrillation is an arrhythmic disorder where the electrical signals of the heart become irregular. The probability of atrial fibrillation (binary response) is often time varying in a structured fashion, as is the influence of associated risk factors. A generalized nonlinear mixed effects model is presented to estimate the time-related probability of atrial fibrillation using a temporal decomposition approach to reveal the pattern of the probability of atrial fibrillation and their determinants. This methodology generalizes to patient-specific analysis of longitudinal binary data with possibly time-varying effects of covariates and with different patient-specific random effects influencing different temporal phases. The motivation and application of this model is illustrated using longitudinally measured atrial fibrillation data obtained through weekly trans-telephonic monitoring from an NIH sponsored clinical trial being conducted by the Cardiothoracic Surgery Clinical Trials Network. 14. Assessment of different models for computing the probability of a clear line of sight Science.gov (United States) Bojin, Sorin; Paulescu, Marius; Badescu, Viorel 2017-12-01 This paper is focused on modeling the morphological properties of the cloud fields in terms of the probability of a clear line of sight (PCLOS). PCLOS is defined as the probability that a line of sight between observer and a given point of the celestial vault goes freely without intersecting a cloud. A variety of PCLOS models assuming the cloud shape hemisphere, semi-ellipsoid and ellipsoid are tested. The effective parameters (cloud aspect ratio and absolute cloud fraction) are extracted from high-resolution series of sunshine number measurements. The performance of the PCLOS models is evaluated from the perspective of their ability in retrieving the point cloudiness. The advantages and disadvantages of the tested models are discussed, aiming to a simplified parameterization of PCLOS models. 15. Modelling detection probabilities to evaluate management and control tools for an invasive species Science.gov (United States) Christy, M.T.; Yackel Adams, A.A.; Rodda, G.H.; Savidge, J.A.; Tyrrell, C.L. 2010-01-01 For most ecologists, detection probability (p) is a nuisance variable that must be modelled to estimate the state variable of interest (i.e. survival, abundance, or occupancy). However, in the realm of invasive species control, the rate of detection and removal is the rate-limiting step for management of this pervasive environmental problem. For strategic planning of an eradication (removal of every individual), one must identify the least likely individual to be removed, and determine the probability of removing it. To evaluate visual searching as a control tool for populations of the invasive brown treesnake Boiga irregularis, we designed a mark-recapture study to evaluate detection probability as a function of time, gender, size, body condition, recent detection history, residency status, searcher team and environmental covariates. We evaluated these factors using 654 captures resulting from visual detections of 117 snakes residing in a 5-ha semi-forested enclosure on Guam, fenced to prevent immigration and emigration of snakes but not their prey. Visual detection probability was low overall (= 0??07 per occasion) but reached 0??18 under optimal circumstances. Our results supported sex-specific differences in detectability that were a quadratic function of size, with both small and large females having lower detection probabilities than males of those sizes. There was strong evidence for individual periodic changes in detectability of a few days duration, roughly doubling detection probability (comparing peak to non-elevated detections). Snakes in poor body condition had estimated mean detection probabilities greater than snakes with high body condition. Search teams with high average detection rates exhibited detection probabilities about twice that of search teams with low average detection rates. Surveys conducted with bright moonlight and strong wind gusts exhibited moderately decreased probabilities of detecting snakes. Synthesis and applications. By 16. Modeling co-occurrence of northern spotted and barred owls: accounting for detection probability differences Science.gov (United States) Bailey, Larissa L.; Reid, Janice A.; Forsman, Eric D.; Nichols, James D. 2009-01-01 Barred owls (Strix varia) have recently expanded their range and now encompass the entire range of the northern spotted owl (Strix occidentalis caurina). This expansion has led to two important issues of concern for management of northern spotted owls: (1) possible competitive interactions between the two species that could contribute to population declines of northern spotted owls, and (2) possible changes in vocalization behavior and detection probabilities of northern spotted owls induced by presence of barred owls. We used a two-species occupancy model to investigate whether there was evidence of competitive exclusion between the two species at study locations in Oregon, USA. We simultaneously estimated detection probabilities for both species and determined if the presence of one species influenced the detection of the other species. Model selection results and associated parameter estimates provided no evidence that barred owls excluded spotted owls from territories. We found strong evidence that detection probabilities differed for the two species, with higher probabilities for northern spotted owls that are the object of current surveys. Non-detection of barred owls is very common in surveys for northern spotted owls, and detection of both owl species was negatively influenced by the presence of the congeneric species. Our results suggest that analyses directed at hypotheses of barred owl effects on demographic or occupancy vital rates of northern spotted owls need to deal adequately with imperfect and variable detection probabilities for both species. 17. Ruin probabilities DEFF Research Database (Denmark) Asmussen, Søren; Albrecher, Hansjörg The book gives a comprehensive treatment of the classical and modern ruin probability theory. Some of the topics are Lundberg's inequality, the Cramér-Lundberg approximation, exact solutions, other approximations (e.g., for heavy-tailed claim size distributions), finite horizon ruin probabilities......, extensions of the classical compound Poisson model to allow for reserve-dependent premiums, Markov-modulation, periodicity, change of measure techniques, phase-type distributions as a computational vehicle and the connection to other applied probability areas, like queueing theory. In this substantially...... updated and extended second version, new topics include stochastic control, fluctuation theory for Levy processes, Gerber–Shiu functions and dependence.... 18. Blind Students' Learning of Probability through the Use of a Tactile Model Science.gov (United States) Vita, Aida Carvalho; Kataoka, Verônica Yumi 2014-01-01 The objective of this paper is to discuss how blind students learn basic concepts of probability using the tactile model proposed by Vita (2012). Among the activities were part of the teaching sequence "Jefferson's Random Walk", in which students built a tree diagram (using plastic trays, foam cards, and toys), and pictograms in 3D… 19. Modeling tumor control probability for spatially inhomogeneous risk of failure based on clinical outcome data DEFF Research Database (Denmark) Lühr, Armin; Löck, Steffen; Jakobi, Annika 2017-01-01 PURPOSE: Objectives of this work are (1) to derive a general clinically relevant approach to model tumor control probability (TCP) for spatially variable risk of failure and (2) to demonstrate its applicability by estimating TCP for patients planned for photon and proton irradiation. METHODS AND ... 20. On the Probability of Occurrence of Clusters in Abelian Sandpile Model OpenAIRE 2004-01-01 We have performed extensive simulations on the Abelian Sandpile Model (ASM) on square lattice. We have estimated the probability of observation of many clusters. Some are in good agreement with previous analytical results, while some show discrepancies between simulation and analytical results. 1. Developing a Model and Applications for Probabilities of Student Success: A Case Study of Predictive Analytics Science.gov (United States) Calvert, Carol Elaine 2014-01-01 This case study relates to distance learning students on open access courses. It demonstrates the use of predictive analytics to generate a model of the probabilities of success and retention at different points, or milestones, in a student journey. A core set of explanatory variables has been established and their varying relative importance at… 2. A new formulation of the probability density function in random walk models for atmospheric dispersion DEFF Research Database (Denmark) Falk, Anne Katrine Vinther; Gryning, Sven-Erik 1997-01-01 In this model for atmospheric dispersion particles are simulated by the Langevin Equation, which is a stochastic differential equation. It uses the probability density function (PDF) of the vertical velocity fluctuations as input. The PDF is constructed as an expansion after Hermite polynomials... 3. Modeling Perceived Quality, Customer Satisfaction and Probability of Guest Returning to the Destination Directory of Open Access Journals (Sweden) Olivera Blagojevic Popovic 2018-03-01 Full Text Available In the hotel industry, it is a well-known fact that, despite of quality and variety of services provided, there is a low probability that the guests will return. This research is focused on identifying the basic factors of the hotel offer, which could determine the influence on the correlation between the guests’ satisfaction and the probability of their return. The objective of the article is to explore the relationship between the guests’ satisfaction with the quality hotel services in total (including the tourist offer of the place and the probability of his return to the same destination. The questionnaire method was applied in the survey, and the data were analysed based on factor analysis. Thereafter, the model for forecasting the probability of the guests returning to the destination was established, by using the example of Montenegrin tourism. The model represents a defined framework for the guest’s decision-making process. It identifies two main characteristics of guest experiences: satisfaction and rated quality (of the destination’s overall hotel service and tourist offer. The same model evaluates the impact of the above factors on the probability of the guests’ returning to the same destination. The starting hypothesis was the existence of a high degree of correlation between the guests’ satisfaction (with the destination’s hotel services and tourist offer and the probability of returning to the selected Montenegrin destinations. The research confirmed the above-mentioned hypothesis. The results have revealed that there are significant differences in perceived quality, i.e. satisfaction between the target groups of Eastern and Western European tourists 4. On new cautious structural reliability models in the framework of imprecise probabilities DEFF Research Database (Denmark) Utkin, Lev; Kozine, Igor 2010-01-01 measures when the number of events of interest or observations is very small. The main feature of the models is that prior ignorance is not modelled by a fixed single prior distribution, but by a class of priors which is defined by upper and lower probabilities that can converge as statistical data......New imprecise structural reliability models are described in this paper. They are developed based on the imprecise Bayesian inference and are imprecise Dirichlet, imprecise negative binomial, gamma-exponential and normal models. The models are applied to computing cautious structural reliability... 5. Predicting Flow Breakdown Probability and Duration in Stochastic Network Models: Impact on Travel Time Reliability Energy Technology Data Exchange (ETDEWEB) Dong, Jing [ORNL; Mahmassani, Hani S. [Northwestern University, Evanston 2011-01-01 This paper proposes a methodology to produce random flow breakdown endogenously in a mesoscopic operational model, by capturing breakdown probability and duration. Based on previous research findings that probability of flow breakdown can be represented as a function of flow rate and the duration can be characterized by a hazard model. By generating random flow breakdown at various levels and capturing the traffic characteristics at the onset of the breakdown, the stochastic network simulation model provides a tool for evaluating travel time variability. The proposed model can be used for (1) providing reliability related traveler information; (2) designing ITS (intelligent transportation systems) strategies to improve reliability; and (3) evaluating reliability-related performance measures of the system. 6. Probability density function modeling of scalar mixing from concentrated sources in turbulent channel flow OpenAIRE Bakosi, J.; Franzese, P.; Boybeyi, Z. 2010-01-01 Dispersion of a passive scalar from concentrated sources in fully developed turbulent channel flow is studied with the probability density function (PDF) method. The joint PDF of velocity, turbulent frequency and scalar concentration is represented by a large number of Lagrangian particles. A stochastic near-wall PDF model combines the generalized Langevin model of Haworth & Pope with Durbin's method of elliptic relaxation to provide a mathematically exact treatment of convective and viscous ... 7. A stochastic-bayesian model for the fracture probability of PWR pressure vessels Energy Technology Data Exchange (ETDEWEB) Francisco, Alexandre S.; Duran, Jorge Alberto R., E-mail: afrancisco@metal.eeimvr.uff.br, E-mail: duran@metal.eeimvr.uff.br [Universidade Federal Fluminense (UFF), Volta Redonda, RJ (Brazil). Dept. de Engenharia Mecanica 2013-07-01 Fracture probability of pressure vessels containing cracks can be obtained by methodologies of easy understanding, which require a deterministic treatment, complemented by statistical methods. However, more accurate results are required, methodologies need to be better formulated. This paper presents a new methodology to address this problem. First, a more rigorous methodology is obtained by means of the relationship of probability distributions that model crack incidence and nondestructive inspection efficiency using the Bayes' theorem. The result is an updated crack incidence distribution. Further, the accuracy of the methodology is improved by using a stochastic model for the crack growth. The stochastic model incorporates the statistical variability of the crack growth process, combining the stochastic theory with experimental data. Stochastic differential equations are derived by the randomization of empirical equations. From the solution of this equation, a distribution function related to the crack growth is derived. The fracture probability using both probability distribution functions is in agreement with theory, and presents realistic value for pressure vessels. (author) 8. A stochastic-bayesian model for the fracture probability of PWR pressure vessels International Nuclear Information System (INIS) Francisco, Alexandre S.; Duran, Jorge Alberto R. 2013-01-01 Fracture probability of pressure vessels containing cracks can be obtained by methodologies of easy understanding, which require a deterministic treatment, complemented by statistical methods. However, more accurate results are required, methodologies need to be better formulated. This paper presents a new methodology to address this problem. First, a more rigorous methodology is obtained by means of the relationship of probability distributions that model crack incidence and nondestructive inspection efficiency using the Bayes' theorem. The result is an updated crack incidence distribution. Further, the accuracy of the methodology is improved by using a stochastic model for the crack growth. The stochastic model incorporates the statistical variability of the crack growth process, combining the stochastic theory with experimental data. Stochastic differential equations are derived by the randomization of empirical equations. From the solution of this equation, a distribution function related to the crack growth is derived. The fracture probability using both probability distribution functions is in agreement with theory, and presents realistic value for pressure vessels. (author) 9. Mental models of women with breast implants : local complications NARCIS (Netherlands) Byram, S.; Fischhoff, B.; Embrey, M.; Bruine de Bruin, W.J.A.; Thorne, S. 2001-01-01 Twenty-five women with breast implants participated in semistructured interviews designed to reveal their "mental models" of the processes potentially causing local (ie, nonsystemic) problems. The authors analyzed their responses in terms of an "expert model," circumscribing scientifically relevant 10. Mathematical modelling of air-flow in geometrically complicated areas OpenAIRE 2014-01-01 The Charles University Large-eddy Microscale Model (CLMM) and its application are presented. It is a numerical model for simulation of turbulent flow and dispersion in the planetary boundary layer. CLMM solves the incompressible Navier-Stokes equations in the Boussinesq approximation and describes turbulence using the large eddy simulation. Three applications of the model are presented. In the first case, the model is applied to the stable boundary layer over a flat terrain. The second case p... 11. Probability of Detection (POD) as a statistical model for the validation of qualitative methods. Science.gov (United States) Wehling, Paul; LaBudde, Robert A; Brunelle, Sharon L; Nelson, Maria T 2011-01-01 A statistical model is presented for use in validation of qualitative methods. This model, termed Probability of Detection (POD), harmonizes the statistical concepts and parameters between quantitative and qualitative method validation. POD characterizes method response with respect to concentration as a continuous variable. The POD model provides a tool for graphical representation of response curves for qualitative methods. In addition, the model allows comparisons between candidate and reference methods, and provides calculations of repeatability, reproducibility, and laboratory effects from collaborative study data. Single laboratory study and collaborative study examples are given. 12. Dependence in probabilistic modeling Dempster-Shafer theory and probability bounds analysis Energy Technology Data Exchange (ETDEWEB) Ferson, Scott [Applied Biomathematics, Setauket, NY (United States); Nelsen, Roger B. [Lewis & Clark College, Portland OR (United States); Hajagos, Janos [Applied Biomathematics, Setauket, NY (United States); Berleant, Daniel J. [Iowa State Univ., Ames, IA (United States); Zhang, Jianzhong [Iowa State Univ., Ames, IA (United States); Tucker, W. Troy [Applied Biomathematics, Setauket, NY (United States); Ginzburg, Lev R. [Applied Biomathematics, Setauket, NY (United States); Oberkampf, William L. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States) 2015-05-01 This report summarizes methods to incorporate information (or lack of information) about inter-variable dependence into risk assessments that use Dempster-Shafer theory or probability bounds analysis to address epistemic and aleatory uncertainty. The report reviews techniques for simulating correlated variates for a given correlation measure and dependence model, computation of bounds on distribution functions under a specified dependence model, formulation of parametric and empirical dependence models, and bounding approaches that can be used when information about the intervariable dependence is incomplete. The report also reviews several of the most pervasive and dangerous myths among risk analysts about dependence in probabilistic models. 13. On New Cautious Structural Reliability Models in the Framework of imprecise Probabilities DEFF Research Database (Denmark) Utkin, Lev V.; Kozine, Igor 2010-01-01 models and gen-eralizing conventional ones to imprecise probabili-ties. The theoretical setup employed for this purpose is imprecise statistical reasoning (Walley 1991), whose general framework is provided by upper and lower previsions (expectations). The appeal of this theory is its ability to capture......Uncertainty of parameters in engineering design has been modeled in different frameworks such as inter-val analysis, fuzzy set and possibility theories, ran-dom set theory and imprecise probability theory. The authors of this paper for many years have been de-veloping new imprecise reliability...... both aleatory (stochas-tic) and epistemic uncertainty and the flexibility with which information can be represented. The previous research of the authors related to generalizing structural reliability models to impre-cise statistical measures is summarized in Utkin & Kozine (2002) and Utkin (2004... 14. The return period analysis of natural disasters with statistical modeling of bivariate joint probability distribution. Science.gov (United States) Li, Ning; Liu, Xueqin; Xie, Wei; Wu, Jidong; Zhang, Peng 2013-01-01 New features of natural disasters have been observed over the last several years. The factors that influence the disasters' formation mechanisms, regularity of occurrence and main characteristics have been revealed to be more complicated and diverse in nature than previously thought. As the uncertainty involved increases, the variables need to be examined further. This article discusses the importance and the shortage of multivariate analysis of natural disasters and presents a method to estimate the joint probability of the return periods and perform a risk analysis. Severe dust storms from 1990 to 2008 in Inner Mongolia were used as a case study to test this new methodology, as they are normal and recurring climatic phenomena on Earth. Based on the 79 investigated events and according to the dust storm definition with bivariate, the joint probability distribution of severe dust storms was established using the observed data of maximum wind speed and duration. The joint return periods of severe dust storms were calculated, and the relevant risk was analyzed according to the joint probability. The copula function is able to simulate severe dust storm disasters accurately. The joint return periods generated are closer to those observed in reality than the univariate return periods and thus have more value in severe dust storm disaster mitigation, strategy making, program design, and improvement of risk management. This research may prove useful in risk-based decision making. The exploration of multivariate analysis methods can also lay the foundation for further applications in natural disaster risk analysis. © 2012 Society for Risk Analysis. 15. Bayesian selection of misspecified models is overconfident and may cause spurious posterior probabilities for phylogenetic trees. Science.gov (United States) Yang, Ziheng; Zhu, Tianqi 2018-02-20 The Bayesian method is noted to produce spuriously high posterior probabilities for phylogenetic trees in analysis of large datasets, but the precise reasons for this overconfidence are unknown. In general, the performance of Bayesian selection of misspecified models is poorly understood, even though this is of great scientific interest since models are never true in real data analysis. Here we characterize the asymptotic behavior of Bayesian model selection and show that when the competing models are equally wrong, Bayesian model selection exhibits surprising and polarized behaviors in large datasets, supporting one model with full force while rejecting the others. If one model is slightly less wrong than the other, the less wrong model will eventually win when the amount of data increases, but the method may become overconfident before it becomes reliable. We suggest that this extreme behavior may be a major factor for the spuriously high posterior probabilities for evolutionary trees. The philosophical implications of our results to the application of Bayesian model selection to evaluate opposing scientific hypotheses are yet to be explored, as are the behaviors of non-Bayesian methods in similar situations. 16. Exploring the Subtleties of Inverse Probability Weighting and Marginal Structural Models. Science.gov (United States) Breskin, Alexander; Cole, Stephen R; Westreich, Daniel 2018-05-01 Since being introduced to epidemiology in 2000, marginal structural models have become a commonly used method for causal inference in a wide range of epidemiologic settings. In this brief report, we aim to explore three subtleties of marginal structural models. First, we distinguish marginal structural models from the inverse probability weighting estimator, and we emphasize that marginal structural models are not only for longitudinal exposures. Second, we explore the meaning of the word "marginal" in "marginal structural model." Finally, we show that the specification of a marginal structural model can have important implications for the interpretation of its parameters. Each of these concepts have important implications for the use and understanding of marginal structural models, and thus providing detailed explanations of them may lead to better practices for the field of epidemiology. 17. Scaling Qualitative Probability OpenAIRE Burgin, Mark 2017-01-01 There are different approaches to qualitative probability, which includes subjective probability. We developed a representation of qualitative probability based on relational systems, which allows modeling uncertainty by probability structures and is more coherent than existing approaches. This setting makes it possible proving that any comparative probability is induced by some probability structure (Theorem 2.1), that classical probability is a probability structure (Theorem 2.2) and that i... 18. Fishnet model for failure probability tail of nacre-like imbricated lamellar materials Science.gov (United States) Luo, Wen; Bažant, Zdeněk P. 2017-12-01 Nacre, the iridescent material of the shells of pearl oysters and abalone, consists mostly of aragonite (a form of CaCO3), a brittle constituent of relatively low strength (≈10 MPa). Yet it has astonishing mean tensile strength (≈150 MPa) and fracture energy (≈350 to 1,240 J/m2). The reasons have recently become well understood: (i) the nanoscale thickness (≈300 nm) of nacre's building blocks, the aragonite lamellae (or platelets), and (ii) the imbricated, or staggered, arrangement of these lamellea, bound by biopolymer layers only ≈25 nm thick, occupying engineering applications, however, the failure probability of ≤10-6 is generally required. To guarantee it, the type of probability density function (pdf) of strength, including its tail, must be determined. This objective, not pursued previously, is hardly achievable by experiments alone, since >10^8 tests of specimens would be needed. Here we outline a statistical model of strength that resembles a fishnet pulled diagonally, captures the tail of pdf of strength and, importantly, allows analytical safety assessments of nacreous materials. The analysis shows that, in terms of safety, the imbricated lamellar structure provides a major additional advantage—˜10% strength increase at tail failure probability 10^-6 and a 1 to 2 orders of magnitude tail probability decrease at fixed stress. Another advantage is that a high scatter of microstructure properties diminishes the strength difference between the mean and the probability tail, compared with the weakest link model. These advantages of nacre-like materials are here justified analytically and supported by millions of Monte Carlo simulations. 19. Probabilities and energies to obtain the counting efficiency of electron-capture nuclides, KLMN model International Nuclear Information System (INIS) Casas Galiano, G.; Grau Malonda, A. 1994-01-01 An intelligent computer program has been developed to obtain the mathematical formulae to compute the probabilities and reduced energies of the different atomic rearrangement pathways following electron-capture decay. Creation and annihilation operators for Auger and X processes have been introduced. Taking into account the symmetries associated with each process, 262 different pathways were obtained. This model allows us to obtain the influence of the M-electron-capture in the counting efficiency when the atomic number of the nuclide is high 20. Probability-based collaborative filtering model for predicting gene–disease associations OpenAIRE Zeng, Xiangxiang; Ding, Ningxiang; Rodríguez-Patón, Alfonso; Zou, Quan 2017-01-01 Background Accurately predicting pathogenic human genes has been challenging in recent research. Considering extensive gene–disease data verified by biological experiments, we can apply computational methods to perform accurate predictions with reduced time and expenses. Methods We propose a probability-based collaborative filtering model (PCFM) to predict pathogenic human genes. Several kinds of data sets, containing data of humans and data of other nonhuman species, are integrated in our mo... 1. How to model a negligible probability under the WTO sanitary and phytosanitary agreement? Science.gov (United States) Powell, Mark R 2013-06-01 Since the 1997 EC--Hormones decision, World Trade Organization (WTO) Dispute Settlement Panels have wrestled with the question of what constitutes a negligible risk under the Sanitary and Phytosanitary Agreement. More recently, the 2010 WTO Australia--Apples Panel focused considerable attention on the appropriate quantitative model for a negligible probability in a risk assessment. The 2006 Australian Import Risk Analysis for Apples from New Zealand translated narrative probability statements into quantitative ranges. The uncertainty about a "negligible" probability was characterized as a uniform distribution with a minimum value of zero and a maximum value of 10(-6) . The Australia - Apples Panel found that the use of this distribution would tend to overestimate the likelihood of "negligible" events and indicated that a triangular distribution with a most probable value of zero and a maximum value of 10⁻⁶ would correct the bias. The Panel observed that the midpoint of the uniform distribution is 5 × 10⁻⁷ but did not consider that the triangular distribution has an expected value of 3.3 × 10⁻⁷. Therefore, if this triangular distribution is the appropriate correction, the magnitude of the bias found by the Panel appears modest. The Panel's detailed critique of the Australian risk assessment, and the conclusions of the WTO Appellate Body about the materiality of flaws found by the Panel, may have important implications for the standard of review for risk assessments under the WTO SPS Agreement. © 2012 Society for Risk Analysis. 2. Model-assisted probability of detection of flaws in aluminum blocks using polynomial chaos expansions Science.gov (United States) Du, Xiaosong; Leifsson, Leifur; Grandin, Robert; Meeker, William; Roberts, Ronald; Song, Jiming 2018-04-01 Probability of detection (POD) is widely used for measuring reliability of nondestructive testing (NDT) systems. Typically, POD is determined experimentally, while it can be enhanced by utilizing physics-based computational models in combination with model-assisted POD (MAPOD) methods. With the development of advanced physics-based methods, such as ultrasonic NDT testing, the empirical information, needed for POD methods, can be reduced. However, performing accurate numerical simulations can be prohibitively time-consuming, especially as part of stochastic analysis. In this work, stochastic surrogate models for computational physics-based measurement simulations are developed for cost savings of MAPOD methods while simultaneously ensuring sufficient accuracy. The stochastic surrogate is used to propagate the random input variables through the physics-based simulation model to obtain the joint probability distribution of the output. The POD curves are then generated based on those results. Here, the stochastic surrogates are constructed using non-intrusive polynomial chaos (NIPC) expansions. In particular, the NIPC methods used are the quadrature, ordinary least-squares (OLS), and least-angle regression sparse (LARS) techniques. The proposed approach is demonstrated on the ultrasonic testing simulation of a flat bottom hole flaw in an aluminum block. The results show that the stochastic surrogates have at least two orders of magnitude faster convergence on the statistics than direct Monte Carlo sampling (MCS). Moreover, the evaluation of the stochastic surrogate models is over three orders of magnitude faster than the underlying simulation model for this case, which is the UTSim2 model. 3. A statistical model for investigating binding probabilities of DNA nucleotide sequences using microarrays. Science.gov (United States) Lee, Mei-Ling Ting; Bulyk, Martha L; Whitmore, G A; Church, George M 2002-12-01 There is considerable scientific interest in knowing the probability that a site-specific transcription factor will bind to a given DNA sequence. Microarray methods provide an effective means for assessing the binding affinities of a large number of DNA sequences as demonstrated by Bulyk et al. (2001, Proceedings of the National Academy of Sciences, USA 98, 7158-7163) in their study of the DNA-binding specificities of Zif268 zinc fingers using microarray technology. In a follow-up investigation, Bulyk, Johnson, and Church (2002, Nucleic Acid Research 30, 1255-1261) studied the interdependence of nucleotides on the binding affinities of transcription proteins. Our article is motivated by this pair of studies. We present a general statistical methodology for analyzing microarray intensity measurements reflecting DNA-protein interactions. The log probability of a protein binding to a DNA sequence on an array is modeled using a linear ANOVA model. This model is convenient because it employs familiar statistical concepts and procedures and also because it is effective for investigating the probability structure of the binding mechanism. 4. An extended car-following model considering the appearing probability of truck and driver's characteristics Science.gov (United States) Rong, Ying; Wen, Huiying 2018-05-01 In this paper, the appearing probability of truck is introduced and an extended car-following model is presented to analyze the traffic flow based on the consideration of driver's characteristics, under honk environment. The stability condition of this proposed model is obtained through linear stability analysis. In order to study the evolution properties of traffic wave near the critical point, the mKdV equation is derived by the reductive perturbation method. The results show that the traffic flow will become more disorder for the larger appearing probability of truck. Besides, the appearance of leading truck affects not only the stability of traffic flow, but also the effect of other aspects on traffic flow, such as: driver's reaction and honk effect. The effects of them on traffic flow are closely correlated with the appearing probability of truck. Finally, the numerical simulations under the periodic boundary condition are carried out to verify the proposed model. And they are consistent with the theoretical findings. 5. Application of a weighted spatial probability model in GIS to analyse landslides in Penang Island, Malaysia Directory of Open Access Journals (Sweden) Samy Ismail Elmahdy 2016-01-01 Full Text Available In the current study, Penang Island, which is one of the several mountainous areas in Malaysia that is often subjected to landslide hazard, was chosen for further investigation. A multi-criteria Evaluation and the spatial probability weighted approach and model builder was applied to map and analyse landslides in Penang Island. A set of automated algorithms was used to construct new essential geological and morphometric thematic maps from remote sensing data. The maps were ranked using the weighted probability spatial model based on their contribution to the landslide hazard. Results obtained showed that sites at an elevation of 100–300 m, with steep slopes of 10°–37° and slope direction (aspect in the E and SE directions were areas of very high and high probability for the landslide occurrence; the total areas were 21.393 km2 (11.84% and 58.690 km2 (32.48%, respectively. The obtained map was verified by comparing variogram models of the mapped and the occurred landslide locations and showed a strong correlation with the locations of occurred landslides, indicating that the proposed method can successfully predict the unpredictable landslide hazard. The method is time and cost effective and can be used as a reference for geological and geotechnical engineers. 6. Pulmonary Complications Resulting from Genetic Cardiovascular Disease in Two Rat Models Science.gov (United States) Underlying cardiovascular disease (CVD) has been considered a risk factor for exacerbation of air pollution health effects. Therefore, rodent models of CVD are increasingly used to examine mechanisms of variation in susceptibility. Pulmonary complications and altered iron homeost... 7. Moritz enhancements for visualization of complicated geometry models International Nuclear Information System (INIS) Van Riper, K. A. 2009-01-01 We describe new features implemented in the Moritz geometry editing and visualization program to enhance the accuracy and efficiency of viewing complex geometry models. The 3D display is based on OpenGL and requires conversion of the combinatorial surface and solid body geometry used by MCNP and other transport codes to a set of polygons. Calculation of those polygons can take many minutes for complex models. Once calculated, the polygons can be saved to a file and reused when the same or a derivative model is loaded; the file can be read and processed in under a second. A cell can be filled with a collection of other cells constituting a universe. A new option bypasses use of the filled cell's boundaries when calculating the polygons for the filling universe. This option, when applicable, speeds processing, improves the 3D image, and permits reuse of the universe's polygons when other cells are filled with transformed instances of the universe. Surfaces and solid bodies used in a cell description must be converted to polygons before calculating the polygonal representation of a cell; this conversion requires truncation of infinite surfaces. A new method for truncating transformed surfaces ensures the finite surface intersects the entire model. When a surface or solid body is processed in a cell description, an optional test detects when that object does not contribute additional polygons; if so, that object May be extraneous for the cell description. (authors) 8. Interpreting and Understanding Logits, Probits, and other Non-Linear Probability Models DEFF Research Database (Denmark) Breen, Richard; Karlson, Kristian Bernt; Holm, Anders 2018-01-01 Methods textbooks in sociology and other social sciences routinely recommend the use of the logit or probit model when an outcome variable is binary, an ordered logit or ordered probit when it is ordinal, and a multinomial logit when it has more than two categories. But these methodological...... guidelines take little or no account of a body of work that, over the past 30 years, has pointed to problematic aspects of these nonlinear probability models and, particularly, to difficulties in interpreting their parameters. In this chapterreview, we draw on that literature to explain the problems, show... 9. A cellular automata model of traffic flow with variable probability of randomization International Nuclear Information System (INIS) Zheng Wei-Fan; Zhang Ji-Ye 2015-01-01 Research on the stochastic behavior of traffic flow is important to understand the intrinsic evolution rules of a traffic system. By introducing an interactional potential of vehicles into the randomization step, an improved cellular automata traffic flow model with variable probability of randomization is proposed in this paper. In the proposed model, the driver is affected by the interactional potential of vehicles before him, and his decision-making process is related to the interactional potential. Compared with the traditional cellular automata model, the modeling is more suitable for the driver’s random decision-making process based on the vehicle and traffic situations in front of him in actual traffic. From the improved model, the fundamental diagram (flow–density relationship) is obtained, and the detailed high-density traffic phenomenon is reproduced through numerical simulation. (paper) 10. The limiting conditional probability distribution in a stochastic model of T cell repertoire maintenance. Science.gov (United States) Stirk, Emily R; Lythe, Grant; van den Berg, Hugo A; Hurst, Gareth A D; Molina-París, Carmen 2010-04-01 The limiting conditional probability distribution (LCD) has been much studied in the field of mathematical biology, particularly in the context of epidemiology and the persistence of epidemics. However, it has not yet been applied to the immune system. One of the characteristic features of the T cell repertoire is its diversity. This diversity declines in old age, whence the concepts of extinction and persistence are also relevant to the immune system. In this paper we model T cell repertoire maintenance by means of a continuous-time birth and death process on the positive integers, where the origin is an absorbing state. We show that eventual extinction is guaranteed. The late-time behaviour of the process before extinction takes place is modelled by the LCD, which we prove always exists for the process studied here. In most cases, analytic expressions for the LCD cannot be computed but the probability distribution may be approximated by means of the stationary probability distributions of two related processes. We show how these approximations are related to the LCD of the original process and use them to study the LCD in two special cases. We also make use of the large N expansion to derive a further approximation to the LCD. The accuracy of the various approximations is then analysed. (c) 2009 Elsevier Inc. All rights reserved. 11. Predicting medical complications after spine surgery: a validated model using a prospective surgical registry. Science.gov (United States) Lee, Michael J; Cizik, Amy M; Hamilton, Deven; Chapman, Jens R 2014-02-01 The possibility and likelihood of a postoperative medical complication after spine surgery undoubtedly play a major role in the decision making of the surgeon and patient alike. Although prior study has determined relative risk and odds ratio values to quantify risk factors, these values may be difficult to translate to the patient during counseling of surgical options. Ideally, a model that predicts absolute risk of medical complication, rather than relative risk or odds ratio values, would greatly enhance the discussion of safety of spine surgery. To date, there is no risk stratification model that specifically predicts the risk of medical complication. The purpose of this study was to create and validate a predictive model for the risk of medical complication during and after spine surgery. Statistical analysis using a prospective surgical spine registry that recorded extensive demographic, surgical, and complication data. Outcomes examined are medical complications that were specifically defined a priori. This analysis is a continuation of statistical analysis of our previously published report. Using a prospectively collected surgical registry of more than 1,476 patients with extensive demographic, comorbidity, surgical, and complication detail recorded for 2 years after surgery, we previously identified several risk factor for medical complications. Using the beta coefficients from those log binomial regression analyses, we created a model to predict the occurrence of medical complication after spine surgery. We split our data into two subsets for internal and cross-validation of our model. We created two predictive models: one predicting the occurrence of any medical complication and the other predicting the occurrence of a major medical complication. The final predictive model for any medical complications had a receiver operator curve characteristic of 0.76, considered to be a fair measure. The final predictive model for any major medical complications had 12. A Novel Probability Model for Suppressing Multipath Ghosts in GPR and TWI Imaging: A Numerical Study Directory of Open Access Journals (Sweden) Tan Yun-hua 2015-10-01 Full Text Available A novel concept for suppressing the problem of multipath ghosts in Ground Penetrating Radar (GPR and Through-Wall Imaging (TWI is presented. Ghosts (i.e., false targets mainly arise from the use of the Born or single-scattering approximations that lead to linearized imaging algorithms; however, these approximations neglect the effect of multiple scattering (or multipath between the electromagnetic wavefield and the object under investigation. In contrast to existing methods of suppressing multipath ghosts, the proposed method models for the first time the reflectivity of the probed objects as a probability function up to a normalized factor and introduces the concept of random subaperture by randomly picking up measurement locations from the entire aperture. Thus, the final radar image is a joint probability distribution that corresponds to radar images derived from multiple random subapertures. Finally, numerical experiments are used to demonstrate the performance of the proposed methodology in GPR and TWI imaging. 13. Probability modeling of high flow extremes in Yingluoxia watershed, the upper reaches of Heihe River basin Science.gov (United States) Li, Zhanling; Li, Zhanjie; Li, Chengcheng 2014-05-01 Probability modeling of hydrological extremes is one of the major research areas in hydrological science. Most basins in humid and semi-humid south and east of China are concerned for probability modeling analysis of high flow extremes. While, for the inland river basin which occupies about 35% of the country area, there is a limited presence of such studies partly due to the limited data availability and a relatively low mean annual flow. The objective of this study is to carry out probability modeling of high flow extremes in the upper reach of Heihe River basin, the second largest inland river basin in China, by using the peak over threshold (POT) method and Generalized Pareto Distribution (GPD), in which the selection of threshold and inherent assumptions for POT series are elaborated in details. For comparison, other widely used probability distributions including generalized extreme value (GEV), Lognormal, Log-logistic and Gamma are employed as well. Maximum likelihood estimate is used for parameter estimations. Daily flow data at Yingluoxia station from 1978 to 2008 are used. Results show that, synthesizing the approaches of mean excess plot, stability features of model parameters, return level plot and the inherent independence assumption of POT series, an optimum threshold of 340m3/s is finally determined for high flow extremes in Yingluoxia watershed. The resulting POT series is proved to be stationary and independent based on Mann-Kendall test, Pettitt test and autocorrelation test. In terms of Kolmogorov-Smirnov test, Anderson-Darling test and several graphical diagnostics such as quantile and cumulative density function plots, GPD provides the best fit to high flow extremes in the study area. The estimated high flows for long return periods demonstrate that, as the return period increasing, the return level estimates are probably more uncertain. The frequency of high flow extremes exhibits a very slight but not significant decreasing trend from 1978 to 14. MASTER: a model to improve and standardize clinical breakpoints for antimicrobial susceptibility testing using forecast probabilities. Science.gov (United States) Blöchliger, Nicolas; Keller, Peter M; Böttger, Erik C; Hombach, Michael 2017-09-01 The procedure for setting clinical breakpoints (CBPs) for antimicrobial susceptibility has been poorly standardized with respect to population data, pharmacokinetic parameters and clinical outcome. Tools to standardize CBP setting could result in improved antibiogram forecast probabilities. We propose a model to estimate probabilities for methodological categorization errors and defined zones of methodological uncertainty (ZMUs), i.e. ranges of zone diameters that cannot reliably be classified. The impact of ZMUs on methodological error rates was used for CBP optimization. The model distinguishes theoretical true inhibition zone diameters from observed diameters, which suffer from methodological variation. True diameter distributions are described with a normal mixture model. The model was fitted to observed inhibition zone diameters of clinical Escherichia coli strains. Repeated measurements for a quality control strain were used to quantify methodological variation. For 9 of 13 antibiotics analysed, our model predicted error rates of  0.1% for ampicillin, cefoxitin, cefuroxime and amoxicillin/clavulanic acid. Increasing the susceptible CBP (cefoxitin) and introducing ZMUs (ampicillin, cefuroxime, amoxicillin/clavulanic acid) decreased error rates to < 0.1%. ZMUs contained low numbers of isolates for ampicillin and cefuroxime (3% and 6%), whereas the ZMU for amoxicillin/clavulanic acid contained 41% of all isolates and was considered not practical. We demonstrate that CBPs can be improved and standardized by minimizing methodological categorization error rates. ZMUs may be introduced if an intermediate zone is not appropriate for pharmacokinetic/pharmacodynamic or drug dosing reasons. Optimized CBPs will provide a standardized antibiotic susceptibility testing interpretation at a defined level of probability. © The Author 2017. Published by Oxford University Press on behalf of the British Society for Antimicrobial Chemotherapy. All rights reserved. For 15. Protein single-model quality assessment by feature-based probability density functions. Science.gov (United States) Cao, Renzhi; Cheng, Jianlin 2016-04-04 Protein quality assessment (QA) has played an important role in protein structure prediction. We developed a novel single-model quality assessment method-Qprob. Qprob calculates the absolute error for each protein feature value against the true quality scores (i.e. GDT-TS scores) of protein structural models, and uses them to estimate its probability density distribution for quality assessment. Qprob has been blindly tested on the 11th Critical Assessment of Techniques for Protein Structure Prediction (CASP11) as MULTICOM-NOVEL server. The official CASP result shows that Qprob ranks as one of the top single-model QA methods. In addition, Qprob makes contributions to our protein tertiary structure predictor MULTICOM, which is officially ranked 3rd out of 143 predictors. The good performance shows that Qprob is good at assessing the quality of models of hard targets. These results demonstrate that this new probability density distribution based method is effective for protein single-model quality assessment and is useful for protein structure prediction. The webserver of Qprob is available at: http://calla.rnet.missouri.edu/qprob/. The software is now freely available in the web server of Qprob. 16. Probabilities in quantum cosmological models: A decoherent histories analysis using a complex potential International Nuclear Information System (INIS) Halliwell, J. J. 2009-01-01 In the quantization of simple cosmological models (minisuperspace models) described by the Wheeler-DeWitt equation, an important step is the construction, from the wave function, of a probability distribution answering various questions of physical interest, such as the probability of the system entering a given region of configuration space at any stage in its entire history. A standard but heuristic procedure is to use the flux of (components of) the wave function in a WKB approximation. This gives sensible semiclassical results but lacks an underlying operator formalism. In this paper, we address the issue of constructing probability distributions linked to the Wheeler-DeWitt equation using the decoherent histories approach to quantum theory. The key step is the construction of class operators characterizing questions of physical interest. Taking advantage of a recent decoherent histories analysis of the arrival time problem in nonrelativistic quantum mechanics, we show that the appropriate class operators in quantum cosmology are readily constructed using a complex potential. The class operator for not entering a region of configuration space is given by the S matrix for scattering off a complex potential localized in that region. We thus derive the class operators for entering one or more regions in configuration space. The class operators commute with the Hamiltonian, have a sensible classical limit, and are closely related to an intersection number operator. The definitions of class operators given here handle the key case in which the underlying classical system has multiple crossings of the boundaries of the regions of interest. We show that oscillatory WKB solutions to the Wheeler-DeWitt equation give approximate decoherence of histories, as do superpositions of WKB solutions, as long as the regions of configuration space are sufficiently large. The corresponding probabilities coincide, in a semiclassical approximation, with standard heuristic procedures 17. Various models for pion probability distributions from heavy-ion collisions International Nuclear Information System (INIS) Mekjian, A.Z.; Mekjian, A.Z.; Schlei, B.R.; Strottman, D.; Schlei, B.R. 1998-01-01 Various models for pion multiplicity distributions produced in relativistic heavy ion collisions are discussed. The models include a relativistic hydrodynamic model, a thermodynamic description, an emitting source pion laser model, and a description which generates a negative binomial description. The approach developed can be used to discuss other cases which will be mentioned. The pion probability distributions for these various cases are compared. Comparison of the pion laser model and Bose-Einstein condensation in a laser trap and with the thermal model are made. The thermal model and hydrodynamic model are also used to illustrate why the number of pions never diverges and why the Bose-Einstein correction effects are relatively small. The pion emission strength η of a Poisson emitter and a critical density η c are connected in a thermal model by η/n c =e -m/T <1, and this fact reduces any Bose-Einstein correction effects in the number and number fluctuation of pions. Fluctuations can be much larger than Poisson in the pion laser model and for a negative binomial description. The clan representation of the negative binomial distribution due to Van Hove and Giovannini is discussed using the present description. Applications to CERN/NA44 and CERN/NA49 data are discussed in terms of the relativistic hydrodynamic model. copyright 1998 The American Physical Society 18. Finite element model updating of concrete structures based on imprecise probability Science.gov (United States) Biswal, S.; Ramaswamy, A. 2017-09-01 Imprecise probability based methods are developed in this study for the parameter estimation, in finite element model updating for concrete structures, when the measurements are imprecisely defined. Bayesian analysis using Metropolis Hastings algorithm for parameter estimation is generalized to incorporate the imprecision present in the prior distribution, in the likelihood function, and in the measured responses. Three different cases are considered (i) imprecision is present in the prior distribution and in the measurements only, (ii) imprecision is present in the parameters of the finite element model and in the measurement only, and (iii) imprecision is present in the prior distribution, in the parameters of the finite element model, and in the measurements. Procedures are also developed for integrating the imprecision in the parameters of the finite element model, in the finite element software Abaqus. The proposed methods are then verified against reinforced concrete beams and prestressed concrete beams tested in our laboratory as part of this study. 19. Probable biofilm formation in the cheek as a complication of soft tissue filler resulting from improper endodontic treatment of tooth 16 Directory of Open Access Journals (Sweden) Marusza W 2012-03-01 Full Text Available Wojciech Marusza1, Grazyna Mlynarczyk2, Romuald Olszanski3, Irina Netsvyetayeva2, Michael Obrowski1, Tommaso Iannitti4, Beniamino Palmieri51Academy of Face Sculpturing, 2Department of Microbiology, Medical University of Warsaw, 3Military Institute of Health Services, Warsaw, Poland; 4Department of Physiology, School of Medicine, University of Kentucky Medical Center, KY, USA; 5Department of General Surgery and Surgical Specialties, University of Modena and Reggio Emilia Medical School, Surgical Clinic, Modena, ItalyAbstract: Injectable filling agents offer the promise of a better appearance without surgery and, among them, hyaluronic acid is the most commonly used. Although complications are rare, it is necessary to know the possible side effects and complications in order to be prepared for their management. That is why many researchers have been focusing on the interactions between hyaluronic acid and pathogens, inflammatory mediators, the immune system, and markers of oxidative stress to achieve efficient drug delivery, given that hyaluronic acid has widening applications in the field of nanomedicine. Here we report the case of a 37-year-old female patient who returned to our clinic with an abscess in her left cheek 3 months after a deep injection of 1 mL of stabilized hyaluronic acid in both cheeks. Steroid and antibiotic therapy was initiated without success, and abscess drainage was performed. Extraction of tooth 16 was performed 11 days after insertion of drains into the abscess. Laboratory blood tests showed acute inflammation of presumed bacterial etiology. Microbiological examination of pus was negative. Bacterial cultures were found in the extracted tooth. After antibiotic therapy, a complete reversal of the pathological process was observed. The present report highlights the need to assess periodontal problems prior to any aesthetic facial treatment. Analyses of further case reports and clinical studies are necessary to understand 20. Analytical models of probability distribution and excess noise factor of solid state photomultiplier signals with crosstalk International Nuclear Information System (INIS) 2012-01-01 Silicon Photomultipliers (SiPM), also called Solid State Photomultipliers (SSPM), are based on Geiger mode avalanche breakdown that is limited by a strong negative feedback. An SSPM can detect and resolve single photons due to the high gain and ultra-low excess noise of avalanche multiplication in this mode. Crosstalk and afterpulsing processes associated with the high gain introduce specific excess noise and deteriorate the photon number resolution of the SSPM. The probabilistic features of these processes are widely studied because of its significance for the SSPM design, characterization, optimization and application, but the process modeling is mostly based on Monte Carlo simulations and numerical methods. In this study, crosstalk is considered to be a branching Poisson process, and analytical models of probability distribution and excess noise factor (ENF) of SSPM signals based on the Borel distribution as an advance on the geometric distribution models are presented and discussed. The models are found to be in a good agreement with the experimental probability distributions for dark counts and a few photon spectrums in a wide range of fired pixels number as well as with observed super-linear behavior of crosstalk ENF. 1. Capacity analysis in multi-state synaptic models: a retrieval probability perspective. Science.gov (United States) Huang, Yibi; Amit, Yali 2011-06-01 We define the memory capacity of networks of binary neurons with finite-state synapses in terms of retrieval probabilities of learned patterns under standard asynchronous dynamics with a predetermined threshold. The threshold is set to control the proportion of non-selective neurons that fire. An optimal inhibition level is chosen to stabilize network behavior. For any local learning rule we provide a computationally efficient and highly accurate approximation to the retrieval probability of a pattern as a function of its age. The method is applied to the sequential models (Fusi and Abbott, Nat Neurosci 10:485-493, 2007) and meta-plasticity models (Fusi et al., Neuron 45(4):599-611, 2005; Leibold and Kempter, Cereb Cortex 18:67-77, 2008). We show that as the number of synaptic states increases, the capacity, as defined here, either plateaus or decreases. In the few cases where multi-state models exceed the capacity of binary synapse models the improvement is small. 2. A formalism to generate probability distributions for performance-assessment modeling International Nuclear Information System (INIS) Kaplan, P.G. 1990-01-01 A formalism is presented for generating probability distributions of parameters used in performance-assessment modeling. The formalism is used when data are either sparse or nonexistent. The appropriate distribution is a function of the known or estimated constraints and is chosen to maximize a quantity known as Shannon's informational entropy. The formalism is applied to a parameter used in performance-assessment modeling. The functional form of the model that defines the parameter, data from the actual field site, and natural analog data are analyzed to estimate the constraints. A beta probability distribution of the example parameter is generated after finding four constraints. As an example of how the formalism is applied to the site characterization studies of Yucca Mountain, the distribution is generated for an input parameter in a performance-assessment model currently used to estimate compliance with disposal of high-level radioactive waste in geologic repositories, 10 CFR 60.113(a)(2), commonly known as the ground water travel time criterion. 8 refs., 2 figs 3. Influences of variables on ship collision probability in a Bayesian belief network model International Nuclear Information System (INIS) Hänninen, Maria; Kujala, Pentti 2012-01-01 The influences of the variables in a Bayesian belief network model for estimating the role of human factors on ship collision probability in the Gulf of Finland are studied for discovering the variables with the largest influences and for examining the validity of the network. The change in the so-called causation probability is examined while observing each state of the network variables and by utilizing sensitivity and mutual information analyses. Changing course in an encounter situation is the most influential variable in the model, followed by variables such as the Officer of the Watch's action, situation assessment, danger detection, personal condition and incapacitation. The least influential variables are the other distractions on bridge, the bridge view, maintenance routines and the officer's fatigue. In general, the methods are found to agree on the order of the model variables although some disagreements arise due to slightly dissimilar approaches to the concept of variable influence. The relative values and the ranking of variables based on the values are discovered to be more valuable than the actual numerical values themselves. Although the most influential variables seem to be plausible, there are some discrepancies between the indicated influences in the model and literature. Thus, improvements are suggested to the network. 4. Empirical probability model of cold plasma environment in the Jovian magnetosphere Science.gov (United States) Futaana, Yoshifumi; Wang, Xiao-Dong; Barabash, Stas; Roussos, Elias; Truscott, Pete 2015-04-01 We analyzed the Galileo PLS dataset to produce a new cold plasma environment model for the Jovian magneto- sphere. Although there exist many sophisticated radiation models, treating energetic plasma (e.g. JOSE, GIRE, or Salammbo), only a limited number of simple models has been utilized for cold plasma environment. By extend- ing the existing cold plasma models toward the probability domain, we can predict the extreme periods of Jovian environment by specifying the percentile of the environmental parameters. The new model was produced in the following procedure. We first referred to the existing cold plasma models of Divine and Garrett, 1983 (DG83) or Bagenal and Delamere 2011 (BD11). These models are scaled to fit the statistical median of the parameters obtained from Galileo PLS data. The scaled model (also called as "mean model") indicates the median environment of Jovian magnetosphere. Then, assuming that the deviations in the Galileo PLS parameters are purely due to variations in the environment, we extended the mean model toward the percentile domain. The input parameter of the model is simply the position of the spacecraft (distance, magnetic longitude and lati- tude) and the specific percentile (e.g. 0.5 for the mean model). All the parameters in the model are described in mathematical forms; therefore the needed computational resources are quite low. The new model can be used for assessing the JUICE mission profile. The spatial extent of the model covers the main phase of the JUICE mission; namely from the Europa orbit to 40 Rj (where Rj is the radius of Jupiter). In addition, theoretical extensions toward the latitudinal direction are also included in the model to support the high latitude orbit of the JUICE spacecraft. 5. Fast Outage Probability Simulation for FSO Links with a Generalized Pointing Error Model KAUST Repository Ben Issaid, Chaouki 2017-02-07 Over the past few years, free-space optical (FSO) communication has gained significant attention. In fact, FSO can provide cost-effective and unlicensed links, with high-bandwidth capacity and low error rate, making it an exciting alternative to traditional wireless radio-frequency communication systems. However, the system performance is affected not only by the presence of atmospheric turbulences, which occur due to random fluctuations in the air refractive index but also by the existence of pointing errors. Metrics, such as the outage probability which quantifies the probability that the instantaneous signal-to-noise ratio is smaller than a given threshold, can be used to analyze the performance of this system. In this work, we consider weak and strong turbulence regimes, and we study the outage probability of an FSO communication system under a generalized pointing error model with both a nonzero boresight component and different horizontal and vertical jitter effects. More specifically, we use an importance sampling approach which is based on the exponential twisting technique to offer fast and accurate results. 6. Fixation Probability in a Two-Locus Model by the Ancestral Recombination–Selection Graph Science.gov (United States) Lessard, Sabin; Kermany, Amir R. 2012-01-01 We use the ancestral influence graph (AIG) for a two-locus, two-allele selection model in the limit of a large population size to obtain an analytic approximation for the probability of ultimate fixation of a single mutant allele A. We assume that this new mutant is introduced at a given locus into a finite population in which a previous mutant allele B is already segregating with a wild type at another linked locus. We deduce that the fixation probability increases as the recombination rate increases if allele A is either in positive epistatic interaction with B and allele B is beneficial or in no epistatic interaction with B and then allele A itself is beneficial. This holds at least as long as the recombination fraction and the selection intensity are small enough and the population size is large enough. In particular this confirms the Hill–Robertson effect, which predicts that recombination renders more likely the ultimate fixation of beneficial mutants at different loci in a population in the presence of random genetic drift even in the absence of epistasis. More importantly, we show that this is true from weak negative epistasis to positive epistasis, at least under weak selection. In the case of deleterious mutants, the fixation probability decreases as the recombination rate increases. This supports Muller’s ratchet mechanism to explain the accumulation of deleterious mutants in a population lacking recombination. PMID:22095080 7. Zebrafish as a Model for the Study of Microvascular Complications of Diabetes and Their Mechanisms Directory of Open Access Journals (Sweden) Karl Heckler 2017-09-01 Full Text Available Diabetes mellitus (DM is a crucial metabolic disease that leads to severe disorders. These include macrovascular complications such as myocardial infarction, stroke, and peripheral artery disease and microvascular complications including diabetic nephropathy, neuropathy, and retinopathy. Diabetes mellitus, along with its associated organ pathologies, is one of the key problems in today’s medicine. Zebrafish is an upcoming disease model organism in diabetes research. Its glucose metabolism and the pathways of reactive metabolite formation are very similar to those of humans. Moreover, several physiological and pathophysiological pathways that also exist in humans and other mammals have been identified in this species or are currently under intense investigation. Zebrafish offer sophisticated imaging techniques and allow simple and fast genetic and pharmacological approaches with a high throughput. In this review, we highlight achievements and mechanisms concerning microvascular complications discovered in zebrafish, and we discuss the advantages and disadvantages of zebrafish as a model for studying diabetic complications. 8. Precipitation intensity probability distribution modelling for hydrological and construction design purposes International Nuclear Information System (INIS) Koshinchanov, Georgy; Dimitrov, Dobri 2008-01-01 The characteristics of rainfall intensity are important for many purposes, including design of sewage and drainage systems, tuning flood warning procedures, etc. Those estimates are usually statistical estimates of the intensity of precipitation realized for certain period of time (e.g. 5, 10 min., etc) with different return period (e.g. 20, 100 years, etc). The traditional approach in evaluating the mentioned precipitation intensities is to process the pluviometer's records and fit probability distribution to samples of intensities valid for certain locations ore regions. Those estimates further become part of the state regulations to be used for various economic activities. Two problems occur using the mentioned approach: 1. Due to various factors the climate conditions are changed and the precipitation intensity estimates need regular update; 2. As far as the extremes of the probability distribution are of particular importance for the practice, the methodology of the distribution fitting needs specific attention to those parts of the distribution. The aim of this paper is to make review of the existing methodologies for processing the intensive rainfalls and to refresh some of the statistical estimates for the studied areas. The methodologies used in Bulgaria for analyzing the intensive rainfalls and produce relevant statistical estimates: - The method of the maximum intensity, used in the National Institute of Meteorology and Hydrology to process and decode the pluviometer's records, followed by distribution fitting for each precipitation duration period; - As the above, but with separate modeling of probability distribution for the middle and high probability quantiles. - Method is similar to the first one, but with a threshold of 0,36 mm/min of intensity; - Another method proposed by the Russian hydrologist G. A. Aleksiev for regionalization of estimates over some territory, improved and adapted by S. Gerasimov for Bulgaria; - Next method is considering only 9. Neighbor-dependent Ramachandran probability distributions of amino acids developed from a hierarchical Dirichlet process model. Directory of Open Access Journals (Sweden) Daniel Ting 2010-04-01 Full Text Available Distributions of the backbone dihedral angles of proteins have been studied for over 40 years. While many statistical analyses have been presented, only a handful of probability densities are publicly available for use in structure validation and structure prediction methods. The available distributions differ in a number of important ways, which determine their usefulness for various purposes. These include: 1 input data size and criteria for structure inclusion (resolution, R-factor, etc.; 2 filtering of suspect conformations and outliers using B-factors or other features; 3 secondary structure of input data (e.g., whether helix and sheet are included; whether beta turns are included; 4 the method used for determining probability densities ranging from simple histograms to modern nonparametric density estimation; and 5 whether they include nearest neighbor effects on the distribution of conformations in different regions of the Ramachandran map. In this work, Ramachandran probability distributions are presented for residues in protein loops from a high-resolution data set with filtering based on calculated electron densities. Distributions for all 20 amino acids (with cis and trans proline treated separately have been determined, as well as 420 left-neighbor and 420 right-neighbor dependent distributions. The neighbor-independent and neighbor-dependent probability densities have been accurately estimated using Bayesian nonparametric statistical analysis based on the Dirichlet process. In particular, we used hierarchical Dirichlet process priors, which allow sharing of information between densities for a particular residue type and different neighbor residue types. The resulting distributions are tested in a loop modeling benchmark with the program Rosetta, and are shown to improve protein loop conformation prediction significantly. The distributions are available at http://dunbrack.fccc.edu/hdp. 10. A Bayesian joint probability modeling approach for seasonal forecasting of streamflows at multiple sites Science.gov (United States) Wang, Q. J.; Robertson, D. E.; Chiew, F. H. S. 2009-05-01 Seasonal forecasting of streamflows can be highly valuable for water resources management. In this paper, a Bayesian joint probability (BJP) modeling approach for seasonal forecasting of streamflows at multiple sites is presented. A Box-Cox transformed multivariate normal distribution is proposed to model the joint distribution of future streamflows and their predictors such as antecedent streamflows and El Niño-Southern Oscillation indices and other climate indicators. Bayesian inference of model parameters and uncertainties is implemented using Markov chain Monte Carlo sampling, leading to joint probabilistic forecasts of streamflows at multiple sites. The model provides a parametric structure for quantifying relationships between variables, including intersite correlations. The Box-Cox transformed multivariate normal distribution has considerable flexibility for modeling a wide range of predictors and predictands. The Bayesian inference formulated allows the use of data that contain nonconcurrent and missing records. The model flexibility and data-handling ability means that the BJP modeling approach is potentially of wide practical application. The paper also presents a number of statistical measures and graphical methods for verification of probabilistic forecasts of continuous variables. Results for streamflows at three river gauges in the Murrumbidgee River catchment in southeast Australia show that the BJP modeling approach has good forecast quality and that the fitted model is consistent with observed data. 11. Pre-Service Mathematics Teachers' Use of Probability Models in Making Informal Inferences about a Chance Game Science.gov (United States) Kazak, Sibel; Pratt, Dave 2017-01-01 This study considers probability models as tools for both making informal statistical inferences and building stronger conceptual connections between data and chance topics in teaching statistics. In this paper, we aim to explore pre-service mathematics teachers' use of probability models for a chance game, where the sum of two dice matters in… 12. Probabilities and energies to obtain the counting efficiency of electron-capture nuclides. KLMN model International Nuclear Information System (INIS) Galiano, G.; Grau, A. 1994-01-01 An intelligent computer program has been developed to obtain the mathematical formulae to compute the probabilities and reduced energies of the different atomic rearrangement pathways following electron-capture decay. Creation and annihilation operators for Auger and X processes have been introduced. Taking into account the symmetries associated with each process, 262 different pathways were obtained. This model allows us to obtain the influence of the M-electro capture in the counting efficiency when the atomic number of the nuclide is high. (Author) 13. Modeling Stress Strain Relationships and Predicting Failure Probabilities For Graphite Core Components Energy Technology Data Exchange (ETDEWEB) Duffy, Stephen [Cleveland State Univ., Cleveland, OH (United States) 2013-09-09 This project will implement inelastic constitutive models that will yield the requisite stress-strain information necessary for graphite component design. Accurate knowledge of stress states (both elastic and inelastic) is required to assess how close a nuclear core component is to failure. Strain states are needed to assess deformations in order to ascertain serviceability issues relating to failure, e.g., whether too much shrinkage has taken place for the core to function properly. Failure probabilities, as opposed to safety factors, are required in order to capture the bariability in failure strength in tensile regimes. The current stress state is used to predict the probability of failure. Stochastic failure models will be developed that can accommodate possible material anisotropy. This work will also model material damage (i.e., degradation of mechanical properties) due to radiation exposure. The team will design tools for components fabricated from nuclear graphite. These tools must readily interact with finite element software--in particular, COMSOL, the software algorithm currently being utilized by the Idaho National Laboratory. For the eleastic response of graphite, the team will adopt anisotropic stress-strain relationships available in COMSO. Data from the literature will be utilized to characterize the appropriate elastic material constants. 14. Modeling Stress Strain Relationships and Predicting Failure Probabilities For Graphite Core Components International Nuclear Information System (INIS) Duffy, Stephen 2013-01-01 This project will implement inelastic constitutive models that will yield the requisite stress-strain information necessary for graphite component design. Accurate knowledge of stress states (both elastic and inelastic) is required to assess how close a nuclear core component is to failure. Strain states are needed to assess deformations in order to ascertain serviceability issues relating to failure, e.g., whether too much shrinkage has taken place for the core to function properly. Failure probabilities, as opposed to safety factors, are required in order to capture the bariability in failure strength in tensile regimes. The current stress state is used to predict the probability of failure. Stochastic failure models will be developed that can accommodate possible material anisotropy. This work will also model material damage (i.e., degradation of mechanical properties) due to radiation exposure. The team will design tools for components fabricated from nuclear graphite. These tools must readily interact with finite element software--in particular, COMSOL, the software algorithm currently being utilized by the Idaho National Laboratory. For the eleastic response of graphite, the team will adopt anisotropic stress-strain relationships available in COMSO. Data from the literature will be utilized to characterize the appropriate elastic material constants. 15. Probability-based collaborative filtering model for predicting gene-disease associations. Science.gov (United States) Zeng, Xiangxiang; Ding, Ningxiang; Rodríguez-Patón, Alfonso; Zou, Quan 2017-12-28 Accurately predicting pathogenic human genes has been challenging in recent research. Considering extensive gene-disease data verified by biological experiments, we can apply computational methods to perform accurate predictions with reduced time and expenses. We propose a probability-based collaborative filtering model (PCFM) to predict pathogenic human genes. Several kinds of data sets, containing data of humans and data of other nonhuman species, are integrated in our model. Firstly, on the basis of a typical latent factorization model, we propose model I with an average heterogeneous regularization. Secondly, we develop modified model II with personal heterogeneous regularization to enhance the accuracy of aforementioned models. In this model, vector space similarity or Pearson correlation coefficient metrics and data on related species are also used. We compared the results of PCFM with the results of four state-of-arts approaches. The results show that PCFM performs better than other advanced approaches. PCFM model can be leveraged for predictions of disease genes, especially for new human genes or diseases with no known relationships. 16. Multivariate quantile mapping bias correction: an N-dimensional probability density function transform for climate model simulations of multiple variables Science.gov (United States) Cannon, Alex J. 2018-01-01 Most bias correction algorithms used in climatology, for example quantile mapping, are applied to univariate time series. They neglect the dependence between different variables. Those that are multivariate often correct only limited measures of joint dependence, such as Pearson or Spearman rank correlation. Here, an image processing technique designed to transfer colour information from one image to another—the N-dimensional probability density function transform—is adapted for use as a multivariate bias correction algorithm (MBCn) for climate model projections/predictions of multiple climate variables. MBCn is a multivariate generalization of quantile mapping that transfers all aspects of an observed continuous multivariate distribution to the corresponding multivariate distribution of variables from a climate model. When applied to climate model projections, changes in quantiles of each variable between the historical and projection period are also preserved. The MBCn algorithm is demonstrated on three case studies. First, the method is applied to an image processing example with characteristics that mimic a climate projection problem. Second, MBCn is used to correct a suite of 3-hourly surface meteorological variables from the Canadian Centre for Climate Modelling and Analysis Regional Climate Model (CanRCM4) across a North American domain. Components of the Canadian Forest Fire Weather Index (FWI) System, a complicated set of multivariate indices that characterizes the risk of wildfire, are then calculated and verified against observed values. Third, MBCn is used to correct biases in the spatial dependence structure of CanRCM4 precipitation fields. Results are compared against a univariate quantile mapping algorithm, which neglects the dependence between variables, and two multivariate bias correction algorithms, each of which corrects a different form of inter-variable correlation structure. MBCn outperforms these alternatives, often by a large margin 17. Leveraging electronic health records for predictive modeling of post-surgical complications. Science.gov (United States) Weller, Grant B; Lovely, Jenna; Larson, David W; Earnshaw, Berton A; Huebner, Marianne 2017-01-01 Hospital-specific electronic health record systems are used to inform clinical practice about best practices and quality improvements. Many surgical centers have developed deterministic clinical decision rules to discover adverse events (e.g. postoperative complications) using electronic health record data. However, these data provide opportunities to use probabilistic methods for early prediction of adverse health events, which may be more informative than deterministic algorithms. Electronic health record data from a set of 9598 colorectal surgery cases from 2010 to 2014 were used to predict the occurrence of selected complications including surgical site infection, ileus, and bleeding. Consistent with previous studies, we find a high rate of missing values for both covariates and complication information (4-90%). Several machine learning classification methods are trained on an 80% random sample of cases and tested on a remaining holdout set. Predictive performance varies by complication, although an area under the receiver operating characteristic curve as high as 0.86 on testing data was achieved for bleeding complications, and accuracy for all complications compares favorably to existing clinical decision rules. Our results confirm that electronic health records provide opportunities for improved risk prediction of surgical complications; however, consideration of data quality and consistency standards is an important step in predictive modeling with such data. 18. Probability Models Based on Soil Properties for Predicting Presence-Absence of Pythium in Soybean Roots. Science.gov (United States) Zitnick-Anderson, Kimberly K; Norland, Jack E; Del Río Mendoza, Luis E; Fortuna, Ann-Marie; Nelson, Berlin D 2017-10-01 Associations between soil properties and Pythium groups on soybean roots were investigated in 83 commercial soybean fields in North Dakota. A data set containing 2877 isolates of Pythium which included 26 known spp. and 1 unknown spp. and 13 soil properties from each field were analyzed. A Pearson correlation analysis was performed with all soil properties to observe any significant correlation between properties. Hierarchical clustering, indicator spp., and multi-response permutation procedures were used to identify groups of Pythium. Logistic regression analysis using stepwise selection was employed to calculate probability models for presence of groups based on soil properties. Three major Pythium groups were identified and three soil properties were associated with these groups. Group 1, characterized by P. ultimum, was associated with zinc levels; as zinc increased, the probability of group 1 being present increased (α = 0.05). Pythium group 2, characterized by Pythium kashmirense and an unknown Pythium sp., was associated with cation exchange capacity (CEC) (α < 0.05); as CEC increased, these spp. increased. Group 3, characterized by Pythium heterothallicum and Pythium irregulare, were associated with CEC and calcium carbonate exchange (CCE); as CCE increased and CEC decreased, these spp. increased (α = 0.05). The regression models may have value in predicting pathogenic Pythium spp. in soybean fields in North Dakota and adjacent states. 19. [Application of Bayes Probability Model in Differentiation of Yin and Yang Jaundice Syndromes in Neonates]. Science.gov (United States) Mu, Chun-sun; Zhang, Ping; Kong, Chun-yan; Li, Yang-ning 2015-09-01 To study the application of Bayes probability model in differentiating yin and yang jaundice syndromes in neonates. Totally 107 jaundice neonates who admitted to hospital within 10 days after birth were assigned to two groups according to syndrome differentiation, 68 in the yang jaundice syndrome group and 39 in the yin jaundice syndrome group. Data collected for neonates were factors related to jaundice before, during and after birth. Blood routines, liver and renal functions, and myocardial enzymes were tested on the admission day or the next day. Logistic regression model and Bayes discriminating analysis were used to screen factors important for yin and yang jaundice syndrome differentiation. Finally, Bayes probability model for yin and yang jaundice syndromes was established and assessed. Factors important for yin and yang jaundice syndrome differentiation screened by Logistic regression model and Bayes discriminating analysis included mothers' age, mother with gestational diabetes mellitus (GDM), gestational age, asphyxia, or ABO hemolytic diseases, red blood cell distribution width (RDW-SD), platelet-large cell ratio (P-LCR), serum direct bilirubin (DBIL), alkaline phosphatase (ALP), cholinesterase (CHE). Bayes discriminating analysis was performed by SPSS to obtain Bayes discriminant function coefficient. Bayes discriminant function was established according to discriminant function coefficients. Yang jaundice syndrome: y1= -21. 701 +2. 589 x mother's age + 1. 037 x GDM-17. 175 x asphyxia + 13. 876 x gestational age + 6. 303 x ABO hemolytic disease + 2.116 x RDW-SD + 0. 831 x DBIL + 0. 012 x ALP + 1. 697 x LCR + 0. 001 x CHE; Yin jaundice syndrome: y2= -33. 511 + 2.991 x mother's age + 3.960 x GDM-12. 877 x asphyxia + 11. 848 x gestational age + 1. 820 x ABO hemolytic disease +2. 231 x RDW-SD +0. 999 x DBIL +0. 023 x ALP +1. 916 x LCR +0. 002 x CHE. Bayes discriminant function was hypothesis tested and got Wilks' λ =0. 393 (P =0. 000). So Bayes 20. Repopulation of interacting tumor cells during fractionated radiotherapy: Stochastic modeling of the tumor control probability International Nuclear Information System (INIS) Fakir, Hatim; Hlatky, Lynn; Li, Huamin; Sachs, Rainer 2013-01-01 1. Repopulation of interacting tumor cells during fractionated radiotherapy: stochastic modeling of the tumor control probability. Science.gov (United States) Fakir, Hatim; Hlatky, Lynn; Li, Huamin; Sachs, Rainer 2013-12-01 2. Application of the Tripartite Model to a Complicated Sample of Residential Youth with Externalizing Problems Science.gov (United States) Chin, Eu Gene; Ebesutani, Chad; Young, John 2013-01-01 The tripartite model of anxiety and depression has received strong support among child and adolescent populations. Clinical samples of children and adolescents in these studies, however, have usually been referred for treatment of anxiety and depression. This study investigated the fit of the tripartite model with a complicated sample of… 3. Inferring Pairwise Interactions from Biological Data Using Maximum-Entropy Probability Models. Directory of Open Access Journals (Sweden) Richard R Stein 2015-07-01 Full Text Available Maximum entropy-based inference methods have been successfully used to infer direct interactions from biological datasets such as gene expression data or sequence ensembles. Here, we review undirected pairwise maximum-entropy probability models in two categories of data types, those with continuous and categorical random variables. As a concrete example, we present recently developed inference methods from the field of protein contact prediction and show that a basic set of assumptions leads to similar solution strategies for inferring the model parameters in both variable types. These parameters reflect interactive couplings between observables, which can be used to predict global properties of the biological system. Such methods are applicable to the important problems of protein 3-D structure prediction and association of gene-gene networks, and they enable potential applications to the analysis of gene alteration patterns and to protein design. 4. Probability distributions in conservative energy exchange models of multiple interacting agents International Nuclear Information System (INIS) Scafetta, Nicola; West, Bruce J 2007-01-01 Herein we study energy exchange models of multiple interacting agents that conserve energy in each interaction. The models differ regarding the rules that regulate the energy exchange and boundary effects. We find a variety of stochastic behaviours that manifest energy equilibrium probability distributions of different types and interaction rules that yield not only the exponential distributions such as the familiar Maxwell-Boltzmann-Gibbs distribution of an elastically colliding ideal particle gas, but also uniform distributions, truncated exponential distributions, Gaussian distributions, Gamma distributions, inverse power law distributions, mixed exponential and inverse power law distributions, and evolving distributions. This wide variety of distributions should be of value in determining the underlying mechanisms generating the statistical properties of complex phenomena including those to be found in complex chemical reactions 5. Energies and transition probabilities from the full solution of nuclear quadrupole-octupole model International Nuclear Information System (INIS) Strecker, M.; Lenske, H.; Minkov, N. 2013-01-01 A collective model of nuclear quadrupole-octupole vibrations and rotations, originally restricted to a coherent interplay between quadrupole and octupole modes, is now developed for application beyond this restriction. The eigenvalue problem is solved by diagonalizing the unrestricted Hamiltonian in the basis of the analytic solution obtained in the case of the coherent-mode assumption. Within this scheme the yrast alternating-parity band is constructed by the lowest eigenvalues having the appropriate parity at given angular momentum. Additionally we include the calculation of transition probabilities which are fitted with the energies simultaneously. As a result we obtain a unique set of parameters. The obtained model parameters unambiguously determine the shape of the quadrupole-octupole potential. From the resulting wave functions quadrupole deformation expectation values are calculated which are found to be in agreement with experimental values. (author) 6. The Use of Conditional Probability Integral Transformation Method for Testing Accelerated Failure Time Models Directory of Open Access Journals (Sweden) Abdalla Ahmed Abdel-Ghaly 2016-06-01 Full Text Available This paper suggests the use of the conditional probability integral transformation (CPIT method as a goodness of fit (GOF technique in the field of accelerated life testing (ALT, specifically for validating the underlying distributional assumption in accelerated failure time (AFT model. The method is based on transforming the data into independent and identically distributed (i.i.d Uniform (0, 1 random variables and then applying the modified Watson statistic to test the uniformity of the transformed random variables. This technique is used to validate each of the exponential, Weibull and lognormal distributions' assumptions in AFT model under constant stress and complete sampling. The performance of the CPIT method is investigated via a simulation study. It is concluded that this method performs well in case of exponential and lognormal distributions. Finally, a real life example is provided to illustrate the application of the proposed procedure. 7. Probability density function modeling of scalar mixing from concentrated sources in turbulent channel flow Science.gov (United States) Bakosi, J.; Franzese, P.; Boybeyi, Z. 2007-11-01 Dispersion of a passive scalar from concentrated sources in fully developed turbulent channel flow is studied with the probability density function (PDF) method. The joint PDF of velocity, turbulent frequency and scalar concentration is represented by a large number of Lagrangian particles. A stochastic near-wall PDF model combines the generalized Langevin model of Haworth and Pope [Phys. Fluids 29, 387 (1986)] with Durbin's [J. Fluid Mech. 249, 465 (1993)] method of elliptic relaxation to provide a mathematically exact treatment of convective and viscous transport with a nonlocal representation of the near-wall Reynolds stress anisotropy. The presence of walls is incorporated through the imposition of no-slip and impermeability conditions on particles without the use of damping or wall-functions. Information on the turbulent time scale is supplied by the gamma-distribution model of van Slooten et al. [Phys. Fluids 10, 246 (1998)]. Two different micromixing models are compared that incorporate the effect of small scale mixing on the transported scalar: the widely used interaction by exchange with the mean and the interaction by exchange with the conditional mean model. Single-point velocity and concentration statistics are compared to direct numerical simulation and experimental data at Reτ=1080 based on the friction velocity and the channel half width. The joint model accurately reproduces a wide variety of conditional and unconditional statistics in both physical and composition space. 8. A statistical model for deriving probability distributions of contamination for accidental releases International Nuclear Information System (INIS) ApSimon, H.M.; Davison, A.C. 1986-01-01 Results generated from a detailed long-range transport model, MESOS, simulating dispersal of a large number of hypothetical releases of radionuclides in a variety of meteorological situations over Western Europe have been used to derive a simpler statistical model, MESOSTAT. This model may be used to generate probability distributions of different levels of contamination at a receptor point 100-1000 km or so from the source (for example, across a frontier in another country) without considering individual release and dispersal scenarios. The model is embodied in a series of equations involving parameters which are determined from such factors as distance between source and receptor, nuclide decay and deposition characteristics, release duration, and geostrophic windrose at the source. Suitable geostrophic windrose data have been derived for source locations covering Western Europe. Special attention has been paid to the relatively improbable extreme values of contamination at the top end of the distribution. The MESOSTAT model and its development are described, with illustrations of its use and comparison with the original more detailed modelling techniques. (author) 9. Time series modeling of pathogen-specific disease probabilities with subsampled data. Science.gov (United States) Fisher, Leigh; Wakefield, Jon; Bauer, Cici; Self, Steve 2017-03-01 Many diseases arise due to exposure to one of multiple possible pathogens. We consider the situation in which disease counts are available over time from a study region, along with a measure of clinical disease severity, for example, mild or severe. In addition, we suppose a subset of the cases are lab tested in order to determine the pathogen responsible for disease. In such a context, we focus interest on modeling the probabilities of disease incidence given pathogen type. The time course of these probabilities is of great interest as is the association with time-varying covariates such as meteorological variables. In this set up, a natural Bayesian approach would be based on imputation of the unsampled pathogen information using Markov Chain Monte Carlo but this is computationally challenging. We describe a practical approach to inference that is easy to implement. We use an empirical Bayes procedure in a first step to estimate summary statistics. We then treat these summary statistics as the observed data and develop a Bayesian generalized additive model. We analyze data on hand, foot, and mouth disease (HFMD) in China in which there are two pathogens of primary interest, enterovirus 71 (EV71) and Coxackie A16 (CA16). We find that both EV71 and CA16 are associated with temperature, relative humidity, and wind speed, with reasonably similar functional forms for both pathogens. The important issue of confounding by time is modeled using a penalized B-spline model with a random effects representation. The level of smoothing is addressed by a careful choice of the prior on the tuning variance. © 2016, The International Biometric Society. 10. Watershed erosion modeling using the probability of sediment connectivity in a gently rolling system Science.gov (United States) Mahoney, David Tyler; Fox, James Forrest; Al Aamery, Nabil 2018-06-01 Sediment connectivity has been shown in recent years to explain how the watershed configuration controls sediment transport. However, we find no studies develop a watershed erosion modeling framework based on sediment connectivity, and few, if any, studies have quantified sediment connectivity for gently rolling systems. We develop a new predictive sediment connectivity model that relies on the intersecting probabilities for sediment supply, detachment, transport, and buffers to sediment transport, which is integrated in a watershed erosion model framework. The model predicts sediment flux temporally and spatially across a watershed using field reconnaissance results, a high-resolution digital elevation models, a hydrologic model, and shear-based erosion formulae. Model results validate the capability of the model to predict erosion pathways causing sediment connectivity. More notably, disconnectivity dominates the gently rolling watershed across all morphologic levels of the uplands, including, microtopography from low energy undulating surfaces across the landscape, swales and gullies only active in the highest events, karst sinkholes that disconnect drainage areas, and floodplains that de-couple the hillslopes from the stream corridor. Results show that sediment connectivity is predicted for about 2% or more the watershed's area 37 days of the year, with the remaining days showing very little or no connectivity. Only 12.8 ± 0.7% of the gently rolling watershed shows sediment connectivity on the wettest day of the study year. Results also highlight the importance of urban/suburban sediment pathways in gently rolling watersheds, and dynamic and longitudinal distributions of sediment connectivity might be further investigated in future work. We suggest the method herein provides the modeler with an added tool to account for sediment transport criteria and has the potential to reduce computational costs in watershed erosion modeling. 11. Modeling detection probability to improve marsh bird surveys in southern Canada and the Great Lakes states Directory of Open Access Journals (Sweden) Douglas C. Tozer 2016-12-01 Full Text Available Marsh birds are notoriously elusive, with variation in detection probability across species, regions, seasons, and different times of day and weather. Therefore, it is important to develop regional field survey protocols that maximize detections, but that also produce data for estimating and analytically adjusting for remaining differences in detections. We aimed to improve regional field survey protocols by estimating detection probability of eight elusive marsh bird species throughout two regions that have ongoing marsh bird monitoring programs: the southern Canadian Prairies (Prairie region and the southern portion of the Great Lakes basin and parts of southern Québec (Great Lakes-St. Lawrence region. We accomplished our goal using generalized binomial N-mixture models and data from ~22,300 marsh bird surveys conducted between 2008 and 2014 by Bird Studies Canada's Prairie, Great Lakes, and Québec Marsh Monitoring Programs. Across all species, on average, detection probability was highest in the Great Lakes-St. Lawrence region from the beginning of May until mid-June, and then fell throughout the remainder of the season until the end of June; was lowest in the Prairie region in mid-May and then increased throughout the remainder of the season until the end of June; was highest during darkness compared with light; and did not vary significantly according to temperature (range: 0-30°C, cloud cover (0%-100%, or wind (0-20 kph, or during morning versus evening. We used our results to formulate improved marsh bird survey protocols for each region. Our analysis and recommendations are useful and contribute to conservation of wetland birds at various scales from local single-species studies to the continental North American Marsh Bird Monitoring Program. 12. PHOTOMETRIC REDSHIFTS AND QUASAR PROBABILITIES FROM A SINGLE, DATA-DRIVEN GENERATIVE MODEL International Nuclear Information System (INIS) Bovy, Jo; Hogg, David W.; Weaver, Benjamin A.; Myers, Adam D.; Hennawi, Joseph F.; McMahon, Richard G.; Schiminovich, David; Sheldon, Erin S.; Brinkmann, Jon; Schneider, Donald P. 2012-01-01 We describe a technique for simultaneously classifying and estimating the redshift of quasars. It can separate quasars from stars in arbitrary redshift ranges, estimate full posterior distribution functions for the redshift, and naturally incorporate flux uncertainties, missing data, and multi-wavelength photometry. We build models of quasars in flux-redshift space by applying the extreme deconvolution technique to estimate the underlying density. By integrating this density over redshift, one can obtain quasar flux densities in different redshift ranges. This approach allows for efficient, consistent, and fast classification and photometric redshift estimation. This is achieved by combining the speed obtained by choosing simple analytical forms as the basis of our density model with the flexibility of non-parametric models through the use of many simple components with many parameters. We show that this technique is competitive with the best photometric quasar classification techniques—which are limited to fixed, broad redshift ranges and high signal-to-noise ratio data—and with the best photometric redshift techniques when applied to broadband optical data. We demonstrate that the inclusion of UV and NIR data significantly improves photometric quasar-star separation and essentially resolves all of the redshift degeneracies for quasars inherent to the ugriz filter system, even when included data have a low signal-to-noise ratio. For quasars spectroscopically confirmed by the SDSS 84% and 97% of the objects with Galaxy Evolution Explorer UV and UKIDSS NIR data have photometric redshifts within 0.1 and 0.3, respectively, of the spectroscopic redshift; this amounts to about a factor of three improvement over ugriz-only photometric redshifts. Our code to calculate quasar probabilities and redshift probability distributions is publicly available. 13. Probability of identification: a statistical model for the validation of qualitative botanical identification methods. Science.gov (United States) LaBudde, Robert A; Harnly, James M 2012-01-01 A qualitative botanical identification method (BIM) is an analytical procedure that returns a binary result (1 = Identified, 0 = Not Identified). A BIM may be used by a buyer, manufacturer, or regulator to determine whether a botanical material being tested is the same as the target (desired) material, or whether it contains excessive nontarget (undesirable) material. The report describes the development and validation of studies for a BIM based on the proportion of replicates identified, or probability of identification (POI), as the basic observed statistic. The statistical procedures proposed for data analysis follow closely those of the probability of detection, and harmonize the statistical concepts and parameters between quantitative and qualitative method validation. Use of POI statistics also harmonizes statistical concepts for botanical, microbiological, toxin, and other analyte identification methods that produce binary results. The POI statistical model provides a tool for graphical representation of response curves for qualitative methods, reporting of descriptive statistics, and application of performance requirements. Single collaborator and multicollaborative study examples are given. 14. Global climate change model natural climate variation: Paleoclimate data base, probabilities and astronomic predictors Energy Technology Data Exchange (ETDEWEB) Kukla, G.; Gavin, J. [Columbia Univ., Palisades, NY (United States). Lamont-Doherty Geological Observatory 1994-05-01 This report was prepared at the Lamont-Doherty Geological Observatory of Columbia University at Palisades, New York, under subcontract to Pacific Northwest Laboratory it is a part of a larger project of global climate studies which supports site characterization work required for the selection of a potential high-level nuclear waste repository and forms part of the Performance Assessment Scientific Support (PASS) Program at PNL. The work under the PASS Program is currently focusing on the proposed site at Yucca Mountain, Nevada, and is under the overall direction of the Yucca Mountain Project Office US Department of Energy, Las Vegas, Nevada. The final results of the PNL project will provide input to global atmospheric models designed to test specific climate scenarios which will be used in the site specific modeling work of others. The primary purpose of the data bases compiled and of the astronomic predictive models is to aid in the estimation of the probabilities of future climate states. The results will be used by two other teams working on the global climate study under contract to PNL. They are located at and the University of Maine in Orono, Maine, and the Applied Research Corporation in College Station, Texas. This report presents the results of the third years work on the global climate change models and the data bases describing past climates. 15. Constituent quarks as clusters in quark-gluon-parton model. [Total cross sections, probability distributions Energy Technology Data Exchange (ETDEWEB) Kanki, T [Osaka Univ., Toyonaka (Japan). Coll. of General Education 1976-12-01 We present a quark-gluon-parton model in which quark-partons and gluons make clusters corresponding to two or three constituent quarks (or anti-quarks) in the meson or in the baryon, respectively. We explicitly construct the constituent quark state (cluster), by employing the Kuti-Weisskopf theory and by requiring the scaling. The quark additivity of the hadronic total cross sections and the quark counting rules on the threshold powers of various distributions are satisfied. For small x (Feynman fraction), it is shown that the constituent quarks and quark-partons have quite different probability distributions. We apply our model to hadron-hadron inclusive reactions, and clarify that the fragmentation and the diffractive processes relate to the constituent quark distributions, while the processes in or near the central region are controlled by the quark-partons. Our model gives the reasonable interpretation for the experimental data and much improves the usual ''constituent interchange model'' result near and in the central region (x asymptotically equals x sub(T) asymptotically equals 0). 16. Probability distribution of magnetization in the one-dimensional Ising model: effects of boundary conditions Energy Technology Data Exchange (ETDEWEB) Antal, T [Physics Department, Simon Fraser University, Burnaby, BC V5A 1S6 (Canada); Droz, M [Departement de Physique Theorique, Universite de Geneve, CH 1211 Geneva 4 (Switzerland); Racz, Z [Institute for Theoretical Physics, Eoetvoes University, 1117 Budapest, Pazmany setany 1/a (Hungary) 2004-02-06 Finite-size scaling functions are investigated both for the mean-square magnetization fluctuations and for the probability distribution of the magnetization in the one-dimensional Ising model. The scaling functions are evaluated in the limit of the temperature going to zero (T {yields} 0), the size of the system going to infinity (N {yields} {infinity}) while N[1 - tanh(J/k{sub B}T)] is kept finite (J being the nearest neighbour coupling). Exact calculations using various boundary conditions (periodic, antiperiodic, free, block) demonstrate explicitly how the scaling functions depend on the boundary conditions. We also show that the block (small part of a large system) magnetization distribution results are identical to those obtained for free boundary conditions. 17. Probability Model of Allele Frequency of Alzheimer’s Disease Genetic Risk Factor Directory of Open Access Journals (Sweden) Afshin Fayyaz-Movaghar 2016-06-01 Full Text Available Background and Purpose: The identification of genetics risk factors of human diseases is very important. This study is conducted to model the allele frequencies (AFs of Alzheimer’s disease. Materials and Methods: In this study, several candidate probability distributions are fitted on a data set of Alzheimer’s disease genetic risk factor. Unknown parameters of the considered distributions are estimated, and some criterions of goodness-of-fit are calculated for the sake of comparison. Results: Based on some statistical criterions, the beta distribution gives the best fit on AFs. However, the estimate values of the parameters of beta distribution lead us to the standard uniform distribution. Conclusion: The AFs of Alzheimer’s disease follow the standard uniform distribution. 18. A scan statistic for continuous data based on the normal probability model Directory of Open Access Journals (Sweden) Huang Lan 2009-10-01 Full Text Available Abstract Temporal, spatial and space-time scan statistics are commonly used to detect and evaluate the statistical significance of temporal and/or geographical disease clusters, without any prior assumptions on the location, time period or size of those clusters. Scan statistics are mostly used for count data, such as disease incidence or mortality. Sometimes there is an interest in looking for clusters with respect to a continuous variable, such as lead levels in children or low birth weight. For such continuous data, we present a scan statistic where the likelihood is calculated using the the normal probability model. It may also be used for other distributions, while still maintaining the correct alpha level. In an application of the new method, we look for geographical clusters of low birth weight in New York City. 19. The probability distribution model of air pollution index and its dominants in Kuala Lumpur Science.gov (United States) AL-Dhurafi, Nasr Ahmed; Razali, Ahmad Mahir; Masseran, Nurulkamal; Zamzuri, Zamira Hasanah 2016-11-01 This paper focuses on the statistical modeling for the distributions of air pollution index (API) and its sub-indexes data observed at Kuala Lumpur in Malaysia. Five pollutants or sub-indexes are measured including, carbon monoxide (CO); sulphur dioxide (SO2); nitrogen dioxide (NO2), and; particulate matter (PM10). Four probability distributions are considered, namely log-normal, exponential, Gamma and Weibull in search for the best fit distribution to the Malaysian air pollutants data. In order to determine the best distribution for describing the air pollutants data, five goodness-of-fit criteria's are applied. This will help in minimizing the uncertainty in pollution resource estimates and improving the assessment phase of planning. The conflict in criterion results for selecting the best distribution was overcome by using the weight of ranks method. We found that the Gamma distribution is the best distribution for the majority of air pollutants data in Kuala Lumpur. 20. Animal Models of Diabetic Macrovascular Complications: Key Players in the Development of New Therapeutic Approaches Directory of Open Access Journals (Sweden) Suvi E. Heinonen 2015-01-01 Full Text Available Diabetes mellitus is a lifelong, incapacitating metabolic disease associated with chronic macrovascular complications (coronary heart disease, stroke, and peripheral vascular disease and microvascular disorders leading to damage of the kidneys (nephropathy and eyes (retinopathy. Based on the current trends, the rising prevalence of diabetes worldwide will lead to increased cardiovascular morbidity and mortality. Therefore, novel means to prevent and treat these complications are needed. Under the auspices of the IMI (Innovative Medicines Initiative, the SUMMIT (SUrrogate markers for Micro- and Macrovascular hard end points for Innovative diabetes Tools consortium is working on the development of novel animal models that better replicate vascular complications of diabetes and on the characterization of the available models. In the past years, with the high level of genomic information available and more advanced molecular tools, a very large number of models has been created. Selecting the right model for a specific study is not a trivial task and will have an impact on the study results and their interpretation. This review gathers information on the available experimental animal models of diabetic macrovascular complications and evaluates their pros and cons for research purposes as well as for drug development. 1. Mathematical models of tumor growth: translating absorbed dose to tumor control probability International Nuclear Information System (INIS) Sgouros, G. 1996-01-01 cell loss due to irradiation, the log-kill model, therefore, predicts that incomplete treatment of a kinetically heterogeneous tumor will yield a more proliferative tumor. The probability of tumor control in such a simulation may be obtained from the nadir in tumor cell number. If the nadir is not sufficiently low to yield a high probability of tumor control, then the tumor will re-grow. Since tumors in each sub-population are assumed lost at the same rate, cells comprising the sub-population with the shortest potential doubling time will re-grow the fastest, yielding a recurrent tumor that is more proliferative. A number of assumptions and simplifications are both implicitly and explicitly made in converting absorbed dose to tumor control probability. The modeling analyses described above must, therefore, be viewed in terms of understanding and evaluating different treatment approaches with the goal of treatment optimization rather than outcome prediction 2. A transmission/escape probabilities model for neutral particle transport in the outer regions of a diverted tokamak International Nuclear Information System (INIS) Stacey, W.M. 1992-12-01 A new computational model for neutral particle transport in the outer regions of a diverted tokamak plasma chamber is presented. The model is based on the calculation of transmission and escape probabilities using first-flight integral transport theory and the balancing of fluxes across the surfaces bounding the various regions. The geometrical complexity of the problem is included in precomputed probabilities which depend only on the mean free path of the region 3. Modeling Plan-Related Clinical Complications Using Machine Learning Tools in a Multiplan IMRT Framework International Nuclear Information System (INIS) Zhang, Hao H.; D'Souza, Warren D.; Shi Leyuan; Meyer, Robert R. 2009-01-01 Purpose: To predict organ-at-risk (OAR) complications as a function of dose-volume (DV) constraint settings without explicit plan computation in a multiplan intensity-modulated radiotherapy (IMRT) framework. Methods and Materials: Several plans were generated by varying the DV constraints (input features) on the OARs (multiplan framework), and the DV levels achieved by the OARs in the plans (plan properties) were modeled as a function of the imposed DV constraint settings. OAR complications were then predicted for each of the plans by using the imposed DV constraints alone (features) or in combination with modeled DV levels (plan properties) as input to machine learning (ML) algorithms. These ML approaches were used to model two OAR complications after head-and-neck and prostate IMRT: xerostomia, and Grade 2 rectal bleeding. Two-fold cross-validation was used for model verification and mean errors are reported. Results: Errors for modeling the achieved DV values as a function of constraint settings were 0-6%. In the head-and-neck case, the mean absolute prediction error of the saliva flow rate normalized to the pretreatment saliva flow rate was 0.42% with a 95% confidence interval of (0.41-0.43%). In the prostate case, an average prediction accuracy of 97.04% with a 95% confidence interval of (96.67-97.41%) was achieved for Grade 2 rectal bleeding complications. Conclusions: ML can be used for predicting OAR complications during treatment planning allowing for alternative DV constraint settings to be assessed within the planning framework. 4. MEASURING MODEL FOR BAD LOANS IN BANKS. THE DEFAULT PROBABILITY MODEL. Directory of Open Access Journals (Sweden) 2010-12-01 Full Text Available The banking sectors of the transition countries have progressed remarkably in the last 20 years. In fact, banking in most transition countries has largely shaken off the traumas of the transition eraAt the start of the 21st century banks in these countries look very much like banks elsewhere. That is, they are by no means problem free but they are struggling with the same issues as banks in other emerging market countries during the financial crises conditions. The institutional environment differs considerably among the countries. The goal we set with this article is to examine in terms of methodology the most important assessment criteria of a measuring model for bad loans. 5. Diagnostics of enterprise bankruptcy occurrence probability in an anti-crisis management: modern approaches and classification of models Directory of Open Access Journals (Sweden) I.V. Zhalinska 2015-09-01 Full Text Available Diagnostics of enterprise bankruptcy occurrence probability is defined as an important tool ensuring the viability of an organization under conditions of unpredictable dynamic environment. The paper aims to define the basic features of diagnostics of bankruptcy occurrence probability models and their classification. The article grounds the objective increasing of crisis probability in modern enterprises where such increasing leads to the need to improve the efficiency of anti-crisis enterprise activities. The system of anti-crisis management is based on the subsystem of diagnostics of bankruptcy occurrence probability. Such a subsystem is the main one for further measures to prevent and overcome the crisis. The classification of existing models of enterprise bankruptcy occurrence probability has been suggested. The classification is based on methodical and methodological principles of models. The following main groups of models are determined: the models using financial ratios, aggregates and scores, the models of discriminated analysis, the methods of strategic analysis, informal models, artificial intelligence systems and the combination of the models. The classification made it possible to identify the analytical capabilities of each of the groups of models suggested. 6. Z → bb-bar probability and asymmetry in a model of dynamical electroweak symmetry breaking International Nuclear Information System (INIS) Arbuzov, B.A.; Osipov, M.Yu. 1997-01-01 The deviations from the standard model in the probability of Z → bb-bar decay and in the forward-backward asymmetry in the reaction e + e - → bb-bar are studied in the framework of the model of dynamical electroweak symmetry breaking, the basic point of which is the existence of a triple anomalous W-boson vertex in a region of momenta restricted by a cutoff. A set of equations for additional terms in the W b t-bar vertex is obtained and its solution to the process Z → bb-bar is applied. It is shown that it is possible to obtain a consistent description of both deviations, which is quite nontrivial because these effects are not simply correlated. The necessary value of the anomalous W interaction coupling, λ = -0.22 ± 0.01, is consistent with existing limitations and leads to definite predictions, e.g., for pair W production in e + e - collisions at LEP 200 7. Constraint-based Student Modelling in Probability Story Problems with Scaffolding Techniques Directory of Open Access Journals (Sweden) Nabila Khodeir 2018-01-01 Full Text Available Constraint-based student modelling (CBM is an important technique employed in intelligent tutoring systems to model student knowledge to provide relevant assistance. This paper introduces the Math Story Problem Tutor (MAST, a Web-based intelligent tutoring system for probability story problems, which is able to generate problems of different contexts, types and difficulty levels for self-paced learning. Constraints in MAST are specified at a low-level of granularity to allow fine-grained diagnosis of the student error. Furthermore, MAST extends CBM to address errors due to misunderstanding of the narrative story. It can locate and highlight keywords that may have been overlooked or misunderstood leading to an error. This is achieved by utilizing the role of sentences and keywords that are defined through the Natural Language Generation (NLG methods deployed in the story problem generation. MAST also integrates CBM with scaffolding questions and feedback to provide various forms of help and guidance to the student. This allows the student to discover and correct any errors in his/her solution. MAST has been preliminary evaluated empirically and the results show the potential effectiveness in tutoring students with a decrease in the percentage of violated constraints along the learning curve. Additionally, there is a significant improvement in the results of the post–test exam in comparison to the pre-test exam of the students using MAST in comparison to those relying on the textbook 8. ESTIMATION OF BANKRUPTCY PROBABILITIES BY USING FUZZY LOGIC AND MERTON MODEL: AN APPLICATION ON USA COMPANIES Directory of Open Access Journals (Sweden) Çiğdem ÖZARİ 2018-01-01 Full Text Available In this study, we have worked on developing a brand-new index called Fuzzy-bankruptcy index. The aim of this index is to find out the default probability of any company X, independent from the sector it belongs. Fuzzy logic is used to state the financial ratiointerruption change related with time and inside different sectors, the new index is created to eliminate the number of the relativity of financial ratios. The four input variables inside the five main input variables used for the fuzzy process, are chosen from both factor analysis and clustering and the last input variable calculated from Merton Model. As we analyze in the past cases of the default history of companies, one could explore different reasons such as managerial arrogance, fraud and managerial mistakes, that are responsible for the very poor endings of prestigious companies like Enron, K-Mart. Because of these kind of situations, we try to design a model which one could be able to get a better view of a company’s financial position, and it couldbe prevent credit loan companies from investing in the wrong company and possibly from losing all investments using our Fuzzy-bankruptcy index. 9. Characteristics of the probability function for three random-walk models of reaction--diffusion processes International Nuclear Information System (INIS) Musho, M.K.; Kozak, J.J. 1984-01-01 A method is presented for calculating exactly the relative width (sigma 2 )/sup 1/2// , the skewness γ 1 , and the kurtosis γ 2 characterizing the probability distribution function for three random-walk models of diffusion-controlled processes. For processes in which a diffusing coreactant A reacts irreversibly with a target molecule B situated at a reaction center, three models are considered. The first is the traditional one of an unbiased, nearest-neighbor random walk on a d-dimensional periodic/confining lattice with traps; the second involves the consideration of unbiased, non-nearest-neigh bor (i.e., variable-step length) walks on the same d-dimensional lattice; and, the third deals with the case of a biased, nearest-neighbor walk on a d-dimensional lattice (wherein a walker experiences a potential centered at the deep trap site of the lattice). Our method, which has been described in detail elsewhere [P.A. Politowicz and J. J. Kozak, Phys. Rev. B 28, 5549 (1983)] is based on the use of group theoretic arguments within the framework of the theory of finite Markov processes 10. Rooting phylogenetic trees under the coalescent model using site pattern probabilities. Science.gov (United States) Tian, Yuan; Kubatko, Laura 2017-12-19 Phylogenetic tree inference is a fundamental tool to estimate ancestor-descendant relationships among different species. In phylogenetic studies, identification of the root - the most recent common ancestor of all sampled organisms - is essential for complete understanding of the evolutionary relationships. Rooted trees benefit most downstream application of phylogenies such as species classification or study of adaptation. Often, trees can be rooted by using outgroups, which are species that are known to be more distantly related to the sampled organisms than any other species in the phylogeny. However, outgroups are not always available in evolutionary research. In this study, we develop a new method for rooting species tree under the coalescent model, by developing a series of hypothesis tests for rooting quartet phylogenies using site pattern probabilities. The power of this method is examined by simulation studies and by application to an empirical North American rattlesnake data set. The method shows high accuracy across the simulation conditions considered, and performs well for the rattlesnake data. Thus, it provides a computationally efficient way to accurately root species-level phylogenies that incorporates the coalescent process. The method is robust to variation in substitution model, but is sensitive to the assumption of a molecular clock. Our study establishes a computationally practical method for rooting species trees that is more efficient than traditional methods. The method will benefit numerous evolutionary studies that require rooting a phylogenetic tree without having to specify outgroups. 11. Probability Modeling of Precipitation Extremes over Two River Basins in Northwest of China Directory of Open Access Journals (Sweden) Zhanling Li 2015-01-01 Full Text Available This paper is focused on the probability modeling with a range of distribution models over two inland river basins in China, together with the estimations of return levels on various return periods. Both annual and seasonal maximum precipitations (MP are investigated based on daily precipitation data at 13 stations from 1960 to 2010 in Heihe River and Shiyang River basins. Results show that GEV, Burr, and Weibull distributions provide the best fit to both annual and seasonal MP. Exponential and Pareto 2 distributions show the worst fit. The estimated return levels for spring MP show decreasing trends from the upper to the middle and then to the lower reaches totally speaking. Summer MP approximates to annual MP both in the quantity and in the spatial distributions. Autumn MP shows a little higher value in the estimated return levels than Spring MP, while keeping consistent with spring MP in the spatial distribution. It is also found that the estimated return levels for annual MP derived from various distributions differ by 22%, 36%, and 53% on average at 20-year, 50-year, and 100-year return periods, respectively. 12. A Mechanistic Beta-Binomial Probability Model for mRNA Sequencing Data. Science.gov (United States) Smith, Gregory R; Birtwistle, Marc R 2016-01-01 A main application for mRNA sequencing (mRNAseq) is determining lists of differentially-expressed genes (DEGs) between two or more conditions. Several software packages exist to produce DEGs from mRNAseq data, but they typically yield different DEGs, sometimes markedly so. The underlying probability model used to describe mRNAseq data is central to deriving DEGs, and not surprisingly most softwares use different models and assumptions to analyze mRNAseq data. Here, we propose a mechanistic justification to model mRNAseq as a binomial process, with data from technical replicates given by a binomial distribution, and data from biological replicates well-described by a beta-binomial distribution. We demonstrate good agreement of this model with two large datasets. We show that an emergent feature of the beta-binomial distribution, given parameter regimes typical for mRNAseq experiments, is the well-known quadratic polynomial scaling of variance with the mean. The so-called dispersion parameter controls this scaling, and our analysis suggests that the dispersion parameter is a continually decreasing function of the mean, as opposed to current approaches that impose an asymptotic value to the dispersion parameter at moderate mean read counts. We show how this leads to current approaches overestimating variance for moderately to highly expressed genes, which inflates false negative rates. Describing mRNAseq data with a beta-binomial distribution thus may be preferred since its parameters are relatable to the mechanistic underpinnings of the technique and may improve the consistency of DEG analysis across softwares, particularly for moderately to highly expressed genes. 13. Ruin probability with claims modeled by a stationary ergodic stable process NARCIS (Netherlands) Mikosch, T.; Samorodnitsky, G. 2000-01-01 For a random walk with negative drift we study the exceedance probability (ruin probability) of a high threshold. The steps of this walk (claim sizes) constitute a stationary ergodic stable process. We study how ruin occurs in this situation and evaluate the asymptotic behavior of the ruin 14. Sampling the stream landscape: Improving the applicability of an ecoregion-level capture probability model for stream fishes Science.gov (United States) Mollenhauer, Robert; Mouser, Joshua B.; Brewer, Shannon K. 2018-01-01 Temporal and spatial variability in streams result in heterogeneous gear capture probability (i.e., the proportion of available individuals identified) that confounds interpretation of data used to monitor fish abundance. We modeled tow-barge electrofishing capture probability at multiple spatial scales for nine Ozark Highland stream fishes. In addition to fish size, we identified seven reach-scale environmental characteristics associated with variable capture probability: stream discharge, water depth, conductivity, water clarity, emergent vegetation, wetted width–depth ratio, and proportion of riffle habitat. The magnitude of the relationship between capture probability and both discharge and depth varied among stream fishes. We also identified lithological characteristics among stream segments as a coarse-scale source of variable capture probability. The resulting capture probability model can be used to adjust catch data and derive reach-scale absolute abundance estimates across a wide range of sampling conditions with similar effort as used in more traditional fisheries surveys (i.e., catch per unit effort). Adjusting catch data based on variable capture probability improves the comparability of data sets, thus promoting both well-informed conservation and management decisions and advances in stream-fish ecology. 15. Investigating the Differences of Single-Vehicle and Multivehicle Accident Probability Using Mixed Logit Model Directory of Open Access Journals (Sweden) Bowen Dong 2018-01-01 Full Text Available Road traffic accidents are believed to be associated with not only road geometric feature and traffic characteristic, but also weather condition. To address these safety issues, it is of paramount importance to understand how these factors affect the occurrences of the crashes. Existing studies have suggested that the mechanisms of single-vehicle (SV accidents and multivehicle (MV accidents can be very different. Few studies were conducted to examine the difference of SV and MV accident probability by addressing unobserved heterogeneity at the same time. To investigate the different contributing factors on SV and MV, a mixed logit model is employed using disaggregated data with the response variable categorized as no accidents, SV accidents, and MV accidents. The results indicate that, in addition to speed gap, length of segment, and wet road surfaces which are significant for both SV and MV accidents, most of other variables are significant only for MV accidents. Traffic, road, and surface characteristics are main influence factors of SV and MV accident possibility. Hourly traffic volume, inside shoulder width, and wet road surface are found to produce statistically significant random parameters. Their effects on the possibility of SV and MV accident vary across different road segments. 16. Interval forecasting of cyberattack intensity on informatization objects of industry using probability cluster model Science.gov (United States) Krakovsky, Y. M.; Luzgin, A. N.; Mikhailova, E. A. 2018-05-01 At present, cyber-security issues associated with the informatization objects of industry occupy one of the key niches in the state management system. As a result of functional disruption of these systems via cyberattacks, an emergency may arise related to loss of life, environmental disasters, major financial and economic damage, or disrupted activities of cities and settlements. When cyberattacks occur with high intensity, in these conditions there is the need to develop protection against them, based on machine learning methods. This paper examines interval forecasting and presents results with a pre-set intensity level. The interval forecasting is carried out based on a probabilistic cluster model. This method involves forecasting of one of the two predetermined intervals in which a future value of the indicator will be located; probability estimates are used for this purpose. A dividing bound of these intervals is determined by a calculation method based on statistical characteristics of the indicator. Source data are used that includes a number of hourly cyberattacks using a honeypot from March to September 2013. 17. A Novel Adaptive Conditional Probability-Based Predicting Model for User’s Personality Traits Directory of Open Access Journals (Sweden) Mengmeng Wang 2015-01-01 Full Text Available With the pervasive increase in social media use, the explosion of users’ generated data provides a potentially very rich source of information, which plays an important role in helping online researchers understand user’s behaviors deeply. Since user’s personality traits are the driving force of user’s behaviors, hence, in this paper, along with social network features, we first extract linguistic features, emotional statistical features, and topic features from user’s Facebook status updates, followed by quantifying importance of features via Kendall correlation coefficient. And then, on the basis of weighted features and dynamic updated thresholds of personality traits, we deploy a novel adaptive conditional probability-based predicting model which considers prior knowledge of correlations between user’s personality traits to predict user’s Big Five personality traits. In the experimental work, we explore the existence of correlations between user’s personality traits which provides a better theoretical support for our proposed method. Moreover, on the same Facebook dataset, compared to other methods, our method can achieve an F1-measure of 80.6% when taking into account correlations between user’s personality traits, and there is an impressive improvement of 5.8% over other approaches. 18. Methane emission through ebullition from an estuarine mudflat: 2. Field observations and modeling of occurrence probability Science.gov (United States) Chen, Xi; Schäfer, Karina V. R.; Slater, Lee 2017-08-01 Ebullition can transport methane (CH4) at a much faster rate than other pathways, albeit over limited time and area, in wetland soils and sediments. However, field observations present large uncertainties in ebullition occurrences and statistic models are needed to describe the function relationship between probability of ebullition occurrence and water level changes. A flow-through chamber was designed and installed in a mudflat of an estuarine temperate marsh. Episodic increases in CH4 concentration signaling ebullition events were observed during ebbing tides (15 events over 456 ebbing tides) and occasionally during flooding tides (4 events over 455 flooding tides). Ebullition occurrence functions were defined using logistic regression as the relative initial and end water levels, as well as tidal amplitudes were found to be the key functional variables related to ebullition events. Ebullition of methane was restricted by a surface frozen layer during winter; melting of this layer during spring thaw caused increases in CH4 concentration, with ebullition fluxes similar to those associated with large fluctuations in water level around spring tides. Our findings suggest that initial and end relative water levels, in addition to tidal amplitude, partly regulate ebullition events in tidal wetlands, modulated by the lunar cycle, storage of gas bubbles at different depths and seasonal changes in the surface frozen layer. Maximum tidal strength over a few days, rather than hourly water level, may be more closely associated with the possibility of ebullition occurrence as it represents a trade-off time scale in between hourly and lunar periods. 19. Three-dimensional radiotherapy planning system for esophageal tumors: comparison of treatment techniques and analysis of probability of complications; Planejamento tridimensional para radioterapia de tumores de esofago: comparacao de tecnicas de tratamento e analise de probabilidade de complicacoes Energy Technology Data Exchange (ETDEWEB) Justino, Pitagoras Baskara; Carvalho, Heloisa de Andrade; Ferauche, Debora; Ros, Renato [Sao Paulo Uni., SP (Brazil). Hospital das Clinicas. Instituto de Radioterapia (InRad)]. E-mail: pitagorasb@hotmail.com 2003-06-01 Radiotherapy techniques for esophageal cancer were compared using a three-dimensional planning system. We studied the following treatment techniques used for a patient with squamous cell carcinoma of the middle third of the esophagus: two antero-posterior and two latero-lateral parallel opposed fields, three fields ('Y' and 'T'), and four fields ('X'). Dose-volume histograms were obtained considering spinal cord and lungs as organs at risk. Analysis was performed comparing doses in these organs as recommended by the Normal Tissue Complication Probability (NTCP) and Tumor Control Probability (TCP). When only the lungs were considered the best technique was two antero-posterior parallel opposed fields. The spinal cord was best protected using latero-lateral fields. We suggest the combination of at least two treatment techniques: antero-posterior fields with 'Y' or 'T' techniques, or latero-lateral fields in order to balance the doses in the lungs and the spinal cord. Another option may be the use of any of the three-field techniques during the whole treatment. (author) 20. Toward Big Data Analytics: Review of Predictive Models in Management of Diabetes and Its Complications. Science.gov (United States) Cichosz, Simon Lebech; Johansen, Mette Dencker; Hejlesen, Ole 2015-10-14 Diabetes is one of the top priorities in medical science and health care management, and an abundance of data and information is available on these patients. Whether data stem from statistical models or complex pattern recognition models, they may be fused into predictive models that combine patient information and prognostic outcome results. Such knowledge could be used in clinical decision support, disease surveillance, and public health management to improve patient care. Our aim was to review the literature and give an introduction to predictive models in screening for and the management of prevalent short- and long-term complications in diabetes. Predictive models have been developed for management of diabetes and its complications, and the number of publications on such models has been growing over the past decade. Often multiple logistic or a similar linear regression is used for prediction model development, possibly owing to its transparent functionality. Ultimately, for prediction models to prove useful, they must demonstrate impact, namely, their use must generate better patient outcomes. Although extensive effort has been put in to building these predictive models, there is a remarkable scarcity of impact studies. © 2015 Diabetes Technology Society. 1. Interpretations of probability CERN Document Server Khrennikov, Andrei 2009-01-01 This is the first fundamental book devoted to non-Kolmogorov probability models. It provides a mathematical theory of negative probabilities, with numerous applications to quantum physics, information theory, complexity, biology and psychology. The book also presents an interesting model of cognitive information reality with flows of information probabilities, describing the process of thinking, social, and psychological phenomena. 2. Development of a new model to evaluate the probability of automatic plant trips for pressurized water reactors Energy Technology Data Exchange (ETDEWEB) Shimada, Yoshio [Institute of Nuclear Safety System Inc., Mihama, Fukui (Japan); Kawai, Katsunori; Suzuki, Hiroshi [Mitsubishi Heavy Industries Ltd., Tokyo (Japan) 2001-09-01 In order to improve the reliability of plant operations for pressurized water reactors, a new fault tree model was developed to evaluate the probability of automatic plant trips. This model consists of fault trees for sixteen systems. It has the following features: (1) human errors and transmission line incidents are modeled by the existing data, (2) the repair of failed components is considered to calculate the failure probability of components, (3) uncertainty analysis is performed by an exact method. From the present results, it is confirmed that the obtained upper and lower bound values of the automatic plant trip probability are within the existing data bound in Japan. Thereby this model can be applicable to the prediction of plant performance and reliability. (author) 3. Quantitative non-monotonic modeling of economic uncertainty by probability and possibility distributions DEFF Research Database (Denmark) Schjær-Jacobsen, Hans 2012-01-01 uncertainty can be calculated. The possibility approach is particular well suited for representation of uncertainty of a non-statistical nature due to lack of knowledge and requires less information than the probability approach. Based on the kind of uncertainty and knowledge present, these aspects...... to the understanding of similarities and differences of the two approaches as well as practical applications. The probability approach offers a good framework for representation of randomness and variability. Once the probability distributions of uncertain parameters and their correlations are known the resulting...... are thoroughly discussed in the case of rectangular representation of uncertainty by the uniform probability distribution and the interval, respectively. Also triangular representations are dealt with and compared. Calculation of monotonic as well as non-monotonic functions of variables represented... 4. Path Loss, Shadow Fading, and Line-Of-Sight Probability Models for 5G Urban Macro-Cellular Scenarios DEFF Research Database (Denmark) Sun, Shu; Thomas, Timothy; Rappaport, Theodore S. 2015-01-01 This paper presents key parameters including the line-of-sight (LOS) probability, large-scale path loss, and shadow fading models for the design of future fifth generation (5G) wireless communication systems in urban macro-cellular (UMa) scenarios, using the data obtained from propagation...... measurements in Austin, US, and Aalborg, Denmark, at 2, 10, 18, and 38 GHz. A comparison of different LOS probability models is performed for the Aalborg environment. Both single-slope and dual-slope omnidirectional path loss models are investigated to analyze and contrast their root-mean-square (RMS) errors... 5. Comparison of prospective risk estimates for postoperative complications: human vs computer model. Science.gov (United States) Glasgow, Robert E; Hawn, Mary T; Hosokawa, Patrick W; Henderson, William G; Min, Sung-Joon; Richman, Joshua S; Tomeh, Majed G; Campbell, Darrell; Neumayer, Leigh A 2014-02-01 Surgical quality improvement tools such as NSQIP are limited in their ability to prospectively affect individual patient care by the retrospective audit and feedback nature of their design. We hypothesized that statistical models using patient preoperative characteristics could prospectively provide risk estimates of postoperative adverse events comparable to risk estimates provided by experienced surgeons, and could be useful for stratifying preoperative assessment of patient risk. This was a prospective observational cohort. Using previously developed models for 30-day postoperative mortality, overall morbidity, cardiac, thromboembolic, pulmonary, renal, and surgical site infection (SSI) complications, model and surgeon estimates of risk were compared with each other and with actual 30-day outcomes. The study cohort included 1,791 general surgery patients operated on between June 2010 and January 2012. Observed outcomes were mortality (0.2%), overall morbidity (8.2%), and pulmonary (1.3%), cardiac (0.3%), thromboembolism (0.2%), renal (0.4%), and SSI (3.8%) complications. Model and surgeon risk estimates showed significant correlation (p risk for overall morbidity to be low, the model-predicted risk and observed morbidity rates were 2.8% and 4.1%, respectively, compared with 10% and 18% in perceived high risk patients. Patients in the highest quartile of model-predicted risk accounted for 75% of observed mortality and 52% of morbidity. Across a broad range of general surgical operations, we confirmed that the model risk estimates are in fairly good agreement with risk estimates of experienced surgeons. Using these models prospectively can identify patients at high risk for morbidity and mortality, who could then be targeted for intervention to reduce postoperative complications. Published by Elsevier Inc. 6. Modeling the probability distribution of positional errors incurred by residential address geocoding Directory of Open Access Journals (Sweden) Mazumdar Soumya 2007-01-01 Full Text Available Abstract Background The assignment of a point-level geocode to subjects' residences is an important data assimilation component of many geographic public health studies. Often, these assignments are made by a method known as automated geocoding, which attempts to match each subject's address to an address-ranged street segment georeferenced within a streetline database and then interpolate the position of the address along that segment. Unfortunately, this process results in positional errors. Our study sought to model the probability distribution of positional errors associated with automated geocoding and E911 geocoding. Results Positional errors were determined for 1423 rural addresses in Carroll County, Iowa as the vector difference between each 100%-matched automated geocode and its true location as determined by orthophoto and parcel information. Errors were also determined for 1449 60%-matched geocodes and 2354 E911 geocodes. Huge (> 15 km outliers occurred among the 60%-matched geocoding errors; outliers occurred for the other two types of geocoding errors also but were much smaller. E911 geocoding was more accurate (median error length = 44 m than 100%-matched automated geocoding (median error length = 168 m. The empirical distributions of positional errors associated with 100%-matched automated geocoding and E911 geocoding exhibited a distinctive Greek-cross shape and had many other interesting features that were not capable of being fitted adequately by a single bivariate normal or t distribution. However, mixtures of t distributions with two or three components fit the errors very well. Conclusion Mixtures of bivariate t distributions with few components appear to be flexible enough to fit many positional error datasets associated with geocoding, yet parsimonious enough to be feasible for nascent applications of measurement-error methodology to spatial epidemiology. 7. Characterization of vascular complications in experimental model of fructose-induced metabolic syndrome. Science.gov (United States) El-Bassossy, Hany M; Dsokey, Nora; Fahmy, Ahmed 2014-12-01 Vascular dysfunction is an important complication associated with metabolic syndrome (MS). Here we fully characterized vascular complications in a rat model of fructose-induced MS. MS was induced by adding fructose (10%) to drinking water to male Wistar rats of 6 weeks age. Blood pressure (BP) and isolated aorta responses phenylephrine (PE), KCl, acetylcholine (ACh), and sodium nitroprusside (SNP) were recorded after 6, 9, and 12 weeks of fructose administration. In addition, serum levels of glucose, insulin, uric acid, tumor necrosis factor α (TNFα), lipids, advanced glycation end products (AGEs), and arginase activity were determined. Furthermore, aortic reactive oxygen species (ROS) generation, hemeoxygenase-1 expression, and collagen deposition were examined. Fructose administration resulted in a significant hyperinslinemia after 6 weeks which continued for 12 weeks. It was also associated with a significant increase in BP after 6 weeks which was stable for 12 weeks. Aorta isolated from MS animals showed exaggerated contractility to PE and KCl and impaired relaxation to ACh compared with control after 6 weeks which were clearer at 12 weeks of fructose administration. In addition, MS animals showed significant increases in serum levels of lipids, uric acid, AGEs, TNFα, and arginase enzyme activity after 12 weeks of fructose administration. Furthermore, aortae isolated from MS animals were characterized by increased ROS generation and collagen deposition. In conclusion, adding fructose (10%) to drinking water produces a model of MS with vascular complications after 12 weeks that are characterized by insulin resistance, hypertension, disturbed vascular reactivity and structure, hyperuricemia, dyslipidemia, and low-grade inflammation. 8. Modeling spatial variability of sand-lenses in clay till settings using transition probability and multiple-point geostatistics DEFF Research Database (Denmark) Kessler, Timo Christian; Nilsson, Bertel; Klint, Knud Erik 2010-01-01 (TPROGS) of alternating geological facies. The second method, multiple-point statistics, uses training images to estimate the conditional probability of sand-lenses at a certain location. Both methods respect field observations such as local stratigraphy, however, only the multiple-point statistics can...... of sand-lenses in clay till. Sand-lenses mainly account for horizontal transport and are prioritised in this study. Based on field observations, the distribution has been modeled using two different geostatistical approaches. One method uses a Markov chain model calculating the transition probabilities... 9. Peristomal Skin Complications Are Common, Expensive, and Difficult to Manage: A Population Based Cost Modeling Study Science.gov (United States) Meisner, Søren; Lehur, Paul-Antoine; Moran, Brendan; Martins, Lina; Jemec, Gregor Borut Ernst 2012-01-01 Background Peristomal skin complications (PSCs) are the most common post-operative complications following creation of a stoma. Living with a stoma is a challenge, not only for the patient and their carers, but also for society as a whole. Due to methodological problems of PSC assessment, the associated health-economic burden of medium to longterm complications has been poorly described. Aim The aim of the present study was to create a model to estimate treatment costs of PSCs using the standardized assessment Ostomy Skin Tool as a reference. The resultant model was applied to a real-life global data set of stoma patients (n = 3017) to determine the prevalence and financial burden of PSCs. Methods Eleven experienced stoma care nurses were interviewed to get a global understanding of a treatment algorithm that formed the basis of the cost analysis. The estimated costs were based on a seven week treatment period. PSC costs were estimated for five underlying diagnostic categories and three levels of severity. The estimated treatment costs of severe cases of PSCs were increased 2–5 fold for the different diagnostic categories of PSCs compared with mild cases. French unit costs were applied to the global data set. Results The estimated total average cost for a seven week treatment period (including appliances and accessories) was 263€ for those with PSCs (n = 1742) compared to 215€ for those without PSCs (n = 1172). A co-variance analysis showed that leakage level had a significant impact on PSC cost from ‘rarely/never’ to ‘always/often’ p<0.00001 and from ‘rarely/never’ to ‘sometimes’ p = 0.0115. Conclusion PSCs are common and troublesome and the consequences are substantial, both for the patient and from a health economic viewpoint. PSCs should be diagnosed and treated at an early stage to prevent long term, debilitating and expensive complications. PMID:22679479 10. Peristomal skin complications are common, expensive, and difficult to manage: a population based cost modeling study. Directory of Open Access Journals (Sweden) Søren Meisner Full Text Available BACKGROUND: Peristomal skin complications (PSCs are the most common post-operative complications following creation of a stoma. Living with a stoma is a challenge, not only for the patient and their carers, but also for society as a whole. Due to methodological problems of PSC assessment, the associated health-economic burden of medium to longterm complications has been poorly described. AIM: The aim of the present study was to create a model to estimate treatment costs of PSCs using the standardized assessment Ostomy Skin Tool as a reference. The resultant model was applied to a real-life global data set of stoma patients (n = 3017 to determine the prevalence and financial burden of PSCs. METHODS: Eleven experienced stoma care nurses were interviewed to get a global understanding of a treatment algorithm that formed the basis of the cost analysis. The estimated costs were based on a seven week treatment period. PSC costs were estimated for five underlying diagnostic categories and three levels of severity. The estimated treatment costs of severe cases of PSCs were increased 2-5 fold for the different diagnostic categories of PSCs compared with mild cases. French unit costs were applied to the global data set. RESULTS: The estimated total average cost for a seven week treatment period (including appliances and accessories was 263€ for those with PSCs (n = 1742 compared to 215€ for those without PSCs (n = 1172. A co-variance analysis showed that leakage level had a significant impact on PSC cost from 'rarely/never' to 'always/often' p<0.00001 and from 'rarely/never' to 'sometimes' p = 0.0115. CONCLUSION: PSCs are common and troublesome and the consequences are substantial, both for the patient and from a health economic viewpoint. PSCs should be diagnosed and treated at an early stage to prevent long term, debilitating and expensive complications. 11. Delay or probability discounting in a model of impulsive behavior: effect of alcohol. Science.gov (United States) Richards, J B; Zhang, L; Mitchell, S H; de Wit, H 1999-01-01 Little is known about the acute effects of drugs of abuse on impulsivity and self-control. In this study, impulsivity was assessed in humans using a computer task that measured delay and probability discounting. Discounting describes how much the value of a reward (or punisher) is decreased when its occurrence is either delayed or uncertain. Twenty-four healthy adult volunteers ingested a moderate dose of ethanol (0.5 or 0.8 g/kg ethanol: n = 12 at each dose) or placebo before completing the discounting task. In the task the participants were given a series of choices between a small, immediate, certain amount of money and $10 that was either delayed (0, 2, 30, 180, or 365 days) or probabilistic (i.e., certainty of receipt was 1.0, .9, .75, .5, or .25). The point at which each individual was indifferent between the smaller immediate or certain reward and the$10 delayed or probabilistic reward was identified using an adjusting-amount procedure. The results indicated that (a) delay and probability discounting were well described by a hyperbolic function; (b) delay and probability discounting were positively correlated within subjects; (c) delay and probability discounting were moderately correlated with personality measures of impulsivity; and (d) alcohol had no effect on discounting. PMID:10220927 12. Modelling the impact of creep on the probability of failure of a solid oxidefuel cell stack DEFF Research Database (Denmark) Greco, Fabio; Frandsen, Henrik Lund; Nakajo, Arata 2014-01-01 In solid oxide fuel cell (SOFC) technology a major challenge lies in balancing thermal stresses from an inevitable thermal field. The cells are known to creep, changing over time the stress field. The main objective of this study was to assess the influence of creep on the failure probability of ... 13. Multivariate modeling of complications with data driven variable selection: Guarding against overfitting and effects of data set size International Nuclear Information System (INIS) Schaaf, Arjen van der; Xu Chengjian; Luijk, Peter van; Veld, Aart A. van’t; Langendijk, Johannes A.; Schilstra, Cornelis 2012-01-01 Purpose: Multivariate modeling of complications after radiotherapy is frequently used in conjunction with data driven variable selection. This study quantifies the risk of overfitting in a data driven modeling method using bootstrapping for data with typical clinical characteristics, and estimates the minimum amount of data needed to obtain models with relatively high predictive power. Materials and methods: To facilitate repeated modeling and cross-validation with independent datasets for the assessment of true predictive power, a method was developed to generate simulated data with statistical properties similar to real clinical data sets. Characteristics of three clinical data sets from radiotherapy treatment of head and neck cancer patients were used to simulate data with set sizes between 50 and 1000 patients. A logistic regression method using bootstrapping and forward variable selection was used for complication modeling, resulting for each simulated data set in a selected number of variables and an estimated predictive power. The true optimal number of variables and true predictive power were calculated using cross-validation with very large independent data sets. Results: For all simulated data set sizes the number of variables selected by the bootstrapping method was on average close to the true optimal number of variables, but showed considerable spread. Bootstrapping is more accurate in selecting the optimal number of variables than the AIC and BIC alternatives, but this did not translate into a significant difference of the true predictive power. The true predictive power asymptotically converged toward a maximum predictive power for large data sets, and the estimated predictive power converged toward the true predictive power. More than half of the potential predictive power is gained after approximately 200 samples. Our simulations demonstrated severe overfitting (a predicative power lower than that of predicting 50% probability) in a number of small 14. The Surgical Mortality Probability Model: derivation and validation of a simple risk prediction rule for noncardiac surgery. Science.gov (United States) Glance, Laurent G; Lustik, Stewart J; Hannan, Edward L; Osler, Turner M; Mukamel, Dana B; Qian, Feng; Dick, Andrew W 2012-04-01 To develop a 30-day mortality risk index for noncardiac surgery that can be used to communicate risk information to patients and guide clinical management at the "point-of-care," and that can be used by surgeons and hospitals to internally audit their quality of care. Clinicians rely on the Revised Cardiac Risk Index to quantify the risk of cardiac complications in patients undergoing noncardiac surgery. Because mortality from noncardiac causes accounts for many perioperative deaths, there is also a need for a simple bedside risk index to predict 30-day all-cause mortality after noncardiac surgery. Retrospective cohort study of 298,772 patients undergoing noncardiac surgery during 2005 to 2007 using the American College of Surgeons National Surgical Quality Improvement Program database. The 9-point S-MPM (Surgical Mortality Probability Model) 30-day mortality risk index was derived empirically and includes three risk factors: ASA (American Society of Anesthesiologists) physical status, emergency status, and surgery risk class. Patients with ASA physical status I, II, III, IV or V were assigned either 0, 2, 4, 5, or 6 points, respectively; intermediate- or high-risk procedures were assigned 1 or 2 points, respectively; and emergency procedures were assigned 1 point. Patients with risk scores less than 5 had a predicted risk of mortality less than 0.50%, whereas patients with a risk score of 5 to 6 had a risk of mortality between 1.5% and 4.0%. Patients with a risk score greater than 6 had risk of mortality more than 10%. S-MPM exhibited excellent discrimination (C statistic, 0.897) and acceptable calibration (Hosmer-Lemeshow statistic 13.0, P = 0.023) in the validation data set. Thirty-day mortality after noncardiac surgery can be accurately predicted using a simple and accurate risk score based on information readily available at the bedside. This risk index may play a useful role in facilitating shared decision making, developing and implementing risk 15. Modeling and advanced sliding mode controls of crawler cranes considering wire rope elasticity and complicated operations Science.gov (United States) Tuan, Le Anh; Lee, Soon-Geul 2018-03-01 In this study, a new mathematical model of crawler cranes is developed for heavy working conditions, with payload-lifting and boom-hoisting motions simultaneously activated. The system model is built with full consideration of wind disturbances, geometrical nonlinearities, and cable elasticities of cargo lifting and boom luffing. On the basis of this dynamic model, three versions of sliding mode control are analyzed and designed to control five system outputs with only two inputs. When used in complicated operations, the effectiveness of the controllers is analyzed using analytical investigation and numerical simulation. Results indicate the effectiveness of the control algorithms and the proposed dynamic model. The control algorithms asymptotically stabilize the system with finite-time convergences, remaining robust amid disturbances and parametric uncertainties. 16. Developing a Mathematical Model for Scheduling and Determining Success Probability of Research Projects Considering Complex-Fuzzy Networks Directory of Open Access Journals (Sweden) Gholamreza Norouzi 2015-01-01 Full Text Available In project management context, time management is one of the most important factors affecting project success. This paper proposes a new method to solve research project scheduling problems (RPSP containing Fuzzy Graphical Evaluation and Review Technique (FGERT networks. Through the deliverables of this method, a proper estimation of project completion time (PCT and success probability can be achieved. So algorithms were developed to cover all features of the problem based on three main parameters “duration, occurrence probability, and success probability.” These developed algorithms were known as PR-FGERT (Parallel and Reversible-Fuzzy GERT networks. The main provided framework includes simplifying the network of project and taking regular steps to determine PCT and success probability. Simplifications include (1 equivalent making of parallel and series branches in fuzzy network considering the concepts of probabilistic nodes, (2 equivalent making of delay or reversible-to-itself branches and impact of changing the parameters of time and probability based on removing related branches, (3 equivalent making of simple and complex loops, and (4 an algorithm that was provided to resolve no-loop fuzzy network, after equivalent making. Finally, the performance of models was compared with existing methods. The results showed proper and real performance of models in comparison with existing methods. 17. Deregulation of arginase induces bone complications in high-fat/high-sucrose diet diabetic mouse model. Science.gov (United States) Bhatta, Anil; Sangani, Rajnikumar; Kolhe, Ravindra; Toque, Haroldo A; Cain, Michael; Wong, Abby; Howie, Nicole; Shinde, Rahul; Elsalanty, Mohammed; Yao, Lin; Chutkan, Norman; Hunter, Monty; Caldwell, Ruth B; Isales, Carlos; Caldwell, R William; Fulzele, Sadanand 2016-02-15 A balanced diet is crucial for healthy development and prevention of musculoskeletal related diseases. Diets high in fat content are known to cause obesity, diabetes and a number of other disease states. Our group and others have previously reported that activity of the urea cycle enzyme arginase is involved in diabetes-induced dysregulation of vascular function due to decreases in nitric oxide formation. We hypothesized that diabetes may also elevate arginase activity in bone and bone marrow, which could lead to bone-related complications. To test this we determined the effects of diabetes on expression and activity of arginase, in bone and bone marrow stromal cells (BMSCs). We demonstrated that arginase 1 is abundantly present in the bone and BMSCs. We also demonstrated that arginase activity and expression in bone and bone marrow is up-regulated in models of diabetes induced by HFHS diet and streptozotocin (STZ). HFHS diet down-regulated expression of healthy bone metabolism markers (BMP2, COL-1, ALP, and RUNX2) and reduced bone mineral density, bone volume and trabecular thickness. However, treatment with an arginase inhibitor (ABH) prevented these bone-related complications of diabetes. In-vitro study of BMSCs showed that high glucose treatment increased arginase activity and decreased nitric oxide production. These effects were reversed by treatment with an arginase inhibitor (ABH). Our study provides evidence that deregulation of l-arginine metabolism plays a vital role in HFHS diet-induced diabetic complications and that these complications can be prevented by treatment with arginase inhibitors. The modulation of l-arginine metabolism in disease could offer a novel therapeutic approach for osteoporosis and other musculoskeletal related diseases. Published by Elsevier Ireland Ltd. 18. A combined Importance Sampling and Kriging reliability method for small failure probabilities with time-demanding numerical models International Nuclear Information System (INIS) Echard, B.; Gayton, N.; Lemaire, M.; Relun, N. 2013-01-01 Applying reliability methods to a complex structure is often delicate for two main reasons. First, such a structure is fortunately designed with codified rules leading to a large safety margin which means that failure is a small probability event. Such a probability level is difficult to assess efficiently. Second, the structure mechanical behaviour is modelled numerically in an attempt to reproduce the real response and numerical model tends to be more and more time-demanding as its complexity is increased to improve accuracy and to consider particular mechanical behaviour. As a consequence, performing a large number of model computations cannot be considered in order to assess the failure probability. To overcome these issues, this paper proposes an original and easily implementable method called AK-IS for active learning and Kriging-based Importance Sampling. This new method is based on the AK-MCS algorithm previously published by Echard et al. [AK-MCS: an active learning reliability method combining Kriging and Monte Carlo simulation. Structural Safety 2011;33(2):145–54]. It associates the Kriging metamodel and its advantageous stochastic property with the Importance Sampling method to assess small failure probabilities. It enables the correction or validation of the FORM approximation with only a very few mechanical model computations. The efficiency of the method is, first, proved on two academic applications. It is then conducted for assessing the reliability of a challenging aerospace case study submitted to fatigue. 19. Probability model of solid to liquid-like transition of a fluid suspension after a shear flow onset Czech Academy of Sciences Publication Activity Database Nouar, C.; Říha, Pavel 2008-01-01 Roč. 34, č. 5 (2008), s. 477-483 ISSN 0301-9322 R&D Projects: GA AV ČR IAA200600803 Institutional research plan: CEZ:AV0Z20600510 Keywords : laminar suspension flow * liquid-liquid interface * probability model Subject RIV: BK - Fluid Dynamics Impact factor: 1.497, year: 2008 20. Optimal selection for BRCA1 and BRCA2 mutation testing using a combination of ' easy to apply ' probability models NARCIS (Netherlands) Bodmer, D.; Ligtenberg, M. J. L.; van der Hout, A. H.; Gloudemans, S.; Ansink, K.; Oosterwijk, J. C.; Hoogerbrugge, N. 2006-01-01 To establish an efficient, reliable and easy to apply risk assessment tool to select families with breast and/or ovarian cancer patients for BRCA mutation testing, using available probability models. In a retrospective study of 263 families with breast and/or ovarian cancer patients, the utility of 1. Conceptual and Statistical Issues Regarding the Probability of Default and Modeling Default Risk Directory of Open Access Journals (Sweden) Emilia TITAN 2011-03-01 Full Text Available In today’s rapidly evolving financial markets, risk management offers different techniques in order to implement an efficient system against market risk. Probability of default (PD is an essential part of business intelligence and customer relation management systems in the financial institutions. Recent studies indicates that underestimating this important component, and also the loss given default (LGD, might threaten the stability and smooth running of the financial markets. From the perspective of risk management, the result of predictive accuracy of the estimated probability of default is more valuable than the standard binary classification: credible or non credible clients. The Basle II Accord recognizes the methods of reducing credit risk and also PD and LGD as important components of advanced Internal Rating Based (IRB approach. 2. Generalized Probability-Probability Plots NARCIS (Netherlands) Mushkudiani, N.A.; Einmahl, J.H.J. 2004-01-01 We introduce generalized Probability-Probability (P-P) plots in order to study the one-sample goodness-of-fit problem and the two-sample problem, for real valued data.These plots, that are constructed by indexing with the class of closed intervals, globally preserve the properties of classical P-P 3. Determination of probability density functions for parameters in the Munson-Dawson model for creep behavior of salt International Nuclear Information System (INIS) Pfeifle, T.W.; Mellegard, K.D.; Munson, D.E. 1992-10-01 The modified Munson-Dawson (M-D) constitutive model that describes the creep behavior of salt will be used in performance assessment calculations to assess compliance of the Waste Isolation Pilot Plant (WIPP) facility with requirements governing the disposal of nuclear waste. One of these standards requires that the uncertainty of future states of the system, material model parameters, and data be addressed in the performance assessment models. This paper presents a method in which measurement uncertainty and the inherent variability of the material are characterized by treating the M-D model parameters as random variables. The random variables can be described by appropriate probability distribution functions which then can be used in Monte Carlo or structural reliability analyses. Estimates of three random variables in the M-D model were obtained by fitting a scalar form of the model to triaxial compression creep data generated from tests of WIPP salt. Candidate probability distribution functions for each of the variables were then fitted to the estimates and their relative goodness-of-fit tested using the Kolmogorov-Smirnov statistic. A sophisticated statistical software package obtained from BMDP Statistical Software, Inc. was used in the M-D model fitting. A separate software package, STATGRAPHICS, was used in fitting the candidate probability distribution functions to estimates of the variables. Skewed distributions, i.e., lognormal and Weibull, were found to be appropriate for the random variables analyzed 4. Probability, Nondeterminism and Concurrency DEFF Research Database (Denmark) Varacca, Daniele Nondeterminism is modelled in domain theory by the notion of a powerdomain, while probability is modelled by that of the probabilistic powerdomain. Some problems arise when we want to combine them in order to model computation in which both nondeterminism and probability are present. In particula... 5. Probability-1 CERN Document Server Shiryaev, Albert N 2016-01-01 This book contains a systematic treatment of probability from the ground up, starting with intuitive ideas and gradually developing more sophisticated subjects, such as random walks, martingales, Markov chains, the measure-theoretic foundations of probability theory, weak convergence of probability measures, and the central limit theorem. Many examples are discussed in detail, and there are a large number of exercises. The book is accessible to advanced undergraduates and can be used as a text for independent study. To accommodate the greatly expanded material in the third edition of Probability, the book is now divided into two volumes. This first volume contains updated references and substantial revisions of the first three chapters of the second edition. In particular, new material has been added on generating functions, the inclusion-exclusion principle, theorems on monotonic classes (relying on a detailed treatment of “π-λ” systems), and the fundamental theorems of mathematical statistics. 6. Ignition Probability Data.gov (United States) Earth Data Analysis Center, University of New Mexico — USFS, State Forestry, BLM, and DOI fire occurrence point locations from 1987 to 2008 were combined and converted into a fire occurrence probability or density grid... 7. Probability of bystander effect induced by alpha-particles emitted by radon progeny using the analytical model of tracheobronchial tree International Nuclear Information System (INIS) Jovanovic, B.; Nikezic, D. 2010-01-01 Radiation-induced biological bystander effects have become a phenomenon associated with the interaction of radiation with cells. There is a need to include the influence of biological effects in the dosimetry of the human lung. With this aim, the purpose of this work is to calculate the probability of bystander effect induced by alpha-particle radiation on sensitive cells of the human lung. Probability was calculated by applying the analytical model cylinder bifurcation, which was created to simulate the geometry of the human lung with the geometric distribution of cell nuclei in the airway wall of the tracheobronchial tree. This analytical model of the human tracheobronchial tree represents the extension of the ICRP 66 model, and follows it as much as possible. Reported probabilities are calculated for various targets and alpha-particle energies. Probability of bystander effect has been calculated for alpha particles with 6 and 7.69 MeV energies, which are emitted in the 222 Rn chain. The application of these results may enhance current dose risk estimation approaches in the sense of the inclusion of the influence of the biological effects. (authors) 8. Combining 3d Volume and Mesh Models for Representing Complicated Heritage Buildings Science.gov (United States) Tsai, F.; Chang, H.; Lin, Y.-W. 2017-08-01 This study developed a simple but effective strategy to combine 3D volume and mesh models for representing complicated heritage buildings and structures. The idea is to seamlessly integrate 3D parametric or polyhedral models and mesh-based digital surfaces to generate a hybrid 3D model that can take advantages of both modeling methods. The proposed hybrid model generation framework is separated into three phases. Firstly, after acquiring or generating 3D point clouds of the target, these 3D points are partitioned into different groups. Secondly, a parametric or polyhedral model of each group is generated based on plane and surface fitting algorithms to represent the basic structure of that region. A "bare-bones" model of the target can subsequently be constructed by connecting all 3D volume element models. In the third phase, the constructed bare-bones model is used as a mask to remove points enclosed by the bare-bones model from the original point clouds. The remaining points are then connected to form 3D surface mesh patches. The boundary points of each surface patch are identified and these boundary points are projected onto the surfaces of the bare-bones model. Finally, new meshes are created to connect the projected points and original mesh boundaries to integrate the mesh surfaces with the 3D volume model. The proposed method was applied to an open-source point cloud data set and point clouds of a local historical structure. Preliminary results indicated that the reconstructed hybrid models using the proposed method can retain both fundamental 3D volume characteristics and accurate geometric appearance with fine details. The reconstructed hybrid models can also be used to represent targets in different levels of detail according to user and system requirements in different applications. 9. Development of a methodology for probable maximum precipitation estimation over the American River watershed using the WRF model Science.gov (United States) Tan, Elcin A new physically-based methodology for probable maximum precipitation (PMP) estimation is developed over the American River Watershed (ARW) using the Weather Research and Forecast (WRF-ARW) model. A persistent moisture flux convergence pattern, called Pineapple Express, is analyzed for 42 historical extreme precipitation events, and it is found that Pineapple Express causes extreme precipitation over the basin of interest. An average correlation between moisture flux convergence and maximum precipitation is estimated as 0.71 for 42 events. The performance of the WRF model is verified for precipitation by means of calibration and independent validation of the model. The calibration procedure is performed only for the first ranked flood event 1997 case, whereas the WRF model is validated for 42 historical cases. Three nested model domains are set up with horizontal resolutions of 27 km, 9 km, and 3 km over the basin of interest. As a result of Chi-square goodness-of-fit tests, the hypothesis that "the WRF model can be used in the determination of PMP over the ARW for both areal average and point estimates" is accepted at the 5% level of significance. The sensitivities of model physics options on precipitation are determined using 28 microphysics, atmospheric boundary layer, and cumulus parameterization schemes combinations. It is concluded that the best triplet option is Thompson microphysics, Grell 3D ensemble cumulus, and YSU boundary layer (TGY), based on 42 historical cases, and this TGY triplet is used for all analyses of this research. Four techniques are proposed to evaluate physically possible maximum precipitation using the WRF: 1. Perturbations of atmospheric conditions; 2. Shift in atmospheric conditions; 3. Replacement of atmospheric conditions among historical events; and 4. Thermodynamically possible worst-case scenario creation. Moreover, climate change effect on precipitation is discussed by emphasizing temperature increase in order to determine the 10. Hydrological model calibration for derived flood frequency analysis using stochastic rainfall and probability distributions of peak flows Science.gov (United States) 2014-01-01 Derived flood frequency analysis allows the estimation of design floods with hydrological modeling for poorly observed basins considering change and taking into account flood protection measures. There are several possible choices regarding precipitation input, discharge output and consequently the calibration of the model. The objective of this study is to compare different calibration strategies for a hydrological model considering various types of rainfall input and runoff output data sets and to propose the most suitable approach. Event based and continuous, observed hourly rainfall data as well as disaggregated daily rainfall and stochastically generated hourly rainfall data are used as input for the model. As output, short hourly and longer daily continuous flow time series as well as probability distributions of annual maximum peak flow series are employed. The performance of the strategies is evaluated using the obtained different model parameter sets for continuous simulation of discharge in an independent validation period and by comparing the model derived flood frequency distributions with the observed one. The investigations are carried out for three mesoscale catchments in northern Germany with the hydrological model HEC-HMS (Hydrologic Engineering Center's Hydrologic Modeling System). The results show that (I) the same type of precipitation input data should be used for calibration and application of the hydrological model, (II) a model calibrated using a small sample of extreme values works quite well for the simulation of continuous time series with moderate length but not vice versa, and (III) the best performance with small uncertainty is obtained when stochastic precipitation data and the observed probability distribution of peak flows are used for model calibration. This outcome suggests to calibrate a hydrological model directly on probability distributions of observed peak flows using stochastic rainfall as input if its purpose is the 11. Inpatient Consults and Complications During Primary Total Joint Arthroplasty in a Bundled Care Model. Science.gov (United States) Baumgartner, Billy T; Karas, Vasili; Kildow, Beau J; Cunningham, Daniel J; Klement, Mitchell R; Green, Cindy L; Attarian, David E; Seyler, Thorsten M 2018-04-01 The Centers for Medicare and Medicaid Services (CMS) are implementing changes in hospital reimbursement models for total joint arthroplasty (TJA), moving to value-based bundled payments from the fee-for-service model. The purpose of this study is to identify consults and complications during the perioperative period that increase financial burden. We combined CMS payment data for inpatient, professional, and postoperative with retrospective review of patients undergoing primary TJA and developed profiles of patients included in the Comprehensive Care for Joint Replacement bundle undergoing TJA. Statistical comparison of episode inpatient events and payments was conducted. Multiple regression analysis was adjusted for length of stay, disposition, and Charlson-Deyo comorbidity profile. Median total payment was $21,577.36, which exceeded the median bundle target payment of$20,625.00. Adjusted analyses showed that psychiatry consults (increase of $73,123.32; P care unit admission ($14,078.37; P care unit admission, and medical/psychiatric consultation exceeded the CMS target. Although study results showed typical complication rates, acute inpatient consultation significantly increased utilization beyond the CMS target even when adjusted for length of stay, patient comorbidities, and discharge. Needed medical care should continue to be a priority for inpatients, and allowance for individual outliers should be considered in policy discussions. Copyright © 2017 Elsevier Inc. All rights reserved. 12. Grading the probabilities of credit default risk for Malaysian listed companies by using the KMV-Merton model Science.gov (United States) Anuwar, Muhammad Hafidz; Jaffar, Maheran Mohd 2017-08-01 This paper provides an overview for the assessment of credit risk specific to the banks. In finance, risk is a term to reflect the potential of financial loss. The risk of default on loan may increase when a company does not make a payment on that loan when the time comes. Hence, this framework analyses the KMV-Merton model to estimate the probabilities of default for Malaysian listed companies. In this way, banks can verify the ability of companies to meet their loan commitments in order to overcome bad investments and financial losses. This model has been applied to all Malaysian listed companies in Bursa Malaysia for estimating the credit default probabilities of companies and compare with the rating given by the rating agency, which is RAM Holdings Berhad to conform to reality. Then, the significance of this study is a credit risk grade is proposed by using the KMV-Merton model for the Malaysian listed companies. 13. Three-dimensional analytic probabilities of coupled vibrational-rotational-translational energy transfer for DSMC modeling of nonequilibrium flows International Nuclear Information System (INIS) 2014-01-01 A three-dimensional, nonperturbative, semiclassical analytic model of vibrational energy transfer in collisions between a rotating diatomic molecule and an atom, and between two rotating diatomic molecules (Forced Harmonic Oscillator–Free Rotation model) has been extended to incorporate rotational relaxation and coupling between vibrational, translational, and rotational energy transfer. The model is based on analysis of semiclassical trajectories of rotating molecules interacting by a repulsive exponential atom-to-atom potential. The model predictions are compared with the results of three-dimensional close-coupled semiclassical trajectory calculations using the same potential energy surface. The comparison demonstrates good agreement between analytic and numerical probabilities of rotational and vibrational energy transfer processes, over a wide range of total collision energies, rotational energies, and impact parameter. The model predicts probabilities of single-quantum and multi-quantum vibrational-rotational transitions and is applicable up to very high collision energies and quantum numbers. Closed-form analytic expressions for these transition probabilities lend themselves to straightforward incorporation into DSMC nonequilibrium flow codes 14. a Probability Model for Drought Prediction Using Fusion of Markov Chain and SAX Methods Science.gov (United States) 2017-09-01 Drought is one of the most powerful natural disasters which are affected on different aspects of the environment. Most of the time this phenomenon is immense in the arid and semi-arid area. Monitoring and prediction the severity of the drought can be useful in the management of the natural disaster caused by drought. Many indices were used in predicting droughts such as SPI, VCI, and TVX. In this paper, based on three data sets (rainfall, NDVI, and land surface temperature) which are acquired from MODIS satellite imagery, time series of SPI, VCI, and TVX in time limited between winters 2000 to summer 2015 for the east region of Isfahan province were created. Using these indices and fusion of symbolic aggregation approximation and hidden Markov chain drought was predicted for fall 2015. For this purpose, at first, each time series was transformed into the set of quality data based on the state of drought (5 group) by using SAX algorithm then the probability matrix for the future state was created by using Markov hidden chain. The fall drought severity was predicted by fusion the probability matrix and state of drought severity in summer 2015. The prediction based on the likelihood for each state of drought includes severe drought, middle drought, normal drought, severe wet and middle wet. The analysis and experimental result from proposed algorithm show that the product of this algorithm is acceptable and the proposed algorithm is appropriate and efficient for predicting drought using remote sensor data. 15. Quantum Probabilities as Behavioral Probabilities Directory of Open Access Journals (Sweden) Vyacheslav I. Yukalov 2017-03-01 Full Text Available We demonstrate that behavioral probabilities of human decision makers share many common features with quantum probabilities. This does not imply that humans are some quantum objects, but just shows that the mathematics of quantum theory is applicable to the description of human decision making. The applicability of quantum rules for describing decision making is connected with the nontrivial process of making decisions in the case of composite prospects under uncertainty. Such a process involves deliberations of a decision maker when making a choice. In addition to the evaluation of the utilities of considered prospects, real decision makers also appreciate their respective attractiveness. Therefore, human choice is not based solely on the utility of prospects, but includes the necessity of resolving the utility-attraction duality. In order to justify that human consciousness really functions similarly to the rules of quantum theory, we develop an approach defining human behavioral probabilities as the probabilities determined by quantum rules. We show that quantum behavioral probabilities of humans do not merely explain qualitatively how human decisions are made, but they predict quantitative values of the behavioral probabilities. Analyzing a large set of empirical data, we find good quantitative agreement between theoretical predictions and observed experimental data. 16. Probabilistic sensitivity analysis on Markov models with uncertain transition probabilities: an application in evaluating treatment decisions for type 2 diabetes. Science.gov (United States) Zhang, Yuanhui; Wu, Haipeng; Denton, Brian T; Wilson, James R; Lobo, Jennifer M 2017-10-27 Markov models are commonly used for decision-making studies in many application domains; however, there are no widely adopted methods for performing sensitivity analysis on such models with uncertain transition probability matrices (TPMs). This article describes two simulation-based approaches for conducting probabilistic sensitivity analysis on a given discrete-time, finite-horizon, finite-state Markov model using TPMs that are sampled over a specified uncertainty set according to a relevant probability distribution. The first approach assumes no prior knowledge of the probability distribution, and each row of a TPM is independently sampled from the uniform distribution on the row's uncertainty set. The second approach involves random sampling from the (truncated) multivariate normal distribution of the TPM's maximum likelihood estimators for its rows subject to the condition that each row has nonnegative elements and sums to one. The two sampling methods are easily implemented and have reasonable computation times. A case study illustrates the application of these methods to a medical decision-making problem involving the evaluation of treatment guidelines for glycemic control of patients with type 2 diabetes, where natural variation in a patient's glycated hemoglobin (HbA1c) is modeled as a Markov chain, and the associated TPMs are subject to uncertainty. 17. OL-DEC-MDP Model for Multiagent Online Scheduling with a Time-Dependent Probability of Success Directory of Open Access Journals (Sweden) Cheng Zhu 2014-01-01 Full Text Available Focusing on the on-line multiagent scheduling problem, this paper considers the time-dependent probability of success and processing duration and proposes an OL-DEC-MDP (opportunity loss-decentralized Markov Decision Processes model to include opportunity loss into scheduling decision to improve overall performance. The success probability of job processing as well as the process duration is dependent on the time at which the processing is started. The probability of completing the assigned job by an agent would be higher when the process is started earlier, but the opportunity loss could also be high due to the longer engaging duration. As a result, OL-DEC-MDP model introduces a reward function considering the opportunity loss, which is estimated based on the prediction of the upcoming jobs by a sampling method on the job arrival. Heuristic strategies are introduced in computing the best starting time for an incoming job by each agent, and an incoming job will always be scheduled to the agent with the highest reward among all agents with their best starting policies. The simulation experiments show that the OL-DEC-MDP model will improve the overall scheduling performance compared with models not considering opportunity loss in heavy-loading environment. 18. Toward a generalized probability theory: conditional probabilities International Nuclear Information System (INIS) Cassinelli, G. 1979-01-01 The main mathematical object of interest in the quantum logic approach to the foundations of quantum mechanics is the orthomodular lattice and a set of probability measures, or states, defined by the lattice. This mathematical structure is studied per se, independently from the intuitive or physical motivation of its definition, as a generalized probability theory. It is thought that the building-up of such a probability theory could eventually throw light on the mathematical structure of Hilbert-space quantum mechanics as a particular concrete model of the generalized theory. (Auth.) 19. A probability distribution model of tooth pits for evaluating time-varying mesh stiffness of pitting gears Science.gov (United States) Lei, Yaguo; Liu, Zongyao; Wang, Delong; Yang, Xiao; Liu, Huan; Lin, Jing 2018-06-01 Tooth damage often causes a reduction in gear mesh stiffness. Thus time-varying mesh stiffness (TVMS) can be treated as an indication of gear health conditions. This study is devoted to investigating the mesh stiffness variations of a pair of external spur gears with tooth pitting, and proposes a new model for describing tooth pitting based on probability distribution. In the model, considering the appearance and development process of tooth pitting, we model the pitting on the surface of spur gear teeth as a series of pits with a uniform distribution in the direction of tooth width and a normal distribution in the direction of tooth height, respectively. In addition, four pitting degrees, from no pitting to severe pitting, are modeled. Finally, influences of tooth pitting on TVMS are analyzed in details and the proposed model is validated by comparing with a finite element model. The comparison results show that the proposed model is effective for the TVMS evaluations of pitting gears. 20. A Study of Probability Models in Monitoring Environmental Pollution in Nigeria Directory of Open Access Journals (Sweden) P. E. Oguntunde 2014-01-01 Full Text Available In Lagos State, Nigeria, pollutant emissions were monitored across the state to detect any significant change which may cause harm to human health and the environment at large. In this research, three theoretical distributions, Weibull, lognormal, and gamma distributions, were examined on the carbon monoxide observations to determine the best fit. The characteristics of the pollutant observation were established and the probabilities of exceeding the Lagos State Environmental Protection Agency (LASEPA and the Federal Environmental Protection Agency (FEPA acceptable limits have been successfully predicted. Increase in the use of vehicles and increase in the establishment of industries have been found not to contribute significantly to the high level of carbon monoxide concentration in Lagos State for the period studied. 1. A Model to Determinate the Influence of Probability Density Functions (PDFs of Input Quantities in Measurements Directory of Open Access Journals (Sweden) Jesús Caja 2016-06-01 Full Text Available A method for analysing the effect of different hypotheses about the type of the input quantities distributions of a measurement model is presented here so that the developed algorithms can be simplified. As an example, a model of indirect measurements with optical coordinate measurement machine was employed to evaluate these different hypotheses. As a result of the different experiments, the assumption that the different variables of the model can be modelled as normal distributions is proved. 2. The Ins2Akita mouse as a model of early retinal complications in diabetes. Science.gov (United States) Barber, Alistair J; Antonetti, David A; Kern, Timothy S; Reiter, Chad E N; Soans, Rohit S; Krady, J Kyle; Levison, Steven W; Gardner, Thomas W; Bronson, Sarah K 2005-06-01 This study tested the Ins2(Akita) mouse as an animal model of retinal complications in diabetes. The Ins2(Akita) mutation results in a single amino acid substitution in the insulin 2 gene that causes misfolding of the insulin protein. The mutation arose and is maintained on the C57BL/6J background. Male mice heterozygous for this mutation have progressive loss of beta-cell function, decreased pancreatic beta-cell density, and significant hyperglycemia, as early as 4 weeks of age. Heterozygous Ins2(Akita) mice were bred to C57BL/6J mice, and male offspring were monitored for hyperglycemia, beginning at 4.5 weeks of age. After 4 to 36 weeks of hyperglycemia, the retinas were analyzed for vascular permeability, vascular lesions, leukostasis, morphologic changes of micro- and macroglia, apoptosis, retinal degeneration, and insulin receptor kinase activity. The mean blood glucose of Ins2(Akita) mice was significantly elevated, whereas the body weight at death was reduced compared with that of control animals. Compared with sibling control mice, the Ins2(Akita) mice had increased retinal vascular permeability after 12 weeks of hyperglycemia (P microglia, but no changes in expression of Muller cell glial fibrillary acidic protein. Increased apoptosis was identified by immunoreactivity for active caspase-3 after 4 weeks of hyperglycemia (P cell bodies in the retinal ganglion cell layer (P retinal complications of diabetes. 3. Internal Fixation of Complicated Acetabular Fractures Directed by Preoperative Surgery with 3D Printing Models. Science.gov (United States) Liu, Zhao-Jie; Jia, Jian; Zhang, Yin-Guang; Tian, Wei; Jin, Xin; Hu, Yong-Cheng 2017-05-01 The purpose of this article is to evaluate the efficacy and feasibility of preoperative surgery with 3D printing-assisted internal fixation of complicated acetabular fractures. A retrospective case review was performed for the above surgical procedure. A 23-year-old man was confirmed by radiological examination to have fractures of multiple ribs, with hemopneumothorax and communicated fractures of the left acetabulum. According to the Letounel and Judet classification, T-shaped fracture involving posterior wall was diagnosed. A 3D printing pelvic model was established using CT digital imaging and communications in medicine (DICOM) data preoperatively, with which surgical procedures were simulated in preoperative surgery to confirm the sequence of the reduction and fixation as well as the position and length of the implants. Open reduction with internal fixation (ORIF) of the acetabular fracture using modified ilioinguinal and Kocher-Langenbeck approaches was performed 25 days after injury. Plates that had been pre-bent in the preoperative surgery were positioned and screws were tightened in the directions determined in the preoperative planning following satisfactory reduction. The duration of the operation was 170 min and blood loss was 900 mL. Postoperative X-rays showed that anatomical reduction of the acetabulum was achieved and the hip joint was congruous. The position and length of the implants were not different when compared with those in preoperative surgery on 3D printing models. We believe that preoperative surgery using 3D printing models is beneficial for confirming the reduction and fixation sequence, determining the reduction quality, shortening the operative time, minimizing preoperative difficulties, and predicting the prognosis for complicated fractures of acetabulam. © 2017 Chinese Orthopaedic Association and John Wiley & Sons Australia, Ltd. 4. Uniform Estimate of the Finite-Time Ruin Probability for All Times in a Generalized Compound Renewal Risk Model Directory of Open Access Journals (Sweden) Qingwu Gao 2012-01-01 Full Text Available We discuss the uniformly asymptotic estimate of the finite-time ruin probability for all times in a generalized compound renewal risk model, where the interarrival times of successive accidents and all the claim sizes caused by an accident are two sequences of random variables following a wide dependence structure. This wide dependence structure allows random variables to be either negatively dependent or positively dependent. 5. Application of wildfire spread and behavior models to assess fire probability and severity in the Mediterranean region Science.gov (United States) Salis, Michele; Arca, Bachisio; Bacciu, Valentina; Spano, Donatella; Duce, Pierpaolo; Santoni, Paul; Ager, Alan; Finney, Mark 2010-05-01 Characterizing the spatial pattern of large fire occurrence and severity is an important feature of the fire management planning in the Mediterranean region. The spatial characterization of fire probabilities, fire behavior distributions and value changes are key components for quantitative risk assessment and for prioritizing fire suppression resources, fuel treatments and law enforcement. Because of the growing wildfire severity and frequency in recent years (e.g.: Portugal, 2003 and 2005; Italy and Greece, 2007 and 2009), there is an increasing demand for models and tools that can aid in wildfire prediction and prevention. Newer wildfire simulation systems offer promise in this regard, and allow for fine scale modeling of wildfire severity and probability. Several new applications has resulted from the development of a minimum travel time (MTT) fire spread algorithm (Finney, 2002), that models the fire growth searching for the minimum time for fire to travel among nodes in a 2D network. The MTT approach makes computationally feasible to simulate thousands of fires and generate burn probability and fire severity maps over large areas. The MTT algorithm is imbedded in a number of research and fire modeling applications. High performance computers are typically used for MTT simulations, although the algorithm is also implemented in the FlamMap program (www.fire.org). In this work, we described the application of the MTT algorithm to estimate spatial patterns of burn probability and to analyze wildfire severity in three fire prone areas of the Mediterranean Basin, specifically Sardinia (Italy), Sicily (Italy) and Corsica (France) islands. We assembled fuels and topographic data for the simulations in 500 x 500 m grids for the study areas. The simulations were run using 100,000 ignitions under weather conditions that replicated severe and moderate weather conditions (97th and 70th percentile, July and August weather, 1995-2007). We used both random ignition locations 6. Numerical modelling of local deposition patients, activity distributions and cellular hit probabilities of inhaled radon progenies in human airways International Nuclear Information System (INIS) Farkas, A.; Balashazy, I.; Szoeke, I. 2003-01-01 The general objective of our research is modelling the biophysical processes of the effects of inhaled radon progenies. This effort is related to the rejection or support of the linear no threshold (LNT) dose-effect hypothesis, which seems to be one of the most challenging tasks of current radiation protection. Our approximation and results may also serve as a useful tool for lung cancer models. In this study, deposition patterns, activity distributions and alpha-hit probabilities of inhaled radon progenies in the large airways of the human tracheobronchial tree are computed. The airflow fields and related particle deposition patterns strongly depend on the shape of airway geometry and breathing pattern. Computed deposition patterns of attached an unattached radon progenies are strongly inhomogeneous creating hot spots at the carinal regions and downstream of the inner sides of the daughter airways. The results suggest that in the vicinity of the carinal regions the multiple hit probabilities are quite high even at low average doses and increase exponentially in the low-dose range. Thus, even the so-called low doses may present high doses for large clusters of cells. The cell transformation probabilities are much higher in these regions and this phenomenon cannot be modeled with average burdens. (authors) 7. Improved process for calculating the probability of being hit by crashing aircraft by the Balfanz-model International Nuclear Information System (INIS) Hennings, W. 1988-01-01 For calculating the probability of being hit by crashing military aircraft on different buildings, a model was introduced, which has already been used in the conventional fields. In the context of converting the research reactor BER II, this model was also used in the nuclear field. The report introduces this model and shows the application to a vertical cylinder as an example. Compared to the previous model, an exact and also simpler solution of the model attempt for determining the shade surface for different shapes of buildings is derived. The problems of the distribution of crashes given by the previous model is treated via the vertical angle and an attempt to solve these problems is given. (orig./HP) [de 8. Monitoring Least Bitterns (Ixobrychis exilis) in Vermont: Detection probability and occupancy modeling Science.gov (United States) Cherukuri, Aswini; Strong, Allan; Donovan, Therese M. 2018-01-01 Ixobrychus exillis (Least Bittern) is listed as a species of high concern in the North American Waterbird Conservation Plan and is a US Fish and Wildlife Service migratory bird species of conservation concern in the Northeast. Little is known about the population of Least Bitterns in the Northeast because of their low population density, tendency to nest in dense wetland vegetation, and secretive behavior. Urban and agricultural development is expected to encroach on and degrade suitable wetland habitat; however, we cannot predict the effects on Least Bittern populations without more accurate information on their abundance and distribution. We conducted surveys of wetlands in Vermont to assess the efficacy of a monitoring protocol and to establish baseline Least Bittern abundance and distribution data at a sample of 29 wetland sites. Surveys yielded detections of 31 individuals at 15 of 29 sites across 3 biophysical regions and at 5 sites where occupancy had not been previously reported. Probability of occupancy was positively related to wetland size and number of patches, though the relationships were not strong enough to conclude if these were true determinants of occupancy. Call—response broadcast surveys yielded 30 detections, while passive surveys yielded 13. Call—response broadcasts (P = 0.897) increased the rate of detection by 55% compared to passive surveys (P = 0.577). Our results suggest that call—response broadcast surveys are an effective means of assessing Least Bittern occupancy and may reduce bias in long-term monitoring programs. 9. Modeling of Kidney Hemodynamics: Probability-Based Topology of an Arterial Network DEFF Research Database (Denmark) Postnov, Dmitry D; Marsh, Donald J; Postnov, Dmitry E 2016-01-01 CT) data we develop an algorithm for generating the renal arterial network. We then introduce a mathematical model describing blood flow dynamics and nephron to nephron interaction in the network. The model includes an implementation of electrical signal propagation along a vascular wall. Simulation... 10. A comparison of the probability distribution of observed substorm magnitude with that predicted by a minimal substorm model Directory of Open Access Journals (Sweden) S. K. Morley 2007-11-01 Full Text Available We compare the probability distributions of substorm magnetic bay magnitudes from observations and a minimal substorm model. The observed distribution was derived previously and independently using the IL index from the IMAGE magnetometer network. The model distribution is derived from a synthetic AL index time series created using real solar wind data and a minimal substorm model, which was previously shown to reproduce observed substorm waiting times. There are two free parameters in the model which scale the contributions to AL from the directly-driven DP2 electrojet and loading-unloading DP1 electrojet, respectively. In a limited region of the 2-D parameter space of the model, the probability distribution of modelled substorm bay magnitudes is not significantly different to the observed distribution. The ranges of the two parameters giving acceptable (95% confidence level agreement are consistent with expectations using results from other studies. The approximately linear relationship between the two free parameters over these ranges implies that the substorm magnitude simply scales linearly with the solar wind power input at the time of substorm onset. 11. Using probability modelling and genetic parentage assignment to test the role of local mate availability in mating system variation. Science.gov (United States) Blyton, Michaela D J; Banks, Sam C; Peakall, Rod; Lindenmayer, David B 2012-02-01 The formal testing of mating system theories with empirical data is important for evaluating the relative importance of different processes in shaping mating systems in wild populations. Here, we present a generally applicable probability modelling framework to test the role of local mate availability in determining a population's level of genetic monogamy. We provide a significance test for detecting departures in observed mating patterns from model expectations based on mate availability alone, allowing the presence and direction of behavioural effects to be inferred. The assessment of mate availability can be flexible and in this study it was based on population density, sex ratio and spatial arrangement. This approach provides a useful tool for (1) isolating the effect of mate availability in variable mating systems and (2) in combination with genetic parentage analyses, gaining insights into the nature of mating behaviours in elusive species. To illustrate this modelling approach, we have applied it to investigate the variable mating system of the mountain brushtail possum (Trichosurus cunninghami) and compared the model expectations with the outcomes of genetic parentage analysis over an 18-year study. The observed level of monogamy was higher than predicted under the model. Thus, behavioural traits, such as mate guarding or selective mate choice, may increase the population level of monogamy. We show that combining genetic parentage data with probability modelling can facilitate an improved understanding of the complex interactions between behavioural adaptations and demographic dynamics in driving mating system variation. © 2011 Blackwell Publishing Ltd. 12. Probability tales CERN Document Server Grinstead, Charles M; Snell, J Laurie 2011-01-01 This book explores four real-world topics through the lens of probability theory. It can be used to supplement a standard text in probability or statistics. Most elementary textbooks present the basic theory and then illustrate the ideas with some neatly packaged examples. Here the authors assume that the reader has seen, or is learning, the basic theory from another book and concentrate in some depth on the following topics: streaks, the stock market, lotteries, and fingerprints. This extended format allows the authors to present multiple approaches to problems and to pursue promising side discussions in ways that would not be possible in a book constrained to cover a fixed set of topics. To keep the main narrative accessible, the authors have placed the more technical mathematical details in appendices. The appendices can be understood by someone who has taken one or two semesters of calculus. 13. A GRASS GIS Semi-Stochastic Model for Evaluating the Probability of Landslides Impacting Road Networks in Collazzone, Central Italy Science.gov (United States) Taylor, Faith E.; Santangelo, Michele; Marchesini, Ivan; Malamud, Bruce D. 2013-04-01 14. Exit probability of the one-dimensional q-voter model: Analytical results and simulations for large networks Science.gov (United States) Timpanaro, André M.; Prado, Carmen P. C. 2014-05-01 We discuss the exit probability of the one-dimensional q-voter model and present tools to obtain estimates about this probability, both through simulations in large networks (around 107 sites) and analytically in the limit where the network is infinitely large. We argue that the result E(ρ )=ρq/ρq+(1-ρ)q, that was found in three previous works [F. Slanina, K. Sznajd-Weron, and P. Przybyła, Europhys. Lett. 82, 18006 (2008), 10.1209/0295-5075/82/18006; R. Lambiotte and S. Redner, Europhys. Lett. 82, 18007 (2008), 10.1209/0295-5075/82/18007, for the case q =2; and P. Przybyła, K. Sznajd-Weron, and M. Tabiszewski, Phys. Rev. E 84, 031117 (2011), 10.1103/PhysRevE.84.031117, for q >2] using small networks (around 103 sites), is a good approximation, but there are noticeable deviations that appear even for small systems and that do not disappear when the system size is increased (with the notable exception of the case q =2). We also show that, under some simple and intuitive hypotheses, the exit probability must obey the inequality ρq/ρq+(1-ρ)≤E(ρ)≤ρ/ρ +(1-ρ)q in the infinite size limit. We believe this settles in the negative the suggestion made [S. Galam and A. C. R. Martins, Europhys. Lett. 95, 48005 (2001), 10.1209/0295-5075/95/48005] that this result would be a finite size effect, with the exit probability actually being a step function. We also show how the result that the exit probability cannot be a step function can be reconciled with the Galam unified frame, which was also a source of controversy. 15. Analytical Model for the Probability Characteristics of a Crack Penetrating Capsules in Capsule-Based Self-Healing Cementitious Materials Directory of Open Access Journals (Sweden) Zhong LV 2017-08-01 Full Text Available Autonomous crack healing using pre-embedded capsules containing healing agent is becoming a promising approach to restore the strength of damaged structures. In addition to the material properties, the size and volume fraction of capsules influence crack healing in the matrix. Understanding the crack and capsule interaction is critical in the development and design of structures made of capsule-based self-healing materials. Continuing our previous study, in this contribution a more practical rupturing mode of capsules characterizing the rupturing manner of capsules fractured by cracks in cementitious materials is presented, i.e., penetrating mode. With the underlying assumption that a crack penetrating capsules undoubtedly leads to crack healing, geometrical probability theory is employed to develop the quantitative relationship between crack size and capsule size, capsule concentration in capsule-based self-healing virtual cementitious material. Moreover, an analytical expression of probability of a crack penetrating with randomly dispersed capsules is developed in two-dimensional material matrix setup. The influences of the induced rupturing modes of capsules embedded on the self-healing efficiency are analyzed. Much attention is paid to compare the penetrating probability and the hitting probability, in order to assist the designer to make a choice of the optimal rupturing modes of capsules embedded. The accuracy of results of the theoretical model is also compared with Monte-Carlo numerical analysis of crack interacting with capsules. It shows that the developed probability characteristics of a crack interaction with capsules for different rupturing modes is helpful to provide guidelines for designer working with capsule-based self-healing cementitious materials.DOI: http://dx.doi.org/10.5755/j01.ms.23.3.16888 16. Use of a field model to analyze probable fire environments encountered within the complex geometries of nuclear power plants International Nuclear Information System (INIS) Boccio, J.L.; Usher, J.L.; Singhal, A.K.; Tam, L.T. 1985-08-01 A fire in a nuclear power plant (NPP) can damage equipment needed to safely operate the plant and thereby either directly cause an accident or else reduce the plant's margin of safety. The development of a field-model fire code to analyze the probable fire environments encountered within NPP is discussed. A set of fire tests carried out under the aegis of the US Nuclear Regulatory Commission (NRC) is described. The results of these tests are then utilized to validate the field model 17. A dynamic model of liquid containers (tanks) with legs and probability analysis of response to simulated earthquake International Nuclear Information System (INIS) Fujita, Takafumi; Shimosaka, Haruo 1980-01-01 This paper is described on the results of analysis of the response of liquid containers (tanks) to earthquakes. Sine wave oscillation was applied experimentally to model tanks with legs. A model with one degree of freedom is good enough for the analysis. To investigate the reason of this fact, the response multiplication factor of tank displacement was analysed. The shapes of the model tanks were rectangular and cylindrical. Analyses were made by a potential theory. The experimental studies show that the characteristics of attenuation of oscillation was non-linear. The model analysis of this non-linear attenuation was also performed. Good agreement between the experimental and the analytical results was recognized. The probability analysis of the response to earthquake with simulated shock waves was performed, using the above mentioned model, and good agreement between the experiment and the analysis was obtained. (Kato, T.) 18. Building a Model Using Bayesian Network for Assessment of Posterior Probabilities of Falling From Height at Workplaces Directory of Open Access Journals (Sweden) 2014-12-01 Full Text Available Background: Falls from height are one of the main causes of fatal occupational injuries. The objective of this study was to present a model for estimating occurrence probability of falling from height. Methods: In order to make a list of factors affecting falls, we used four expert group's judgment, literature review and an available database. Then the validity and reliability of designed questionnaire were determined and Bayesian networks were built. The built network, nodes and curves were quantified. For network sensitivity analysis, four types of analysis carried out. Results: A Bayesian network for assessment of posterior probabilities of falling from height proposed. The presented Bayesian network model shows the interrelationships among 37 causes affecting the falling from height and can calculate its posterior probabilities. The most important factors affecting falling were Non-compliance with safety instructions for work at height (0.127, Lack of safety equipment for work at height (0.094 and Lack of safety instructions for work at height (0.071 respectively. Conclusion: The proposed Bayesian network used to determine how different causes could affect the falling from height at work. The findings of this study can be used to decide on the falling accident prevention programs. 19. Developing an Empirical Model for Estimating the Probability of Electrical Short Circuits from Tin Whiskers. Part 2 Science.gov (United States) Courey, Karim J.; Asfour, Shihab S.; Onar, Arzu; Bayliss, Jon A.; Ludwig, Larry L.; Wright, Maria C. 2009-01-01 To comply with lead-free legislation, many manufacturers have converted from tin-lead to pure tin finishes of electronic components. However, pure tin finishes have a greater propensity to grow tin whiskers than tin-lead finishes. Since tin whiskers present an electrical short circuit hazard in electronic components, simulations have been developed to quantify the risk of said short circuits occurring. Existing risk simulations make the assumption that when a free tin whisker has bridged two adjacent exposed electrical conductors, the result is an electrical short circuit. This conservative assumption is made because shorting is a random event that had an unknown probability associated with it. Note however that due to contact resistance electrical shorts may not occur at lower voltage levels. In our first article we developed an empirical probability model for tin whisker shorting. In this paper, we develop a more comprehensive empirical model using a refined experiment with a larger sample size, in which we studied the effect of varying voltage on the breakdown of the contact resistance which leads to a short circuit. From the resulting data we estimated the probability distribution of an electrical short, as a function of voltage. In addition, the unexpected polycrystalline structure seen in the focused ion beam (FIB) cross section in the first experiment was confirmed in this experiment using transmission electron microscopy (TEM). The FIB was also used to cross section two card guides to facilitate the measurement of the grain size of each card guide's tin plating to determine its finish. 20. Model-based prognostics for batteries which estimates useful life and uses a probability density function Science.gov (United States) Saha, Bhaskar (Inventor); Goebel, Kai F. (Inventor) 2012-01-01 This invention develops a mathematical model to describe battery behavior during individual discharge cycles as well as over its cycle life. The basis for the form of the model has been linked to the internal processes of the battery and validated using experimental data. Effects of temperature and load current have also been incorporated into the model. Subsequently, the model has been used in a Particle Filtering framework to make predictions of remaining useful life for individual discharge cycles as well as for cycle life. The prediction performance was found to be satisfactory as measured by performance metrics customized for prognostics for a sample case. The work presented here provides initial steps towards a comprehensive health management solution for energy storage devices. 1. From classical to quantum models: the regularising role of integrals, symmetry and probabilities OpenAIRE Gazeau, Jean-Pierre 2018-01-01 In physics, one is often misled in thinking that the mathematical model of a system is part of or is that system itself. Think of expressions commonly used in physics like "point" particle, motion "on the line", "smooth" observables, wave function, and even "going to infinity", without forgetting perplexing phrases like "classical world" versus "quantum world".... On the other hand, when a mathematical model becomes really inoperative with regard to correct predictions, one is forced to repla... 2. A study of quantum mechanical probabilities in the classical Hodgkin-Huxley model. Science.gov (United States) Moradi, N; Scholkmann, F; Salari, V 2015-03-01 The Hodgkin-Huxley (HH) model is a powerful model to explain different aspects of spike generation in excitable cells. However, the HH model was proposed in 1952 when the real structure of the ion channel was unknown. It is now common knowledge that in many ion-channel proteins the flow of ions through the pore is governed by a gate, comprising a so-called "selectivity filter" inside the ion channel, which can be controlled by electrical interactions. The selectivity filter (SF) is believed to be responsible for the selection and fast conduction of particular ions across the membrane of an excitable cell. Other (generally larger) parts of the molecule such as the pore-domain gate control the access of ions to the channel protein. In fact, two types of gates are considered here for ion channels: the "external gate", which is the voltage sensitive gate, and the "internal gate" which is the selectivity filter gate (SFG). Some quantum effects are expected in the SFG due to its small dimensions, which may play an important role in the operation of an ion channel. Here, we examine parameters in a generalized model of HH to see whether any parameter affects the spike generation. Our results indicate that the previously suggested semi-quantum-classical equation proposed by Bernroider and Summhammer (BS) agrees strongly with the HH equation under different conditions and may even provide a better explanation in some cases. We conclude that the BS model can refine the classical HH model substantially. 3. Nonparametric estimation of transition probabilities in the non-Markov illness-death model: A comparative study. Science.gov (United States) 2015-06-01 Multi-state models are often used for modeling complex event history data. In these models the estimation of the transition probabilities is of particular interest, since they allow for long-term predictions of the process. These quantities have been traditionally estimated by the Aalen-Johansen estimator, which is consistent if the process is Markov. Several non-Markov estimators have been proposed in the recent literature, and their superiority with respect to the Aalen-Johansen estimator has been proved in situations in which the Markov condition is strongly violated. However, the existing estimators have the drawback of requiring that the support of the censoring distribution contains the support of the lifetime distribution, which is not often the case. In this article, we propose two new methods for estimating the transition probabilities in the progressive illness-death model. Some asymptotic results are derived. The proposed estimators are consistent regardless the Markov condition and the referred assumption about the censoring support. We explore the finite sample behavior of the estimators through simulations. The main conclusion of this piece of research is that the proposed estimators are much more efficient than the existing non-Markov estimators in most cases. An application to a clinical trial on colon cancer is included. Extensions to progressive processes beyond the three-state illness-death model are discussed. © 2015, The International Biometric Society. 4. Modeling the Speed Choice Behaviors of Drivers on Mountainous Roads with Complicated Shapes Directory of Open Access Journals (Sweden) Yiming Shao 2015-02-01 Full Text Available Roadway geometric features and pavement conditions can significantly affect driver behavior, particularly with regard to vehicle speed. This paper presents the development of an algorithm for speed selection for use in automated passenger car travel (without driver input on mountainous roads with complicated shapes. The relationship between favorable driving speed and the geometric features of horizontal curves was established on the basis of driving experiments and spot speed observation data, and speed control models were established for driving on curves, curve approaches/departures, and tangents. The models developed can be used to calculate a driver's desired speed on any roadway with a defined geometry. The model considers the driver's behavior type and the vehicle's dynamic properties. This paper presents the results of simulation experiments on roads with small curve radii and narrow widths. The algorithms developed may be used for assisted and automated driving. Under automated driving conditions, speed control and speed change based on the algorithms developed make drivers feel natural as if they drive the car themselves. 5. Birth-death models and coalescent point processes: the shape and probability of reconstructed phylogenies. Science.gov (United States) 2013-12-01 6. ABOUT PROBABILITY OF RESEARCH OF THE NN Ser SPECTRUM BY MODEL ATMOSPHERES METHOD OpenAIRE Sakhibullin, N. A.; Shimansky, V. V. 2017-01-01 The spectrum of close binary system NN Ser is investigated by a models atmospheres method. It is show that the atmosphere near the centrum of a hot spot on surface of red dwarf has powerful chromospheres, arising from heating in Laiman continua. Four models of binary system with various of parameters are constructed and their theoretical spectra are obtained. Temperature of white dwarf Tef = 62000 K, radius of the red dwarf RT = 0.20139 and angle inclination of system i = 82“ are determined. ... 7. Choice probability generating functions DEFF Research Database (Denmark) Fosgerau, Mogens; McFadden, Daniel; Bierlaire, Michel 2013-01-01 This paper considers discrete choice, with choice probabilities coming from maximization of preferences from a random utility field perturbed by additive location shifters (ARUM). Any ARUM can be characterized by a choice-probability generating function (CPGF) whose gradient gives the choice...... probabilities, and every CPGF is consistent with an ARUM. We relate CPGF to multivariate extreme value distributions, and review and extend methods for constructing CPGF for applications. The choice probabilities of any ARUM may be approximated by a cross-nested logit model. The results for ARUM are extended... 8. Business model risk analysis: predicting the probability of business network profitability NARCIS (Netherlands) Johnson, Pontus; Iacob, Maria Eugenia; Valja, Margus; van Sinderen, Marten J.; Magnusson, Christer; Ladhe, Tobias; van Sinderen, Marten J.; Oude Luttighuis, P.H.W.M.; Folmer, Erwin Johan Albert; Bosems, S. In the design phase of business collaboration, it is desirable to be able to predict the profitability of the business-to-be. Therefore, techniques to assess qualities such as costs, revenues, risks, and profitability have been previously proposed. However, they do not allow the modeler to properly 9. Probability density function shape sensitivity in the statistical modeling of turbulent particle dispersion Science.gov (United States) Litchford, Ron J.; Jeng, San-Mou 1992-01-01 The performance of a recently introduced statistical transport model for turbulent particle dispersion is studied here for rigid particles injected into a round turbulent jet. Both uniform and isosceles triangle pdfs are used. The statistical sensitivity to parcel pdf shape is demonstrated. 10. Computational Modeling of Statistical Learning: Effects of Transitional Probability versus Frequency and Links to Word Learning Science.gov (United States) Mirman, Daniel; Estes, Katharine Graf; Magnuson, James S. 2010-01-01 Statistical learning mechanisms play an important role in theories of language acquisition and processing. Recurrent neural network models have provided important insights into how these mechanisms might operate. We examined whether such networks capture two key findings in human statistical learning. In Simulation 1, a simple recurrent network… 11. Computational modeling of the probability of destructions in total joint endoprosthesis ceramic heads using Weibull's theory Czech Academy of Sciences Publication Activity Database Janíček, P.; Fuis, Vladimír; Málek, M. 2010-01-01 Roč. 14, č. 4 (2010), s. 42-51 ISSN 1335-2393 Institutional research plan: CEZ:AV0Z20760514 Keywords : computational modeling * ceramic head * in vivo destructions * hip joint endoprosthesis * probabily of rupture Subject RIV: BO - Biophysics 12. Probabilities and Predictions: Modeling the Development of Scientific Problem-Solving Skills Science.gov (United States) 2005-01-01 The IMMEX (Interactive Multi-Media Exercises) Web-based problem set platform enables the online delivery of complex, multimedia simulations, the rapid collection of student performance data, and has already been used in several genetic simulations. The next step is the use of these data to understand and improve student learning in a formative manner. This article describes the development of probabilistic models of undergraduate student problem solving in molecular genetics that detailed the spectrum of strategies students used when problem solving, and how the strategic approaches evolved with experience. The actions of 776 university sophomore biology majors from three molecular biology lecture courses were recorded and analyzed. Each of six simulations were first grouped by artificial neural network clustering to provide individual performance measures, and then sequences of these performances were probabilistically modeled by hidden Markov modeling to provide measures of progress. The models showed that students with different initial problem-solving abilities choose different strategies. Initial and final strategies varied across different sections of the same course and were not strongly correlated with other achievement measures. In contrast to previous studies, we observed no significant gender differences. We suggest that instructor interventions based on early student performances with these simulations may assist students to recognize effective and efficient problem-solving strategies and enhance learning. PMID:15746978 13. On the probability distribution of stock returns in the Mike-Farmer model Science.gov (United States) Gu, G.-F.; Zhou, W.-X. 2009-02-01 Recently, Mike and Farmer have constructed a very powerful and realistic behavioral model to mimick the dynamic process of stock price formation based on the empirical regularities of order placement and cancelation in a purely order-driven market, which can successfully reproduce the whole distribution of returns, not only the well-known power-law tails, together with several other important stylized facts. There are three key ingredients in the Mike-Farmer (MF) model: the long memory of order signs characterized by the Hurst index Hs, the distribution of relative order prices x in reference to the same best price described by a Student distribution (or Tsallis’ q-Gaussian), and the dynamics of order cancelation. They showed that different values of the Hurst index Hs and the freedom degree αx of the Student distribution can always produce power-law tails in the return distribution fr(r) with different tail exponent αr. In this paper, we study the origin of the power-law tails of the return distribution fr(r) in the MF model, based on extensive simulations with different combinations of the left part L(x) for x 0 of fx(x). We find that power-law tails appear only when L(x) has a power-law tail, no matter R(x) has a power-law tail or not. In addition, we find that the distributions of returns in the MF model at different timescales can be well modeled by the Student distributions, whose tail exponents are close to the well-known cubic law and increase with the timescale. 14. Collapse susceptibility mapping in karstified gypsum terrain (Sivas basin - Turkey) by conditional probability, logistic regression, artificial neural network models Science.gov (United States) Yilmaz, Isik; Keskin, Inan; Marschalko, Marian; Bednarik, Martin 2010-05-01 This study compares the GIS based collapse susceptibility mapping methods such as; conditional probability (CP), logistic regression (LR) and artificial neural networks (ANN) applied in gypsum rock masses in Sivas basin (Turkey). Digital Elevation Model (DEM) was first constructed using GIS software. Collapse-related factors, directly or indirectly related to the causes of collapse occurrence, such as distance from faults, slope angle and aspect, topographical elevation, distance from drainage, topographic wetness index- TWI, stream power index- SPI, Normalized Difference Vegetation Index (NDVI) by means of vegetation cover, distance from roads and settlements were used in the collapse susceptibility analyses. In the last stage of the analyses, collapse susceptibility maps were produced from CP, LR and ANN models, and they were then compared by means of their validations. Area Under Curve (AUC) values obtained from all three methodologies showed that the map obtained from ANN model looks like more accurate than the other models, and the results also showed that the artificial neural networks is a usefull tool in preparation of collapse susceptibility map and highly compatible with GIS operating features. Key words: Collapse; doline; susceptibility map; gypsum; GIS; conditional probability; logistic regression; artificial neural networks. 15. Extinction probabilities and stationary distributions of mobile genetic elements in prokaryotes: The birth-death-diversification model. Science.gov (United States) Drakos, Nicole E; Wahl, Lindi M 2015-12-01 Theoretical approaches are essential to our understanding of the complex dynamics of mobile genetic elements (MGEs) within genomes. Recently, the birth-death-diversification model was developed to describe the dynamics of mobile promoters (MPs), a particular class of MGEs in prokaryotes. A unique feature of this model is that genetic diversification of elements was included. To explore the implications of diversification on the longterm fate of MGE lineages, in this contribution we analyze the extinction probabilities, extinction times and equilibrium solutions of the birth-death-diversification model. We find that diversification increases both the survival and growth rate of MGE families, but the strength of this effect depends on the rate of horizontal gene transfer (HGT). We also find that the distribution of MGE families per genome is not necessarily monotonically decreasing, as observed for MPs, but may have a peak in the distribution that is related to the HGT rate. For MPs specifically, we find that new families have a high extinction probability, and predict that the number of MPs is increasing, albeit at a very slow rate. Additionally, we develop an extension of the birth-death-diversification model which allows MGEs in different regions of the genome, for example coding and non-coding, to be described by different rates. This extension may offer a potential explanation as to why the majority of MPs are located in non-promoter regions of the genome. Copyright © 2015 Elsevier Inc. All rights reserved. 16. A novel multi-model probability battery state of charge estimation approach for electric vehicles using H-infinity algorithm International Nuclear Information System (INIS) Lin, Cheng; Mu, Hao; Xiong, Rui; Shen, Weixiang 2016-01-01 Highlights: • A novel multi-model probability battery SOC fusion estimation approach was proposed. • The linear matrix inequality-based H∞ technique is employed to estimate the SOC. • The Bayes theorem has been employed to realize the optimal weight for the fusion. • The robustness of the proposed approach is verified by different batteries. • The results show that the proposed method can promote global estimation accuracy. - Abstract: Due to the strong nonlinearity and complex time-variant property of batteries, the existing state of charge (SOC) estimation approaches based on a single equivalent circuit model (ECM) cannot provide the accurate SOC for the entire discharging period. This paper aims to present a novel SOC estimation approach based on a multiple ECMs fusion method for improving the practical application performance. In the proposed approach, three battery ECMs, namely the Thevenin model, the double polarization model and the 3rd order RC model, are selected to describe the dynamic voltage of lithium-ion batteries and the genetic algorithm is then used to determine the model parameters. The linear matrix inequality-based H-infinity technique is employed to estimate the SOC from the three models and the Bayes theorem-based probability method is employed to determine the optimal weights for synthesizing the SOCs estimated from the three models. Two types of lithium-ion batteries are used to verify the feasibility and robustness of the proposed approach. The results indicate that the proposed approach can improve the accuracy and reliability of the SOC estimation against uncertain battery materials and inaccurate initial states. 17. Implementation of PSA models to estimate the probabilities associated with external event combination International Nuclear Information System (INIS) Burgazzi, Luciano 2014-01-01 This note endeavors to address some significant issues revealed by the Fukushima accident in Japan in 2011, such as the analysis of various dependency aspects arisen in the light of the external event PSA framework, as the treatment of the correlated hazards. To this aim some foundational notions to implement the PSA models related to specific aspects, like the external hazard combination, e.g., earthquake and tsunami as at the Fukushima accident, and the external hazard-caused internal events, e.g., seismic induced fire, are proposed and discussed to be incorporated within the risk assessment structure. Risk assessment of external hazards is required and utilized as an integrated part of PRA for operating and new reactor units. In the light of the Fukushima accident, of special interest are correlated events, whose modelling is proposed in the present study, in the form of some theoretical concepts, which lay the foundations for the PSA framework implementation. An applicative example is presented for illustrative purposes, since the analysis is carried out on the basis of generic numerical values assigned to an oversimplified model and results are achieved without any baseline comparison. Obviously the first step aimed at the process endorsement is the analysis of all available information in order to determine the level of applicability of the observed specific plant site events to the envisaged model and the statistical correlation analysis for event occurrence data that can be used as part of this process. Despite these drawbacks that actually do not qualify the achieved results, the present work represents an exploratory study aimed at resolving current open issues to be resolved in the PSA, like topics related to unanticipated scenarios: the combined external hazards of the earthquake and tsunami in Fukushima, external hazards causing internal events, such as seismic induced fire. These topics are to be resolved among the other ones as emerging from the 18. Inferring the most probable maps of underground utilities using Bayesian mapping model Science.gov (United States) Bilal, Muhammad; Khan, Wasiq; Muggleton, Jennifer; Rustighi, Emiliano; Jenks, Hugo; Pennock, Steve R.; Atkins, Phil R.; Cohn, Anthony 2018-03-01 Mapping the Underworld (MTU), a major initiative in the UK, is focused on addressing social, environmental and economic consequences raised from the inability to locate buried underground utilities (such as pipes and cables) by developing a multi-sensor mobile device. The aim of MTU device is to locate different types of buried assets in real time with the use of automated data processing techniques and statutory records. The statutory records, even though typically being inaccurate and incomplete, provide useful prior information on what is buried under the ground and where. However, the integration of information from multiple sensors (raw data) with these qualitative maps and their visualization is challenging and requires the implementation of robust machine learning/data fusion approaches. An approach for automated creation of revised maps was developed as a Bayesian Mapping model in this paper by integrating the knowledge extracted from sensors raw data and available statutory records. The combination of statutory records with the hypotheses from sensors was for initial estimation of what might be found underground and roughly where. The maps were (re)constructed using automated image segmentation techniques for hypotheses extraction and Bayesian classification techniques for segment-manhole connections. The model consisting of image segmentation algorithm and various Bayesian classification techniques (segment recognition and expectation maximization (EM) algorithm) provided robust performance on various simulated as well as real sites in terms of predicting linear/non-linear segments and constructing refined 2D/3D maps. 19. Betting on change: modeling transitional probabilities to guide therapy development for opioid dependence. Science.gov (United States) Carpenter, Kenneth M; Jiang, Huiping; Sullivan, Maria A; Bisaga, Adam; Comer, Sandra D; Raby, Wilfrid Noel; Brooks, Adam C; Nunes, Edward V 2009-03-01 This study investigated the process of change by modeling transitions among four clinical states encountered in 64 detoxified opiate-dependent individuals treated with daily oral naltrexone: no opiate use, blocked opiate use (i.e., opiate use while adhering to oral naltrexone), unblocked opiate use (i.e., opiate use after having discontinued oral naltrexone), and treatment dropout. The effects of baseline characteristics and two psychosocial interventions of differing intensity, behavioral naltrexone therapy (BNT) and compliance enhancement (CE), on these transitions were studied. Participants using greater quantities of opiates were more likely than other participants to be retained in BNT relative to CE. Markov modeling indicated a transition from abstinence to treatment dropout was approximately 3.56 times greater among participants in CE relative to participants in BNT, indicating the more comprehensive psychosocial intervention kept participants engaged in treatment longer. Transitions to stopping treatment were more likely to occur after unblocked opiate use in both treatments. Continued opiate use while being blocked accounted for a relatively low proportion of transitions to abstinence and may have more deleterious effects later in a treatment episode. (PsycINFO Database Record (c) 2009 APA, all rights reserved). 20. A probabilistic framework for microarray data analysis: fundamental probability models and statistical inference. Science.gov (United States) Ogunnaike, Babatunde A; Gelmi, Claudio A; Edwards, Jeremy S 2010-05-21 Gene expression studies generate large quantities of data with the defining characteristic that the number of genes (whose expression profiles are to be determined) exceed the number of available replicates by several orders of magnitude. Standard spot-by-spot analysis still seeks to extract useful information for each gene on the basis of the number of available replicates, and thus plays to the weakness of microarrays. On the other hand, because of the data volume, treating the entire data set as an ensemble, and developing theoretical distributions for these ensembles provides a framework that plays instead to the strength of microarrays. We present theoretical results that under reasonable assumptions, the distribution of microarray intensities follows the Gamma model, with the biological interpretations of the model parameters emerging naturally. We subsequently establish that for each microarray data set, the fractional intensities can be represented as a mixture of Beta densities, and develop a procedure for using these results to draw statistical inference regarding differential gene expression. We illustrate the results with experimental data from gene expression studies on Deinococcus radiodurans following DNA damage using cDNA microarrays. Copyright (c) 2010 Elsevier Ltd. All rights reserved. 1. Model-Based Calculations of the Probability of a Country's Nuclear Proliferation Decisions International Nuclear Information System (INIS) Li, Jun; Yim, Man-Sung; McNelis, David N. 2007-01-01 explain the occurrences of proliferation decisions. However, predicting major historical proliferation events using model-based predictions has been unreliable. Nuclear proliferation decisions by a country is affected by three main factors: (1) technology; (2) finance; and (3) political motivation [1]. Technological capability is important as nuclear weapons development needs special materials, detonation mechanism, delivery capability, and the supporting human resources and knowledge base. Financial capability is likewise important as the development of the technological capabilities requires a serious financial commitment. It would be difficult for any state with a gross national product (GNP) significantly less than that of about 100 billion to devote enough annual governmental funding to a nuclear weapon program to actually achieve positive results within a reasonable time frame (i.e., 10 years). At the same time, nuclear proliferation is not a matter determined by a mastery of technical details or overcoming financial constraints. Technology or finance is a necessary condition but not a sufficient condition for nuclear proliferation. At the most fundamental level, the proliferation decision by a state is controlled by its political motivation. To effectively address the issue of predicting proliferation events, all three of the factors must be included in the model. To the knowledge of the authors, none of the exiting models considered the 'technology' variable as part of the modeling. This paper presents an attempt to develop a methodology for statistical modeling and predicting a country's nuclear proliferation decisions. The approach is based on the combined use of data on a country's nuclear technical capability profiles economic development status, security environment factors and internal political and cultural factors. All of the information utilized in the study was from open source literature. (authors) 2. Exact valence bond entanglement entropy and probability distribution in the XXX spin chain and the potts model. Science.gov (United States) Jacobsen, J L; Saleur, H 2008-02-29 We determine exactly the probability distribution of the number N_(c) of valence bonds connecting a subsystem of length L>1 to the rest of the system in the ground state of the XXX antiferromagnetic spin chain. This provides, in particular, the asymptotic behavior of the valence-bond entanglement entropy S_(VB)=N_(c)ln2=4ln2/pi(2)lnL disproving a recent conjecture that this should be related with the von Neumann entropy, and thus equal to 1/3lnL. Our results generalize to the Q-state Potts model. 3. Relationship between the generalized equivalent uniform dose formulation and the Poisson statistics-based tumor control probability model International Nuclear Information System (INIS) Zhou Sumin; Das, Shiva; Wang Zhiheng; Marks, Lawrence B. 2004-01-01 The generalized equivalent uniform dose (GEUD) model uses a power-law formalism, where the outcome is related to the dose via a power law. We herein investigate the mathematical compatibility between this GEUD model and the Poisson statistics based tumor control probability (TCP) model. The GEUD and TCP formulations are combined and subjected to a compatibility constraint equation. This compatibility constraint equates tumor control probability from the original heterogeneous target dose distribution to that from the homogeneous dose from the GEUD formalism. It is shown that this constraint equation possesses a unique, analytical closed-form solution which relates radiation dose to the tumor cell survival fraction. It is further demonstrated that, when there is no positive threshold or finite critical dose in the tumor response to radiation, this relationship is not bounded within the realistic cell survival limits of 0%-100%. Thus, the GEUD and TCP formalisms are, in general, mathematically inconsistent. However, when a threshold dose or finite critical dose exists in the tumor response to radiation, there is a unique mathematical solution for the tumor cell survival fraction that allows the GEUD and TCP formalisms to coexist, provided that all portions of the tumor are confined within certain specific dose ranges 4. Survival under uncertainty an introduction to probability models of social structure and evolution CERN Document Server Volchenkov, Dimitri 2016-01-01 This book introduces and studies a number of stochastic models of subsistence, communication, social evolution and political transition that will allow the reader to grasp the role of uncertainty as a fundamental property of our irreversible world. At the same time, it aims to bring about a more interdisciplinary and quantitative approach across very diverse fields of research in the humanities and social sciences. Through the examples treated in this work – including anthropology, demography, migration, geopolitics, management, and bioecology, among other things – evidence is gathered to show that volatile environments may change the rules of the evolutionary selection and dynamics of any social system, creating a situation of adaptive uncertainty, in particular, whenever the rate of change of the environment exceeds the rate of adaptation. Last but not least, it is hoped that this book will contribute to the understanding that inherent randomness can also be a great opportunity – for social systems an... 5. Binary logistic regression modelling: Measuring the probability of relapse cases among drug addict Science.gov (United States) Ismail, Mohd Tahir; Alias, Siti Nor Shadila 2014-07-01 For many years Malaysia faced the drug addiction issues. The most serious case is relapse phenomenon among treated drug addict (drug addict who have under gone the rehabilitation programme at Narcotic Addiction Rehabilitation Centre, PUSPEN). Thus, the main objective of this study is to find the most significant factor that contributes to relapse to happen. The binary logistic regression analysis was employed to model the relationship between independent variables (predictors) and dependent variable. The dependent variable is the status of the drug addict either relapse, (Yes coded as 1) or not, (No coded as 0). Meanwhile the predictors involved are age, age at first taking drug, family history, education level, family crisis, community support and self motivation. The total of the sample is 200 which the data are provided by AADK (National Antidrug Agency). The finding of the study revealed that age and self motivation are statistically significant towards the relapse cases.. 6. Estimating global arthropod species richness: refining probabilistic models using probability bounds analysis Czech Academy of Sciences Publication Activity Database Hamilton, A. J.; Novotný, Vojtěch; Waters, E. K.; Basset, Y.; Benke, K. K.; Grimbacher, P. S.; Miller, S. E.; Samuelson, G. A.; Weiblen, G. D.; Yen, J. D. L.; Stork, N. E. 2013-01-01 Roč. 171, č. 2 (2013), s. 357-365 ISSN 0029-8549 R&D Projects: GA MŠk(CZ) LH11008; GA ČR GA206/09/0115 Grant - others:Czech Ministry of Education(CZ) CZ.1.07/2.3.00/20.0064; National Science Foundarion(US) DEB-0841885; Otto Kinne Foundation, Darwin Initiative(GB) 19-008 Institutional research plan: CEZ:AV0Z50070508 Institutional support: RVO:60077344 Keywords : host specificity * model * Monte Carlo Subject RIV: EH - Ecology, Behaviour Impact factor: 3.248, year: 2013 http://link.springer.com/article/10.1007%2Fs00442-012-2434-5 7. Concepts of probability theory CERN Document Server Pfeiffer, Paul E 1979-01-01 Using the Kolmogorov model, this intermediate-level text discusses random variables, probability distributions, mathematical expectation, random processes, more. For advanced undergraduates students of science, engineering, or math. Includes problems with answers and six appendixes. 1965 edition. 8. Development of a statistical model for the determination of the probability of riverbank erosion in a Meditteranean river basin Science.gov (United States) Varouchakis, Emmanouil; Kourgialas, Nektarios; Karatzas, George; Giannakis, Georgios; Lilli, Maria; Nikolaidis, Nikolaos 2014-05-01 Riverbank erosion affects the river morphology and the local habitat and results in riparian land loss, damage to property and infrastructures, ultimately weakening flood defences. An important issue concerning riverbank erosion is the identification of the areas vulnerable to erosion, as it allows for predicting changes and assists with stream management and restoration. One way to predict the vulnerable to erosion areas is to determine the erosion probability by identifying the underlying relations between riverbank erosion and the geomorphological and/or hydrological variables that prevent or stimulate erosion. A statistical model for evaluating the probability of erosion based on a series of independent local variables and by using logistic regression is developed in this work. The main variables affecting erosion are vegetation index (stability), the presence or absence of meanders, bank material (classification), stream power, bank height, river bank slope, riverbed slope, cross section width and water velocities (Luppi et al. 2009). In statistics, logistic regression is a type of regression analysis used for predicting the outcome of a categorical dependent variable, e.g. binary response, based on one or more predictor variables (continuous or categorical). The probabilities of the possible outcomes are modelled as a function of independent variables using a logistic function. Logistic regression measures the relationship between a categorical dependent variable and, usually, one or several continuous independent variables by converting the dependent variable to probability scores. Then, a logistic regression is formed, which predicts success or failure of a given binary variable (e.g. 1 = "presence of erosion" and 0 = "no erosion") for any value of the independent variables. The regression coefficients are estimated by using maximum likelihood estimation. The erosion occurrence probability can be calculated in conjunction with the model deviance regarding 9. Modeling visual search using three-parameter probability functions in a hierarchical Bayesian framework. Science.gov (United States) Lin, Yi-Shin; Heinke, Dietmar; Humphreys, Glyn W 2015-04-01 In this study, we applied Bayesian-based distributional analyses to examine the shapes of response time (RT) distributions in three visual search paradigms, which varied in task difficulty. In further analyses we investigated two common observations in visual search-the effects of display size and of variations in search efficiency across different task conditions-following a design that had been used in previous studies (Palmer, Horowitz, Torralba, & Wolfe, Journal of Experimental Psychology: Human Perception and Performance, 37, 58-71, 2011; Wolfe, Palmer, & Horowitz, Vision Research, 50, 1304-1311, 2010) in which parameters of the response distributions were measured. Our study showed that the distributional parameters in an experimental condition can be reliably estimated by moderate sample sizes when Monte Carlo simulation techniques are applied. More importantly, by analyzing trial RTs, we were able to extract paradigm-dependent shape changes in the RT distributions that could be accounted for by using the EZ2 diffusion model. The study showed that Bayesian-based RT distribution analyses can provide an important means to investigate the underlying cognitive processes in search, including stimulus grouping and the bottom-up guidance of attention. 10. Continuous time random walk model with asymptotical probability density of waiting times via inverse Mittag-Leffler function Science.gov (United States) Liang, Yingjie; Chen, Wen 2018-04-01 The mean squared displacement (MSD) of the traditional ultraslow diffusion is a logarithmic function of time. Recently, the continuous time random walk model is employed to characterize this ultraslow diffusion dynamics by connecting the heavy-tailed logarithmic function and its variation as the asymptotical waiting time density. In this study we investigate the limiting waiting time density of a general ultraslow diffusion model via the inverse Mittag-Leffler function, whose special case includes the traditional logarithmic ultraslow diffusion model. The MSD of the general ultraslow diffusion model is analytically derived as an inverse Mittag-Leffler function, and is observed to increase even more slowly than that of the logarithmic function model. The occurrence of very long waiting time in the case of the inverse Mittag-Leffler function has the largest probability compared with the power law model and the logarithmic function model. The Monte Carlo simulations of one dimensional sample path of a single particle are also performed. The results show that the inverse Mittag-Leffler waiting time density is effective in depicting the general ultraslow random motion. 11. Evaluation of Presumed Probability-Density-Function Models in Non-Premixed Flames by using Large Eddy Simulation International Nuclear Information System (INIS) Cao Hong-Jun; Zhang Hui-Qiang; Lin Wen-Yi 2012-01-01 Four kinds of presumed probability-density-function (PDF) models for non-premixed turbulent combustion are evaluated in flames with various stoichiometric mixture fractions by using large eddy simulation (LES). The LES code is validated by the experimental data of a classical turbulent jet flame (Sandia flame D). The mean and rms temperatures obtained by the presumed PDF models are compared with the LES results. The β-function model achieves a good prediction for different flames. The predicted rms temperature by using the double-δ function model is very small and unphysical in the vicinity of the maximum mean temperature. The clip-Gaussian model and the multi-δ function model make a worse prediction of the extremely fuel-rich or fuel-lean side due to the clip at the boundary of the mixture fraction space. The results also show that the overall prediction performance of presumed PDF models is better at mediate stoichiometric mixture fractions than that at very small or very large ones. (fundamental areas of phenomenology(including applications)) 12. Hydrological model calibration for flood prediction in current and future climates using probability distributions of observed peak flows and model based rainfall Science.gov (United States) Haberlandt, Uwe; Wallner, Markus; Radtke, Imke 2013-04-01 Derived flood frequency analysis based on continuous hydrological modelling is very demanding regarding the required length and temporal resolution of precipitation input data. Often such flood predictions are obtained using long precipitation time series from stochastic approaches or from regional climate models as input. However, the calibration of the hydrological model is usually done using short time series of observed data. This inconsistent employment of different data types for calibration and application of a hydrological model increases its uncertainty. Here, it is proposed to calibrate a hydrological model directly on probability distributions of observed peak flows using model based rainfall in line with its later application. Two examples are given to illustrate the idea. The first one deals with classical derived flood frequency analysis using input data from an hourly stochastic rainfall model. The second one concerns a climate impact analysis using hourly precipitation from a regional climate model. The results show that: (I) the same type of precipitation input data should be used for calibration and application of the hydrological model, (II) a model calibrated on extreme conditions works quite well for average conditions but not vice versa, (III) the calibration of the hydrological model using regional climate model data works as an implicit bias correction method and (IV) the best performance for flood estimation is usually obtained when model based precipitation and observed probability distribution of peak flows are used for model calibration. 13. Developing logistic regression models using purchase attributes and demographics to predict the probability of purchases of regular and specialty eggs. Science.gov (United States) Bejaei, M; Wiseman, K; Cheng, K M 2015-01-01 Consumers' interest in specialty eggs appears to be growing in Europe and North America. The objective of this research was to develop logistic regression models that utilise purchaser attributes and demographics to predict the probability of a consumer purchasing a specific type of table egg including regular (white and brown), non-caged (free-run, free-range and organic) or nutrient-enhanced eggs. These purchase prediction models, together with the purchasers' attributes, can be used to assess market opportunities of different egg types specifically in British Columbia (BC). An online survey was used to gather data for the models. A total of 702 completed questionnaires were submitted by BC residents. Selected independent variables included in the logistic regression to develop models for different egg types to predict the probability of a consumer purchasing a specific type of table egg. The variables used in the model accounted for 54% and 49% of variances in the purchase of regular and non-caged eggs, respectively. Research results indicate that consumers of different egg types exhibit a set of unique and statistically significant characteristics and/or demographics. For example, consumers of regular eggs were less educated, older, price sensitive, major chain store buyers, and store flyer users, and had lower awareness about different types of eggs and less concern regarding animal welfare issues. However, most of the non-caged egg consumers were less concerned about price, had higher awareness about different types of table eggs, purchased their eggs from local/organic grocery stores, farm gates or farmers markets, and they were more concerned about care and feeding of hens compared to consumers of other eggs types. 14. Computing elastic‐rebound‐motivated rarthquake probabilities in unsegmented fault models: a new methodology supported by physics‐based simulators Science.gov (United States) Field, Edward H. 2015-01-01 A methodology is presented for computing elastic‐rebound‐based probabilities in an unsegmented fault or fault system, which involves computing along‐fault averages of renewal‐model parameters. The approach is less biased and more self‐consistent than a logical extension of that applied most recently for multisegment ruptures in California. It also enables the application of magnitude‐dependent aperiodicity values, which the previous approach does not. Monte Carlo simulations are used to analyze long‐term system behavior, which is generally found to be consistent with that of physics‐based earthquake simulators. Results cast doubt that recurrence‐interval distributions at points on faults look anything like traditionally applied renewal models, a fact that should be considered when interpreting paleoseismic data. We avoid such assumptions by changing the "probability of what" question (from offset at a point to the occurrence of a rupture, assuming it is the next event to occur). The new methodology is simple, although not perfect in terms of recovering long‐term rates in Monte Carlo simulations. It represents a reasonable, improved way to represent first‐order elastic‐rebound predictability, assuming it is there in the first place, and for a system that clearly exhibits other unmodeled complexities, such as aftershock triggering. 15. Binomial probability distribution model-based protein identification algorithm for tandem mass spectrometry utilizing peak intensity information. Science.gov (United States) Xiao, Chuan-Le; Chen, Xiao-Zhou; Du, Yang-Li; Sun, Xuesong; Zhang, Gong; He, Qing-Yu 2013-01-04 Mass spectrometry has become one of the most important technologies in proteomic analysis. Tandem mass spectrometry (LC-MS/MS) is a major tool for the analysis of peptide mixtures from protein samples. The key step of MS data processing is the identification of peptides from experimental spectra by searching public sequence databases. Although a number of algorithms to identify peptides from MS/MS data have been already proposed, e.g. Sequest, OMSSA, X!Tandem, Mascot, etc., they are mainly based on statistical models considering only peak-matches between experimental and theoretical spectra, but not peak intensity information. Moreover, different algorithms gave different results from the same MS data, implying their probable incompleteness and questionable reproducibility. We developed a novel peptide identification algorithm, ProVerB, based on a binomial probability distribution model of protein tandem mass spectrometry combined with a new scoring function, making full use of peak intensity information and, thus, enhancing the ability of identification. Compared with Mascot, Sequest, and SQID, ProVerB identified significantly more peptides from LC-MS/MS data sets than the current algorithms at 1% False Discovery Rate (FDR) and provided more confident peptide identifications. ProVerB is also compatible with various platforms and experimental data sets, showing its robustness and versatility. The open-source program ProVerB is available at http://bioinformatics.jnu.edu.cn/software/proverb/ . 16. Survival modeling for the estimation of transition probabilities in model-based economic evaluations in the absence of individual patient data: a tutorial. Science.gov (United States) Diaby, Vakaramoko; Adunlin, Georges; Montero, Alberto J 2014-02-01 Survival modeling techniques are increasingly being used as part of decision modeling for health economic evaluations. As many models are available, it is imperative for interested readers to know about the steps in selecting and using the most suitable ones. The objective of this paper is to propose a tutorial for the application of appropriate survival modeling techniques to estimate transition probabilities, for use in model-based economic evaluations, in the absence of individual patient data (IPD). An illustration of the use of the tutorial is provided based on the final progression-free survival (PFS) analysis of the BOLERO-2 trial in metastatic breast cancer (mBC). An algorithm was adopted from Guyot and colleagues, and was then run in the statistical package R to reconstruct IPD, based on the final PFS analysis of the BOLERO-2 trial. It should be emphasized that the reconstructed IPD represent an approximation of the original data. Afterwards, we fitted parametric models to the reconstructed IPD in the statistical package Stata. Both statistical and graphical tests were conducted to verify the relative and absolute validity of the findings. Finally, the equations for transition probabilities were derived using the general equation for transition probabilities used in model-based economic evaluations, and the parameters were estimated from fitted distributions. The results of the application of the tutorial suggest that the log-logistic model best fits the reconstructed data from the latest published Kaplan-Meier (KM) curves of the BOLERO-2 trial. Results from the regression analyses were confirmed graphically. An equation for transition probabilities was obtained for each arm of the BOLERO-2 trial. In this paper, a tutorial was proposed and used to estimate the transition probabilities for model-based economic evaluation, based on the results of the final PFS analysis of the BOLERO-2 trial in mBC. The results of our study can serve as a basis for any model 17. Lamb wave-based damage quantification and probability of detection modeling for fatigue life assessment of riveted lap joint Science.gov (United States) He, Jingjing; Wang, Dengjiang; Zhang, Weifang 2015-03-01 This study presents an experimental and modeling study for damage detection and quantification in riveted lap joints. Embedded lead zirconate titanate piezoelectric (PZT) ceramic wafer-type sensors are employed to perform in-situ non-destructive testing during fatigue cyclical loading. A multi-feature integration method is developed to quantify the crack size using signal features of correlation coefficient, amplitude change, and phase change. In addition, probability of detection (POD) model is constructed to quantify the reliability of the developed sizing method. Using the developed crack size quantification method and the resulting POD curve, probabilistic fatigue life prediction can be performed to provide comprehensive information for decision-making. The effectiveness of the overall methodology is demonstrated and validated using several aircraft lap joint specimens from different manufactures and under different loading conditions. 18. On the shapes of the presumed probability density function for the modeling of turbulence-radiation interactions International Nuclear Information System (INIS) Liu, L.H.; Xu, X.; Chen, Y.L. 2004-01-01 The laminar flamelet equations in combination with the joint probability density function (PDF) transport equation of mixture fraction and turbulence frequency have been used to simulate turbulent jet diffusion flames. To check the suitability of the presumed shapes of the PDF for the modeling of turbulence-radiation interactions (TRI), two types of presumed joint PDFs are constructed by using the second-order moments of temperature and the species concentrations, which are derived by the laminar flamelet model. The time-averaged radiative source terms and the time-averaged absorption coefficients are calculated by the presumed joint PDF approaches, and compared with those obtained by the laminar flamelet model. By comparison, it is shown that there are obvious differences between the results of the independent PDF approach and the laminar flamelet model. Generally, the results of the dependent PDF approach agree better with those of the flamelet model. For the modeling of TRI, the dependent PDF approach is superior to the independent PDF approach 19. Simple Urban Simulation Atop Complicated Models: Multi-Scale Equation-Free Computing of Sprawl Using Geographic Automata Directory of Open Access Journals (Sweden) Yu Zou 2013-07-01 Full Text Available Reconciling competing desires to build urban models that can be simple and complicated is something of a grand challenge for urban simulation. It also prompts difficulties in many urban policy situations, such as urban sprawl, where simple, actionable ideas may need to be considered in the context of the messily complex and complicated urban processes and phenomena that work within cities. In this paper, we present a novel architecture for achieving both simple and complicated realizations of urban sprawl in simulation. Fine-scale simulations of sprawl geography are run using geographic automata to represent the geographical drivers of sprawl in intricate detail and over fine resolutions of space and time. We use Equation-Free computing to deploy population as a coarse observable of sprawl, which can be leveraged to run automata-based models as short-burst experiments within a meta-simulation framework. 20. SU-E-T-144: Bayesian Inference of Local Relapse Data Using a Poisson-Based Tumour Control Probability Model Energy Technology Data Exchange (ETDEWEB) La Russa, D [The Ottawa Hospital Cancer Centre, Ottawa, ON (Canada) 2015-06-15 Purpose: The purpose of this project is to develop a robust method of parameter estimation for a Poisson-based TCP model using Bayesian inference. Methods: Bayesian inference was performed using the PyMC3 probabilistic programming framework written in Python. A Poisson-based TCP regression model that accounts for clonogen proliferation was fit to observed rates of local relapse as a function of equivalent dose in 2 Gy fractions for a population of 623 stage-I non-small-cell lung cancer patients. The Slice Markov Chain Monte Carlo sampling algorithm was used to sample the posterior distributions, and was initiated using the maximum of the posterior distributions found by optimization. The calculation of TCP with each sample step required integration over the free parameter α, which was performed using an adaptive 24-point Gauss-Legendre quadrature. Convergence was verified via inspection of the trace plot and posterior distribution for each of the fit parameters, as well as with comparisons of the most probable parameter values with their respective maximum likelihood estimates. Results: Posterior distributions for α, the standard deviation of α (σ), the average tumour cell-doubling time (Td), and the repopulation delay time (Tk), were generated assuming α/β = 10 Gy, and a fixed clonogen density of 10{sup 7} cm−{sup 3}. Posterior predictive plots generated from samples from these posterior distributions are in excellent agreement with the observed rates of local relapse used in the Bayesian inference. The most probable values of the model parameters also agree well with maximum likelihood estimates. Conclusion: A robust method of performing Bayesian inference of TCP data using a complex TCP model has been established. 1. The HYDROMED model and its application to semi-arid Mediterranean catchments with hill reservoirs 3: Reservoir storage capacity and probability of failure model Directory of Open Access Journals (Sweden) R. Ragab 2001-01-01 Full Text Available This paper addresses the issue of "what reservoir storage capacity is required to maintain a yield with a given probability of failure?". It is an important issue in terms of construction and cost. HYDROMED offers a solution based on the modified Gould probability matrix method. This method has the advantage of sampling all years data without reference to the sequence and is therefore particularly suitable for catchments with patchy data. In the HYDROMED model, the probability of failure is calculated on a monthly basis. The model has been applied to the El-Gouazine catchment in Tunisia using a long rainfall record from Kairouan together with the estimated Hortonian runoff, class A pan evaporation data and estimated abstraction data. Generally, the probability of failure differed from winter to summer. Generally, the probability of failure approaches zero when the reservoir capacity is 500,000 m3. The 25% probability of failure (75% success is achieved with a reservoir capacity of 58,000 m3 in June and 95,000 m3 in January. The probability of failure for a 240,000 m3 capacity reservoir (closer to storage capacity of El-Gouazine 233,000 m3, is approximately 5% in November, December and January, 3% in March, and 1.1% in May and June. Consequently there is no high risk of El-Gouazine being unable to meet its requirements at a capacity of 233,000 m3. Subsequently the benefit, in terms of probability of failure, by increasing the reservoir volume of El-Gouazine to greater than the 250,000 m3 is not high. This is important for the design engineers and the funding organizations. However, the analysis is based on the existing water abstraction policy, absence of siltation rate data and on the assumption that the present climate will prevail during the lifetime of the reservoir. Should these conditions change, a new analysis should be carried out. Keywords: HYDROMED, reservoir, storage capacity, probability of failure, Mediterranean 2. WE-AB-204-02: Molecular-Imaging Based Assessment of Liver Complications for Yttrium-90 Microsphere Treatments: Can Existing NTCP Models Explain Clinical Outcomes? Energy Technology Data Exchange (ETDEWEB) Lin, M; Choi, E; Chuong, M; Saboury, B; Moeslein, F; D’Souza, W; Guerrero, M [University of Maryland School of Medicine, Baltimore, MD (United States) 2015-06-15 Purpose: To evaluate weather the current radiobiological models can predict the normal liver complications of radioactive Yttrium-90 ({sup 90}Y) selective-internal-radiation-treatment (SIRT) for metastatic liver lesions based on the post-infusion {sup 90}Y PET images. Methods: A total of 20 patients with metastatic liver tumors treated with SIRT that received a post-infusion {sup 90}Y-PET/CT scan were analyzed in this work. The 3D activity distribution of the PET images was converted into a 3D dose distribution via a kernel convolution process. The physical dose distribution was converted into the equivalent dose (EQ2) delivered at 2 Gy based on the linear-quadratic (LQ) model considering the dose rate effect. The biological endpoint of this work was radiation-induce liver disease (RILD). The NTCPs were calculated with four different repair-times (T1/2-Liver-Repair= 0,0.5,1.0,2.0 hr) and three published NTCP models (Lyman-external-RT, Lyman 90Y-HCC-SIRT, parallel model) were compared to the incidence of RILD of the recruited patients to evaluate their ability of outcome prediction. Results: The mean normal liver physical dose (avg. 51.9 Gy, range 31.9–69.8 Gy) is higher than the suggested liver dose constraint for external beam treatment (∼30 Gy). However, none of the patients in our study developed RILD after the SIRT. The estimated probability of ‘no patient developing RILD’ obtained from the two Lyman models are 46.3% to 48.3% (T1/2-Liver-Repair= 0hr) and <1% for all other repair times. For the parallel model, the estimated probability is 97.3% (0hr), 51.7% (0.5hr), 2.0% (1.0hr) and <1% (2.0hr). Conclusion: Molecular-images providing the distribution of {sup 90}Y enable the dose-volume based dose/outcome analysis for SIRT. Current NTCP models fail to predict RILD complications in our patient population, unless a very short repair-time for the liver is assumed. The discrepancy between the Lyman {sup 90}Y-HCC-SIRT model predicted and the clinically 3. Probabilities in physics CERN Document Server Hartmann, Stephan 2011-01-01 Many results of modern physics--those of quantum mechanics, for instance--come in a probabilistic guise. But what do probabilistic statements in physics mean? Are probabilities matters of objective fact and part of the furniture of the world, as objectivists think? Or do they only express ignorance or belief, as Bayesians suggest? And how are probabilistic hypotheses justified and supported by empirical evidence? Finally, what does the probabilistic nature of physics imply for our understanding of the world? This volume is the first to provide a philosophical appraisal of probabilities in all of physics. Its main aim is to make sense of probabilistic statements as they occur in the various physical theories and models and to provide a plausible epistemology and metaphysics of probabilities. The essays collected here consider statistical physics, probabilistic modelling, and quantum mechanics, and critically assess the merits and disadvantages of objectivist and subjectivist views of probabilities in these fie... 4. Tennis Elbow Diagnosis Using Equivalent Uniform Voltage to Fit the Logistic and the Probit Diseased Probability Models Directory of Open Access Journals (Sweden) Tsair-Fwu Lee 2015-01-01 Full Text Available To develop the logistic and the probit models to analyse electromyographic (EMG equivalent uniform voltage- (EUV- response for the tenderness of tennis elbow. In total, 78 hands from 39 subjects were enrolled. In this study, surface EMG (sEMG signal is obtained by an innovative device with electrodes over forearm region. The analytical endpoint was defined as Visual Analog Score (VAS 3+ tenderness of tennis elbow. The logistic and the probit diseased probability (DP models were established for the VAS score and EMG absolute voltage-time histograms (AVTH. TV50 is the threshold equivalent uniform voltage predicting a 50% risk of disease. Twenty-one out of 78 samples (27% developed VAS 3+ tenderness of tennis elbow reported by the subject and confirmed by the physician. The fitted DP parameters were TV50 = 153.0 mV (CI: 136.3–169.7 mV, γ50 = 0.84 (CI: 0.78–0.90 and TV50 = 155.6 mV (CI: 138.9–172.4 mV, m = 0.54 (CI: 0.49–0.59 for logistic and probit models, respectively. When the EUV ≥ 153 mV, the DP of the patient is greater than 50% and vice versa. The logistic and the probit models are valuable tools to predict the DP of VAS 3+ tenderness of tennis elbow. 5. Tennis Elbow Diagnosis Using Equivalent Uniform Voltage to Fit the Logistic and the Probit Diseased Probability Models Science.gov (United States) Lin, Wei-Chun; Lin, Shu-Yuan; Wu, Li-Fu; Guo, Shih-Sian; Huang, Hsiang-Jui; Chao, Pei-Ju 2015-01-01 To develop the logistic and the probit models to analyse electromyographic (EMG) equivalent uniform voltage- (EUV-) response for the tenderness of tennis elbow. In total, 78 hands from 39 subjects were enrolled. In this study, surface EMG (sEMG) signal is obtained by an innovative device with electrodes over forearm region. The analytical endpoint was defined as Visual Analog Score (VAS) 3+ tenderness of tennis elbow. The logistic and the probit diseased probability (DP) models were established for the VAS score and EMG absolute voltage-time histograms (AVTH). TV50 is the threshold equivalent uniform voltage predicting a 50% risk of disease. Twenty-one out of 78 samples (27%) developed VAS 3+ tenderness of tennis elbow reported by the subject and confirmed by the physician. The fitted DP parameters were TV50 = 153.0 mV (CI: 136.3–169.7 mV), γ 50 = 0.84 (CI: 0.78–0.90) and TV50 = 155.6 mV (CI: 138.9–172.4 mV), m = 0.54 (CI: 0.49–0.59) for logistic and probit models, respectively. When the EUV ≥ 153 mV, the DP of the patient is greater than 50% and vice versa. The logistic and the probit models are valuable tools to predict the DP of VAS 3+ tenderness of tennis elbow. PMID:26380281 6. Probability of Ship on Collision Courses Based on the New PAW Using MMG Model and AIS Data Directory of Open Access Journals (Sweden) I Putu Sindhu Asmara 2015-03-01 Full Text Available This paper proposes an estimation method for ships on collision courses taking crash astern maneuvers based on a new potential area of water (PAW for maneuvering. A crash astern maneuver is an emergency option a ship can take when exposed to the risk of a collision with other ships that have lost control. However, lateral forces and yaw moments exerted by the reversing propeller, as well as the uncertainty of the initial speed and initial yaw rate, will move the ship out of the intended stopping position landing it in a dangerous area. A new PAW for crash astern maneuvers is thus introduced. The PAW is developed based on a probability density function of the initial yaw rate. Distributions of the yaw rates and speeds are analyzed from automatic identification system (AIS data in Madura Strait, and estimated paths of the maneuvers are simulated using a mathematical maneuvering group model. 7. Ruin Probabilities in a Dependent Discrete-Time Risk Model With Gamma-Like Tailed Insurance Risks Directory of Open Access Journals (Sweden) Xing-Fang Huang 2017-03-01 Full Text Available This paper considered a dependent discrete-time risk model, in which the insurance risks are represented by a sequence of independent and identically distributed real-valued random variables with a common Gamma-like tailed distribution; the financial risks are denoted by another sequence of independent and identically distributed positive random variables with a finite upper endpoint, but a general dependence structure exists between each pair of the insurance risks and the financial risks. Following the works of Yang and Yuen in 2016, we derive some asymptotic relations for the finite-time and infinite-time ruin probabilities. As a complement, we demonstrate our obtained result through a Crude Monte Carlo (CMC simulation with asymptotics. 8. Influence of Coloured Correlated Noises on Probability Distribution and Mean of Tumour Cell Number in the Logistic Growth Model Institute of Scientific and Technical Information of China (English) HAN Li-Bo; GONG Xiao-Long; CAO Li; WU Da-Jin 2007-01-01 An approximate Fokker-P1anck equation for the logistic growth model which is driven by coloured correlated noises is derived by applying the Novikov theorem and the Fox approximation. The steady-state probability distribution (SPD) and the mean of the tumour cell number are analysed. It is found that the SPD is the single extremum configuration when the degree of correlation between the multiplicative and additive noises, λ, is in -1<λ ≤ 0 and can be the double extrema in 0<λ<1. A configuration transition occurs because of the variation of noise parameters. A minimum appears in the curve of the mean of the steady-state tumour cell number, 〈x〉, versus λ. The position and the value of the minimum are controlled by the noise-correlated times. 9. Estimating inverse probability weights using super learner when weight-model specification is unknown in a marginal structural Cox model context. Science.gov (United States) Karim, Mohammad Ehsanul; Platt, Robert W 2017-06-15 Correct specification of the inverse probability weighting (IPW) model is necessary for consistent inference from a marginal structural Cox model (MSCM). In practical applications, researchers are typically unaware of the true specification of the weight model. Nonetheless, IPWs are commonly estimated using parametric models, such as the main-effects logistic regression model. In practice, assumptions underlying such models may not hold and data-adaptive statistical learning methods may provide an alternative. Many candidate statistical learning approaches are available in the literature. However, the optimal approach for a given dataset is impossible to predict. Super learner (SL) has been proposed as a tool for selecting an optimal learner from a set of candidates using cross-validation. In this study, we evaluate the usefulness of a SL in estimating IPW in four different MSCM simulation scenarios, in which we varied the specification of the true weight model specification (linear and/or additive). Our simulations show that, in the presence of weight model misspecification, with a rich and diverse set of candidate algorithms, SL can generally offer a better alternative to the commonly used statistical learning approaches in terms of MSE as well as the coverage probabilities of the estimated effect in an MSCM. The findings from the simulation studies guided the application of the MSCM in a multiple sclerosis cohort from British Columbia, Canada (1995-2008), to estimate the impact of beta-interferon treatment in delaying disability progression. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd. 10. Optimum parameters in a model for tumour control probability, including interpatient heterogeneity: evaluation of the log-normal distribution International Nuclear Information System (INIS) Keall, P J; Webb, S 2007-01-01 The heterogeneity of human tumour radiation response is well known. Researchers have used the normal distribution to describe interpatient tumour radiosensitivity. However, many natural phenomena show a log-normal distribution. Log-normal distributions are common when mean values are low, variances are large and values cannot be negative. These conditions apply to radiosensitivity. The aim of this work was to evaluate the log-normal distribution to predict clinical tumour control probability (TCP) data and to compare the results with the homogeneous (δ-function with single α-value) and normal distributions. The clinically derived TCP data for four tumour types-melanoma, breast, squamous cell carcinoma and nodes-were used to fit the TCP models. Three forms of interpatient tumour radiosensitivity were considered: the log-normal, normal and δ-function. The free parameters in the models were the radiosensitivity mean, standard deviation and clonogenic cell density. The evaluation metric was the deviance of the maximum likelihood estimation of the fit of the TCP calculated using the predicted parameters to the clinical data. We conclude that (1) the log-normal and normal distributions of interpatient tumour radiosensitivity heterogeneity more closely describe clinical TCP data than a single radiosensitivity value and (2) the log-normal distribution has some theoretical and practical advantages over the normal distribution. Further work is needed to test these models on higher quality clinical outcome datasets 11. Choice probability generating functions DEFF Research Database (Denmark) Fosgerau, Mogens; McFadden, Daniel; Bierlaire, Michel 2010-01-01 This paper establishes that every random utility discrete choice model (RUM) has a representation that can be characterized by a choice-probability generating function (CPGF) with specific properties, and that every function with these specific properties is consistent with a RUM. The choice...... probabilities from the RUM are obtained from the gradient of the CPGF. Mixtures of RUM are characterized by logarithmic mixtures of their associated CPGF. The paper relates CPGF to multivariate extreme value distributions, and reviews and extends methods for constructing generating functions for applications....... The choice probabilities of any ARUM may be approximated by a cross-nested logit model. The results for ARUM are extended to competing risk survival models.... 12. Medical Updates Number 5 to the International Space Station Probability Risk Assessment (PRA) Model Using the Integrated Medical Model Science.gov (United States) Butler, Doug; Bauman, David; Johnson-Throop, Kathy 2011-01-01 The Integrated Medical Model (IMM) Project has been developing a probabilistic risk assessment tool, the IMM, to help evaluate in-flight crew health needs and impacts to the mission due to medical events. This package is a follow-up to a data package provided in June 2009. The IMM currently represents 83 medical conditions and associated ISS resources required to mitigate medical events. IMM end state forecasts relevant to the ISS PRA model include evacuation (EVAC) and loss of crew life (LOCL). The current version of the IMM provides the basis for the operational version of IMM expected in the January 2011 timeframe. The objectives of this data package are: 1. To provide a preliminary understanding of medical risk data used to update the ISS PRA Model. The IMM has had limited validation and an initial characterization of maturity has been completed using NASA STD 7009 Standard for Models and Simulation. The IMM has been internally validated by IMM personnel but has not been validated by an independent body external to the IMM Project. 2. To support a continued dialogue between the ISS PRA and IMM teams. To ensure accurate data interpretation, and that IMM output format and content meets the needs of the ISS Risk Management Office and ISS PRA Model, periodic discussions are anticipated between the risk teams. 3. To help assess the differences between the current ISS PRA and IMM medical risk forecasts of EVAC and LOCL. Follow-on activities are anticipated based on the differences between the current ISS PRA medical risk data and the latest medical risk data produced by IMM. 13. The perception of probability. Science.gov (United States) Gallistel, C R; Krishan, Monika; Liu, Ye; Miller, Reilly; Latham, Peter E 2014-01-01 We present a computational model to explain the results from experiments in which subjects estimate the hidden probability parameter of a stepwise nonstationary Bernoulli process outcome by outcome. The model captures the following results qualitatively and quantitatively, with only 2 free parameters: (a) Subjects do not update their estimate after each outcome; they step from one estimate to another at irregular intervals. (b) The joint distribution of step widths and heights cannot be explained on the assumption that a threshold amount of change must be exceeded in order for them to indicate a change in their perception. (c) The mapping of observed probability to the median perceived probability is the identity function over the full range of probabilities. (d) Precision (how close estimates are to the best possible estimate) is good and constant over the full range. (e) Subjects quickly detect substantial changes in the hidden probability parameter. (f) The perceived probability sometimes changes dramatically from one observation to the next. (g) Subjects sometimes have second thoughts about a previous change perception, after observing further outcomes. (h) The frequency with which they perceive changes moves in the direction of the true frequency over sessions. (Explaining this finding requires 2 additional parametric assumptions.) The model treats the perception of the current probability as a by-product of the construction of a compact encoding of the experienced sequence in terms of its change points. It illustrates the why and the how of intermittent Bayesian belief updating and retrospective revision in simple perception. It suggests a reinterpretation of findings in the recent literature on the neurobiology of decision making. (PsycINFO Database Record (c) 2014 APA, all rights reserved). 14. Numerical modelling as a cost-reduction tool for probability of detection of bolt hole eddy current testing Science.gov (United States) Mandache, C.; Khan, M.; Fahr, A.; Yanishevsky, M. 2011-03-01 Probability of detection (PoD) studies are broadly used to determine the reliability of specific nondestructive inspection procedures, as well as to provide data for damage tolerance life estimations and calculation of inspection intervals for critical components. They require inspections on a large set of samples, a fact that makes these statistical assessments time- and cost-consuming. Physics-based numerical simulations of nondestructive testing inspections could be used as a cost-effective alternative to empirical investigations. They realistically predict the inspection outputs as functions of the input characteristics related to the test piece, transducer and instrument settings, which are subsequently used to partially substitute and/or complement inspection data in PoD analysis. This work focuses on the numerical modelling aspects of eddy current testing for the bolt hole inspections of wing box structures typical of the Lockheed Martin C-130 Hercules and P-3 Orion aircraft, found in the air force inventory of many countries. Boundary element-based numerical modelling software was employed to predict the eddy current signal responses when varying inspection parameters related to probe characteristics, crack geometry and test piece properties. Two demonstrator exercises were used for eddy current signal prediction when lowering the driver probe frequency and changing the material's electrical conductivity, followed by subsequent discussions and examination of the implications on using simulated data in the PoD analysis. Despite some simplifying assumptions, the modelled eddy current signals were found to provide similar results to the actual inspections. It is concluded that physics-based numerical simulations have the potential to partially substitute or complement inspection data required for PoD studies, reducing the cost, time, effort and resources necessary for a full empirical PoD assessment. 15. A facilitated diffusion model constrained by the probability isotherm: a pedagogical exercise in intuitive non-equilibrium thermodynamics. Science.gov (United States) Chapman, Brian 2017-06-01 This paper seeks to develop a more thermodynamically sound pedagogy for students of biological transport than is currently available from either of the competing schools of linear non-equilibrium thermodynamics (LNET) or Michaelis-Menten kinetics (MMK). To this end, a minimal model of facilitated diffusion was constructed comprising four reversible steps: cis- substrate binding, cis → trans bound enzyme shuttling, trans -substrate dissociation and trans → cis free enzyme shuttling. All model parameters were subject to the second law constraint of the probability isotherm, which determined the unidirectional and net rates for each step and for the overall reaction through the law of mass action. Rapid equilibration scenarios require sensitive 'tuning' of the thermodynamic binding parameters to the equilibrium substrate concentration. All non-equilibrium scenarios show sigmoidal force-flux relations, with only a minority of cases having their quasi -linear portions close to equilibrium. Few cases fulfil the expectations of MMK relating reaction rates to enzyme saturation. This new approach illuminates and extends the concept of rate-limiting steps by focusing on the free energy dissipation associated with each reaction step and thereby deducing its respective relative chemical impedance. The crucial importance of an enzyme's being thermodynamically 'tuned' to its particular task, dependent on the cis- and trans- substrate concentrations with which it deals, is consistent with the occurrence of numerous isoforms for enzymes that transport a given substrate in physiologically different circumstances. This approach to kinetic modelling, being aligned with neither MMK nor LNET, is best described as intuitive non-equilibrium thermodynamics, and is recommended as a useful adjunct to the design and interpretation of experiments in biotransport. 16. Visualisation of details of a complicated inner structure of model objects by the method of diffusion optical tomography International Nuclear Information System (INIS) Tret'yakov, Evgeniy V; Shuvalov, Vladimir V; Shutov, I V 2002-01-01 An approximate algorithm is tested for solving the problem of diffusion optical tomography in experiments on the visualisation of details of the inner structure of strongly scattering model objects containing scattering and semitransparent inclusions, as well as absorbing inclusions located inside other optical inhomogeneities. The stability of the algorithm to errors is demonstrated, which allows its use for a rapid (2 - 3 min) image reconstruction of the details of objects with a complicated inner structure. (laser biology and medicine) 17. A 3-stage model for assessing the probable economic effects of direct-to-consumer advertising of pharmaceuticals. Science.gov (United States) Vogel, Ronald J; Ramachandran, Sulabha; Zachry, Woodie M 2003-01-01 The pharmaceutical industry employs a variety of marketing strategies that have previously been directed primarily toward physicians. However, mass media direct-to-consumer (DTC) advertising of prescription drugs has emerged as a ubiquitous promotional strategy. This article explores the economics of DTC advertising in greater depth than has been done in the past by using a 3-stage economic model to assess the pertinent literature and to show the probable effects of DTC advertising in the United States. Economics literature on the subject was searched using the Journal of Economic Literature. Health services literature was searched using computer callback devices. Spending on DTC advertising in the United States increased from17 million in 1985 to \$2.5 billion in 2000. Proponents of DTC advertising claim that it provides valuable product-related information to health care professionals and patients, may contribute to better use of medications, and helps patients take charge of their own health care. Opponents argue that DTC advertising provides misleading messages rather than well-balanced, evidence-based information. The literature is replete with opinions about the effects of prescription drug advertising on pharmaceutical drug prices and physician-prescribing patterns, but few studies have addressed the issues beyond opinion surveys. The economic literature on advertising effects in other markets, however, may provide insight. DTC advertising indirectly affects the price and the quantity of production of pharmaceuticals via its effect on changes in consumer demand. 18. Introduction to probability with R CERN Document Server Baclawski, Kenneth 2008-01-01 FOREWORD PREFACE Sets, Events, and Probability The Algebra of Sets The Bernoulli Sample Space The Algebra of Multisets The Concept of Probability Properties of Probability Measures Independent Events The Bernoulli Process The R Language Finite Processes The Basic Models Counting Rules Computing Factorials The Second Rule of Counting Computing Probabilities Discrete Random Variables The Bernoulli Process: Tossing a Coin The Bernoulli Process: Random Walk Independence and Joint Distributions Expectations The Inclusion-Exclusion Principle General Random Variable 19. Estimating Subjective Probabilities DEFF Research Database (Denmark) Andersen, Steffen; Fountain, John; Harrison, Glenn W. 2014-01-01 either construct elicitation mechanisms that control for risk aversion, or construct elicitation mechanisms which undertake 'calibrating adjustments' to elicited reports. We illustrate how the joint estimation of risk attitudes and subjective probabilities can provide the calibration adjustments...... that theory calls for. We illustrate this approach using data from a controlled experiment with real monetary consequences to the subjects. This allows the observer to make inferences about the latent subjective probability, under virtually any well-specified model of choice under subjective risk, while still... 20. Probability Aggregates in Probability Answer Set Programming OpenAIRE 2013-01-01 Probability answer set programming is a declarative programming that has been shown effective for representing and reasoning about a variety of probability reasoning tasks. However, the lack of probability aggregates, e.g. {\\em expected values}, in the language of disjunctive hybrid probability logic programs (DHPP) disallows the natural and concise representation of many interesting problems. In this paper, we extend DHPP to allow arbitrary probability aggregates. We introduce two types of p... 1. Collision Probability Analysis DEFF Research Database (Denmark) Hansen, Peter Friis; Pedersen, Preben Terndrup 1998-01-01 It is the purpose of this report to apply a rational model for prediction of ship-ship collision probabilities as function of the ship and the crew characteristics and the navigational environment for MS Dextra sailing on a route between Cadiz and the Canary Islands.The most important ship and crew...... characteristics are: ship speed, ship manoeuvrability, the layout of the navigational bridge, the radar system, the number and the training of navigators, the presence of a look out etc. The main parameters affecting the navigational environment are ship traffic density, probability distributions of wind speeds...... probability, i.e. a study of the navigator's role in resolving critical situations, a causation factor is derived as a second step.The report documents the first step in a probabilistic collision damage analysis. Future work will inlcude calculation of energy released for crushing of structures giving... 2. Approximation for the Finite-Time Ruin Probability of a General Risk Model with Constant Interest Rate and Extended Negatively Dependent Heavy-Tailed Claims Directory of Open Access Journals (Sweden) Yang Yang 2011-01-01 Full Text Available We propose a general continuous-time risk model with a constant interest rate. In this model, claims arrive according to an arbitrary counting process, while their sizes have dominantly varying tails and fulfill an extended negative dependence structure. We obtain an asymptotic formula for the finite-time ruin probability, which extends a corresponding result of Wang (2008. 3. Impact of Age on the Risk of Advanced Colorectal Neoplasia in a Young Population: An Analysis Using the Predicted Probability Model. Science.gov (United States) Jung, Yoon Suk; Park, Chan Hyuk; Kim, Nam Hee; Lee, Mi Yeon; Park, Dong Il 2017-09-01 The incidence of colorectal cancer is decreasing in adults aged ≥50 years and increasing in those aged probability models for ACRN in a population aged 30-49 years. Of 96,235 participants, 57,635 and 38,600 were included in the derivation and validation cohorts, respectively. The predicted probability model considered age, sex, body mass index, family history of colorectal cancer, and smoking habits, as follows: Y ACRN  = -8.755 + 0.080·X age  - 0.055·X male  + 0.041·X BMI  + 0.200·X family_history_of_CRC  + 0.218·X former_smoker  + 0.644·X current_smoker . The optimal cutoff value for the predicted probability of ACRN by Youden index was 1.14%. The area under the receiver-operating characteristic curve (AUROC) values of our model for ACRN were higher than those of the previously established Asia-Pacific Colorectal Screening (APCS), Korean Colorectal Screening (KCS), and Kaminski's scoring models [AUROC (95% confidence interval): model in the current study, 0.673 (0.648-0.697); vs. APCS, 0.588 (0.564-0.611), P probability model can assess the risk of ACRN more accurately than existing models, including the APCS, KCS, and Kaminski's scoring models. 4. Postpartum complications International Nuclear Information System (INIS) Kronthal, A.J.; Kuhlman, J.E.; Fishman, E.K. 1990-01-01 This paper reports the CT findings of major postpartum complications and determine what role CT plays in their evaluation. The CT scans of nine patients with major postpartum complications were retrospectively reviewed. Patients had been referred to CT for evaluation of postpartum fever, abdominal pain, and elevated results of liver function tests. Complications identified at CT included hepatic infarctions (n = 2), endometritis (n = 2), postoperative wound abscess (n = 1), massive abdominal hemorrhage (n = 1), septic thrombophlebitis (n = 1), and renal vein thrombosis (n = 1). CT findings of hepatic infarction included wedge-shaped areas of decreased enhancement conforming to a vascular distribution 5. A tight-binding model of the transmission probability through a molecular junction; a single molecule vs. a molecular layer International Nuclear Information System (INIS) Landau, A.; Nitzan, A. 2006-01-01 Full Text: Molecular electronics, one of the major fields of the current effort in nano-science, may be de ed as the study of electronic behaviors, devices and applications that depend on the properties of matter at the molecular scale. If the miniaturization trend of microelectronic devices is to continue, elements such as transistors and contacts will soon shrink to single molecules. The promise of these new technological breakthroughs has been major driving force in this ld. Moreover, the consideration of molecular systems as electronic devices has raised new fundamental questions. In particular, while traditional quantum chemistry deals with electronically closed systems, we now face problems involving molecular systems that are open to their electronic environment, moreover, function in far from equilibrium situations. A generic molecular junction is made of two electrodes connected by a molecular spacer that takes the form of a molecular chain of varying length or a molecular layer of varying thickness. We use a simple nearest-neighbors tight-biding model with the non-equilibrium Green's function (NEGF) method to investigate and compare between a self-assembled monolayer (SAM), finite molecular layer (FML), and single molecule (SM) chemisorption to a surface of a metal substrate. In addition, we examine the difference in the transmission probability through a SAM, FML and SM sandwiched between two metallic electrodes. Dramatic differences are observed between the SM, FML and SAM density of electronic states and transmission functions. In addition, we analyze the effects of changing different physical parameters such as molecule-substrate interaction, molecule-molecule interactions, etc; interesting effects that pertain to the conduction properties of single molecules and molecular layers are observed. Intriguing results are attained when we investigate the commensurability of the SAM with the metallic surface 6. Tumor Control Probability Modeling for Stereotactic Body Radiation Therapy of Early-Stage Lung Cancer Using Multiple Bio-physical Models Science.gov (United States) Liu, Feng; Tai, An; Lee, Percy; Biswas, Tithi; Ding, George X.; El Naqa, Isaam; Grimm, Jimm; Jackson, Andrew; Kong, Feng-Ming (Spring); LaCouture, Tamara; Loo, Billy; Miften, Moyed; Solberg, Timothy; Li, X Allen 2017-01-01 Purpose To analyze pooled clinical data using different radiobiological models and to understand the relationship between biologically effective dose (BED) and tumor control probability (TCP) for stereotactic body radiotherapy (SBRT) of early-stage non-small cell lung cancer (NSCLC). Method and Materials The clinical data of 1-, 2-, 3-, and 5-year actuarial or Kaplan-Meier TCP from 46 selected studies were collected for SBRT of NSCLC in the literature. The TCP data were separated for Stage T1 and T2 tumors if possible, otherwise collected for combined stages. BED was calculated at isocenters using six radiobiological models. For each model, the independent model parameters were determined from a fit to the TCP data using the least chi-square (χ2) method with either one set of parameters regardless of tumor stages or two sets for T1 and T2 tumors separately. Results The fits to the clinic data yield consistent results of large α/β ratios of about 20 Gy for all models investigated. The regrowth model that accounts for the tumor repopulation and heterogeneity leads to a better fit to the data, compared to other 5 models where the fits were indistinguishable between the models. The models based on the fitting parameters predict that the T2 tumors require about additional 1 Gy physical dose at isocenters per fraction (≤5 fractions) to achieve the optimal TCP when compared to the T1 tumors. Conclusion This systematic analysis of a large set of published clinical data using different radiobiological models shows that local TCP for SBRT of early-stage NSCLC has strong dependence on BED with large α/β ratios of about 20 Gy. The six models predict that a BED (calculated with α/β of 20) of 90 Gy is sufficient to achieve TCP ≥ 95%. Among the models considered, the regrowth model leads to a better fit to the clinical data. PMID:27871671 7. Skin Complications Science.gov (United States) ... Text Size: A A A Listen En Español Skin Complications Diabetes can affect every part of the ... lipoidica diabeticorum, diabetic blisters, and eruptive xanthomatosis. General Skin Conditions Bacterial Infections Several kinds of bacterial infections ... 8. Diphtheria Complications Science.gov (United States) ... Search Form Controls Cancel Submit Search The CDC Diphtheria Note: Javascript is disabled or is not supported ... message, please visit this page: About CDC.gov . Diphtheria Home About Diphtheria Causes and Transmission Symptoms Complications ... 9. Current recommendations on the estimation of transition probabilities in Markov cohort models for use in health care decision-making: a targeted literature review Directory of Open Access Journals (Sweden) Olariu E 2017-09-01 Full Text Available Elena Olariu,1 Kevin K Cadwell,1 Elizabeth Hancock,1 David Trueman,1 Helene Chevrou-Severac2 1PHMR Ltd, London, UK; 2Takeda Pharmaceuticals International AG, Zurich, Switzerland Objective: Although Markov cohort models represent one of the most common forms of decision-analytic models used in health care decision-making, correct implementation of such models requires reliable estimation of transition probabilities. This study sought to identify consensus statements or guidelines that detail how such transition probability matrices should be estimated. Methods: A literature review was performed to identify relevant publications in the following databases: Medline, Embase, the Cochrane Library, and PubMed. Electronic searches were supplemented by manual-searches of health technology assessment (HTA websites in Australia, Belgium, Canada, France, Germany, Ireland, Norway, Portugal, Sweden, and the UK. One reviewer assessed studies for eligibility. Results: Of the 1,931 citations identified in the electronic searches, no studies met the inclusion criteria for full-text review, and no guidelines on transition probabilities in Markov models were identified. Manual-searching of the websites of HTA agencies identified ten guidelines on economic evaluations (Australia, Belgium, Canada, France, Germany, Ireland, Norway, Portugal, Sweden, and UK. All identified guidelines provided general guidance on how to develop economic models, but none provided guidance on the calculation of transition probabilities. One relevant publication was identified following review of the reference lists of HTA agency guidelines: the International Society for Pharmacoeconomics and Outcomes Research taskforce guidance. This provided limited guidance on the use of rates and probabilities. Conclusions: There is limited formal guidance available on the estimation of transition probabilities for use in decision-analytic models. Given the increasing importance of cost 10. What subject matter questions motivate the use of machine learning approaches compared to statistical models for probability prediction? Science.gov (United States) Binder, Harald 2014-07-01 This is a discussion of the following papers: "Probability estimation with machine learning methods for dichotomous and multicategory outcome: Theory" by Jochen Kruppa, Yufeng Liu, Gérard Biau, Michael Kohler, Inke R. König, James D. Malley, and Andreas Ziegler; and "Probability estimation with machine learning methods for dichotomous and multicategory outcome: Applications" by Jochen Kruppa, Yufeng Liu, Hans-Christian Diener, Theresa Holste, Christian Weimar, Inke R. König, and Andreas Ziegler. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim. 11. [Biometric bases: basic concepts of probability calculation]. Science.gov (United States) Dinya, E 1998-04-26 The author gives or outline of the basic concepts of probability theory. The bases of the event algebra, definition of the probability, the classical probability model and the random variable are presented. 12. Robust estimation of the expected survival probabilities from high-dimensional Cox models with biomarker-by-treatment interactions in randomized clinical trials Directory of Open Access Journals (Sweden) Nils Ternès 2017-05-01 Full Text Available Abstract Background Thanks to the advances in genomics and targeted treatments, more and more prediction models based on biomarkers are being developed to predict potential benefit from treatments in a randomized clinical trial. Despite the methodological framework for the development and validation of prediction models in a high-dimensional setting is getting more and more established, no clear guidance exists yet on how to estimate expected survival probabilities in a penalized model with biomarker-by-treatment interactions. Methods Based on a parsimonious biomarker selection in a penalized high-dimensional Cox model (lasso or adaptive lasso, we propose a unified framework to: estimate internally the predictive accuracy metrics of the developed model (using double cross-validation; estimate the individual survival probabilities at a given timepoint; construct confidence intervals thereof (analytical or bootstrap; and visualize them graphically (pointwise or smoothed with spline. We compared these strategies through a simulation study covering scenarios with or without biomarker effects. We applied the strategies to a large randomized phase III clinical trial that evaluated the effect of adding trastuzumab to chemotherapy in 1574 early breast cancer patients, for which the expression of 462 genes was measured. Results In our simulations, penalized regression models using the adaptive lasso estimated the survival probability of new patients with low bias and standard error; bootstrapped confidence intervals had empirical coverage probability close to the nominal level across very different scenarios. The double cross-validation performed on the training data set closely mimicked the predictive accuracy of the selected models in external validation data. We also propose a useful visual representation of the expected survival probabilities using splines. In the breast cancer trial, the adaptive lasso penalty selected a prediction model with 4 13. Metabolic Disorders and Diabetic Complications in Spontaneously Diabetic Torii Leprfa Rat: A New Obese Type 2 Diabetic Model Directory of Open Access Journals (Sweden) Yusuke Kemmochi 2013-01-01 Full Text Available Spontaneously Diabetic Torii Leprfa (SDT fatty rat, established by introducing the fa allele of the Zucker fatty rat into SDT rat genome, is a new model of obese type 2 diabetes. Both male and female SDT fatty rats show overt obesity, and hyperglycemia and hyperlipidemia are observed at a young age as compared with SDT rats. With early incidence of diabetes mellitus, diabetic complications, such as nephropathy, retinopathy, and neuropathy, in SDT fatty rats were seen at younger ages compared to those in the SDT rats. In this paper, we overview pathophysiological features in SDT fatty rats and also describe new insights regarding the hematology, blood pressure, renal complications, and sexual dysfunction. The SDT fatty rats showed an increase of leukocytes, especially the monocyte count, prominent hypertension associated with salt drinking, end-stage renal disease with aging, and hypogonadism. Unlike other diabetic models, the characteristic of SDT fatty rat is to present an incidence of diabetes in females, hypertension, and retinopathy. SDT fatty rat is a useful model for analysis of various metabolic disorders and the evaluation of drugs related to metabolic disease. 14. WE-AB-202-10: Modelling Individual Tumor-Specific Control Probability for Hypoxia in Rectal Cancer Energy Technology Data Exchange (ETDEWEB) Warren, S; Warren, DR; Wilson, JM; Muirhead, R; Hawkins, MA; Maughan, T; Partridge, M [University of Oxford, Oxford, Oxfordshire (United Kingdom) 2016-06-15 Purpose: To investigate hypoxia-guided dose-boosting for increased tumour control and improved normal tissue sparing using FMISO-PET images Methods: Individual tumor-specific control probability (iTSCP) was calculated using a modified linear-quadratic model with rectal-specific radiosensitivity parameters for three limiting-case assumptions of the hypoxia / FMISO uptake relationship. {sup 18}FMISO-PET images from 2 patients (T3N0M0) from the RHYTHM trial (Investigating Hypoxia in Rectal Tumours NCT02157246) were chosen to delineate a hypoxic region (GTV-MISO defined as tumor-to-muscle ratio > 1.3) within the anatomical GTV. Three VMAT treatment plans were created in Eclipse (Varian): STANDARD (45Gy / 25 fractions to PTV4500); BOOST-GTV (simultaneous integrated boost of 60Gy / 25fr to GTV +0.5cm) and BOOST-MISO (60Gy / 25fr to GTV-MISO+0.5cm). GTV mean dose (in EQD2), iTSCP and normal tissue dose-volume metrics (small bowel, bladder, anus, and femoral heads) were recorded. Results: Patient A showed small hypoxic volume (15.8% of GTV) and Patient B moderate hypoxic volume (40.2% of GTV). Dose escalation to 60Gy was achievable, and doses to femoral heads and small bowel in BOOST plans were comparable to STANDARD plans. For patient A, a reduced maximum bladder dose was observed in BOOST-MISO compared to BOOST-GTV (D0.1cc 49.2Gy vs 54.0Gy). For patient B, a smaller high dose volume was observed for the anus region in BOOST-MISO compared to BOOST-GTV (V55Gy 19.9% vs 100%), which could potentially reduce symptoms of fecal incontinence. For BOOST-MISO, the largest iTSCPs (A: 95.5% / B: 90.0%) assumed local correlation between FMISO uptake and hypoxia, and approached iTSCP values seen for BOOST-GTV (A: 96.1% / B: 90.5%). Conclusion: Hypoxia-guided dose-boosting is predicted to improve local control in rectal tumors when FMISO is spatially correlated to hypoxia, and to reduce dose to organs-at-risk compared to boosting the whole GTV. This could lead to organ 15. Converting dose distributions into tumour control probability International Nuclear Information System (INIS) Nahum, A.E. 1996-01-01 The endpoints in radiotherapy that are truly of relevance are not dose distributions but the probability of local control, sometimes known as the Tumour Control Probability (TCP) and the Probability of Normal Tissue Complications (NTCP). A model for the estimation of TCP based on simple radiobiological considerations is described. It is shown that incorporation of inter-patient heterogeneity into the radiosensitivity parameter a through s a can result in a clinically realistic slope for the dose-response curve. The model is applied to inhomogeneous target dose distributions in order to demonstrate the relationship between dose uniformity and s a . The consequences of varying clonogenic density are also explored. Finally the model is applied to the target-volume DVHs for patients in a clinical trial of conformal pelvic radiotherapy; the effect of dose inhomogeneities on distributions of TCP are shown as well as the potential benefits of customizing the target dose according to normal-tissue DVHs. (author). 37 refs, 9 figs 16. Converting dose distributions into tumour control probability Energy Technology Data Exchange (ETDEWEB) Nahum, A E [The Royal Marsden Hospital, London (United Kingdom). Joint Dept. of Physics 1996-08-01 The endpoints in radiotherapy that are truly of relevance are not dose distributions but the probability of local control, sometimes known as the Tumour Control Probability (TCP) and the Probability of Normal Tissue Complications (NTCP). A model for the estimation of TCP based on simple radiobiological considerations is described. It is shown that incorporation of inter-patient heterogeneity into the radiosensitivity parameter a through s{sub a} can result in a clinically realistic slope for the dose-response curve. The model is applied to inhomogeneous target dose distributions in order to demonstrate the relationship between dose uniformity and s{sub a}. The consequences of varying clonogenic density are also explored. Finally the model is applied to the target-volume DVHs for patients in a clinical trial of conformal pelvic radiotherapy; the effect of dose inhomogeneities on distributions of TCP are shown as well as the potential benefits of customizing the target dose according to normal-tissue DVHs. (author). 37 refs, 9 figs. 17. Light Modulates Ocular Complications in an Albino Rat Model of Type 1 Diabetes Mellitus. Science.gov (United States) Andrawus, Elias; Veildbaum, Gizi; Zemel, Esther; Leibu, Rina; Perlman, Ido; Shehadeh, Naim 2017-07-01 The purpose of the study was to assess potential interactions of light exposure and hyperglycemia upon ocular complications in diabetic rats. Streptozotocin-induced (STZ-induced) diabetic rats ( N = 39) and non-diabetic rats ( N = 9) were distributed into eight groups according to the irradiance and color of the light phase during the 12/12-hour light/dark regime. Follow-up lasted 90 days and included assessment of cataract development and electroretinogram (ERG) recordings. Stress to the retina was also assessed by glial fibrillary acidic protein immunocytochemistry. Cataract development was fast in diabetic rats that were exposed to unattenuated white light or to bright colored lights during the light phase. Diabetic rats that were kept under attenuated brown or yellow light during the light phase exhibited slower rate of cataract development. Electroretinogram responses indicated very severe retinal damage in diabetic rats kept under bright colored lights in the blue-yellow range or bright white light during the light phase. Electroretinogram damage was milder in rats kept under bright red light or attenuated yellow or brown light during the light phase. Glial fibrillary acidic protein expression in retinal Müller cells was consistent with ERG assessment of retinal damage. Attenuating white light and filtering out short wavelengths have a protective effect on the eyes of diabetic rats as evident by slower rate of cataract formation and a smaller degree of retinal damage. Our findings suggest that special glasses attenuating light exposure and filtering out short wavelengths (400-530 nm) may be beneficial for diabetic patients. 18. A composite mouse model of aplastic anemia complicated with iron overload. Science.gov (United States) Wu, Dijiong; Wen, Xiaowen; Liu, Wenbin; Xu, Linlong; Ye, Baodong; Zhou, Yuhong 2018-02-01 Iron overload is commonly encountered during the course of aplastic anemia (AA), but no composite animal model has been developed yet, which hinders drug research. In the present study, the optimal dosage and duration of intraperitoneal iron dextran injection for the development of an iron overload model in mice were explored. A composite model of AA was successfully established on the principle of immune-mediated bone marrow failure. Liver volume, peripheral hemogram, bone marrow pathology, serum iron, serum ferritin, pathological iron deposition in multiple organs (liver, bone marrow, spleen), liver hepcidin, and bone morphogenetic protein 6 (BMP6), SMAD family member 4 (SMAD4) and transferrin receptor 2 (TfR2) mRNA expression levels were compared among the normal control, AA, iron overload and composite model groups to validate the composite model, and explore the pathogenesis and features of iron overload in this model. The results indicated marked increases in iron deposits, with significantly increased liver/body weight ratios as well as serum iron and ferritin in the iron overload and composite model groups as compared with the normal control and AA groups (Poverload and AA was successfully established, and AA was indicated to possibly have a critical role in abnormal iron metabolism, which promoted the development of iron deposits. 19. Neuromuscular complications of thyrotoxicosis. Science.gov (United States) Kung, Annie W C 2007-11-01 Thyroid hormones exert multiple effects on the neuromuscular system and the brain, with the most important being their role in stimulating the development and differentiation of the neuromuscular system and brain in foetal and neonatal life. In the presence of hyperthyroidism, muscular and neurological symptoms may be the presenting clinical features of the disease. The frequency and severity of neuromuscular complications vary considerably and are probably related to the degree of hyperthyroidism, although in some patients the neuromuscular dysfunction is caused by associated disorders rather than by hyperthyroidism per se. This update focuses on the most common neurological and muscular disorders that occur in patients with thyrotoxicosis. It is beyond the scope of this paper to discuss thyroid eye disease and cardiac complications, in themselves separate complications of specific myocytes. 20. The creation and evaluation of a model predicting the probability of conception in seasonal-calving, pasture-based dairy cows. Science.gov (United States) Fenlon, Caroline; O'Grady, Luke; Doherty, Michael L; Dunnion, John; Shalloo, Laurence; Butler, Stephen T 2017-07-01 Reproductive performance in pasture-based production systems has a fundamentally important effect on economic efficiency. The individual factors affecting the probability of submission and conception are multifaceted and have been extensively researched. The present study analyzed some of these factors in relation to service-level probability of conception in seasonal-calving pasture-based dairy cows to develop a predictive model of conception. Data relating to 2,966 services from 737 cows on 2 research farms were used for model development and data from 9 commercial dairy farms were used for model testing, comprising 4,212 services from 1,471 cows. The data spanned a 15-yr period and originated from seasonal-calving pasture-based dairy herds in Ireland. The calving season for the study herds extended from January to June, with peak calving in February and March. A base mixed-effects logistic regression model was created using a stepwise model-building strategy and incorporated parity, days in milk, interservice interval, calving difficulty, and predicted transmitting abilities for calving interval and milk production traits. To attempt to further improve the predictive capability of the model, the addition of effects that were not statistically significant was considered, resulting in a final model composed of the base model with the inclusion of BCS at service. The models' predictions were evaluated using discrimination to measure their ability to correctly classify positive and negative cases. Precision, recall, F-score, and area under the receiver operating characteristic curve (AUC) were calculated. Calibration tests measured the accuracy of the predicted probabilities. These included tests of overall goodness-of-fit, bias, and calibration error. Both models performed better than using the population average probability of conception. Neither of the models showed high levels of discrimination (base model AUC 0.61, final model AUC 0.62), possibly because of the 1. An Empirical Model for Estimating the Probability of Electrical Short Circuits from Tin Whiskers-Part I Science.gov (United States) Courey, Karim; Wright, Clara; Asfour, Shihab; Bayliss, Jon; Ludwig, Larry 2008-01-01 Existing risk simulations make the assumption that when a free tin whisker has bridged two adjacent exposed electrical conductors, the result is an electrical short circuit. This conservative assumption is made because shorting is a random event that has a currently unknown probability associated with it. Due to contact resistance, electrical shorts may not occur at lower voltage levels. In this experiment, we study the effect of varying voltage on the breakdown of the contact resistance which leads to a short circuit. From this data, we can estimate the probability of an electrical short, as a function of voltage, given that a free tin whisker has bridged two adjacent exposed electrical conductors. In addition, three tin whiskers grown from the same Space Shuttle Orbiter card guide used in the aforementioned experiment were cross sectioned and studied using a focused ion beam (FIB). 2. Two-slit experiment: quantum and classical probabilities International Nuclear Information System (INIS) Khrennikov, Andrei 2015-01-01 Inter-relation between quantum and classical probability models is one of the most fundamental problems of quantum foundations. Nowadays this problem also plays an important role in quantum technologies, in quantum cryptography and the theory of quantum random generators. In this letter, we compare the viewpoint of Richard Feynman that the behavior of quantum particles cannot be described by classical probability theory with the viewpoint that quantum–classical inter-relation is more complicated (cf, in particular, with the tomographic model of quantum mechanics developed in detail by Vladimir Man'ko). As a basic example, we consider the two-slit experiment, which played a crucial role in quantum foundational debates at the beginning of quantum mechanics (QM). In particular, its analysis led Niels Bohr to the formulation of the principle of complementarity. First, we demonstrate that in complete accordance with Feynman's viewpoint, the probabilities for the two-slit experiment have the non-Kolmogorovian structure, since they violate one of basic laws of classical probability theory, the law of total probability (the heart of the Bayesian analysis). However, then we show that these probabilities can be embedded in a natural way into the classical (Kolmogorov, 1933) probability model. To do this, one has to take into account the randomness of selection of different experimental contexts, the joint consideration of which led Feynman to a conclusion about the non-classicality of quantum probability. We compare this embedding of non-Kolmogorovian quantum probabilities into the Kolmogorov model with well-known embeddings of non-Euclidean geometries into Euclidean space (e.g., the Poincaré disk model for the Lobachvesky plane). (paper) 3. Constructing probability distributions of uncertain variables in models of the performance of the Waste Isolation Pilot Plant: The 1990 performance simulations International Nuclear Information System (INIS) Tierney, M.S. 1990-12-01 A five-step procedure was used in the 1990 performance simulations to construct probability distributions of the uncertain variables appearing in the mathematical models used to simulate the Waste Isolation Pilot Plant's (WIPP's) performance. This procedure provides a consistent approach to the construction of probability distributions in cases where empirical data concerning a variable are sparse or absent and minimizes the amount of spurious information that is often introduced into a distribution by assumptions of nonspecialists. The procedure gives first priority to the professional judgment of subject-matter experts and emphasizes the use of site-specific empirical data for the construction of the probability distributions when such data are available. In the absence of sufficient empirical data, the procedure employs the Maximum Entropy Formalism and the subject-matter experts' subjective estimates of the parameters of the distribution to construct a distribution that can be used in a performance simulation. (author) 4. Constructing probability distributions of uncertain variables in models of the performance of the Waste Isolation Pilot Plant: The 1990 performance simulations Energy Technology Data Exchange (ETDEWEB) Tierney, M S 1990-12-15 A five-step procedure was used in the 1990 performance simulations to construct probability distributions of the uncertain variables appearing in the mathematical models used to simulate the Waste Isolation Pilot Plant's (WIPP's) performance. This procedure provides a consistent approach to the construction of probability distributions in cases where empirical data concerning a variable are sparse or absent and minimizes the amount of spurious information that is often introduced into a distribution by assumptions of nonspecialists. The procedure gives first priority to the professional judgment of subject-matter experts and emphasizes the use of site-specific empirical data for the construction of the probability distributions when such data are available. In the absence of sufficient empirical data, the procedure employs the Maximum Entropy Formalism and the subject-matter experts' subjective estimates of the parameters of the distribution to construct a distribution that can be used in a performance simulation. (author) 5. Bayesian Probability Theory Science.gov (United States) von der Linden, Wolfgang; Dose, Volker; von Toussaint, Udo 2014-06-01 Preface; Part I. Introduction: 1. The meaning of probability; 2. Basic definitions; 3. Bayesian inference; 4. Combinatrics; 5. Random walks; 6. Limit theorems; 7. Continuous distributions; 8. The central limit theorem; 9. Poisson processes and waiting times; Part II. Assigning Probabilities: 10. Transformation invariance; 11. Maximum entropy; 12. Qualified maximum entropy; 13. Global smoothness; Part III. Parameter Estimation: 14. Bayesian parameter estimation; 15. Frequentist parameter estimation; 16. The Cramer-Rao inequality; Part IV. Testing Hypotheses: 17. The Bayesian way; 18. The frequentist way; 19. Sampling distributions; 20. Bayesian vs frequentist hypothesis tests; Part V. Real World Applications: 21. Regression; 22. Inconsistent data; 23. Unrecognized signal contributions; 24. Change point problems; 25. Function estimation; 26. Integral equations; 27. Model selection; 28. Bayesian experimental design; Part VI. Probabilistic Numerical Techniques: 29. Numerical integration; 30. Monte Carlo methods; 31. Nested sampling; Appendixes; References; Index. 6. Modelling Gaucher disease progression: long-term enzyme replacement therapy reduces the incidence of splenectomy and bone complications NARCIS (Netherlands) van Dussen, Laura; Biegstraaten, Marieke; Dijkgraaf, Marcel Gw; Hollak, Carla Em 2014-01-01 Long-term complications and associated conditions of type 1 Gaucher Disease (GD) can include splenectomy, bone complications, pulmonary hypertension, Parkinson disease and malignancies. Enzyme replacement therapy (ERT) reverses cytopenia and reduces organomegaly. To study the effects of ERT on 7. Evaluating the effect of corridors and landscape heterogeneity on dispersal probability: a comparison of three spatially explicit modelling approaches DEFF Research Database (Denmark) Jepsen, J. U.; Baveco, J. M.; Topping, C. J. 2004-01-01 preferences of the modeller, rather than by a critical evaluation of model performance. We present a comparison of three common spatial simulation approaches (patch-based incidence-function model (IFM), individual-based movement model (IBMM), individual-based population model including detailed behaviour... 8. Probability 1/e Science.gov (United States) Koo, Reginald; Jones, Martin L. 2011-01-01 Quite a number of interesting problems in probability feature an event with probability equal to 1/e. This article discusses three such problems and attempts to explain why this probability occurs with such frequency. 9. Probability an introduction CERN Document Server Goldberg, Samuel 1960-01-01 Excellent basic text covers set theory, probability theory for finite sample spaces, binomial theorem, probability distributions, means, standard deviations, probability function of binomial distribution, more. Includes 360 problems with answers for half. 10. Contributions to quantum probability International Nuclear Information System (INIS) Fritz, Tobias 2010-01-01 Chapter 1: On the existence of quantum representations for two dichotomic measurements. Under which conditions do outcome probabilities of measurements possess a quantum-mechanical model? This kind of problem is solved here for the case of two dichotomic von Neumann measurements which can be applied repeatedly to a quantum system with trivial dynamics. The solution uses methods from the theory of operator algebras and the theory of moment problems. The ensuing conditions reveal surprisingly simple relations between certain quantum-mechanical probabilities. It also shown that generally, none of these relations holds in general probabilistic models. This result might facilitate further experimental discrimination between quantum mechanics and other general probabilistic theories. Chapter 2: Possibilistic Physics. I try to outline a framework for fundamental physics where the concept of probability gets replaced by the concept of possibility. Whereas a probabilistic theory assigns a state-dependent probability value to each outcome of each measurement, a possibilistic theory merely assigns one of the state-dependent labels ''possible to occur'' or ''impossible to occur'' to each outcome of each measurement. It is argued that Spekkens' combinatorial toy theory of quantum mechanics is inconsistent in a probabilistic framework, but can be regarded as possibilistic. Then, I introduce the concept of possibilistic local hidden variable models and derive a class of possibilistic Bell inequalities which are violated for the possibilistic Popescu-Rohrlich boxes. The chapter ends with a philosophical discussion on possibilistic vs. probabilistic. It can be argued that, due to better falsifiability properties, a possibilistic theory has higher predictive power than a probabilistic one. Chapter 3: The quantum region for von Neumann measurements with postselection. It is determined under which conditions a probability distribution on a finite set can occur as the outcome 11. Contributions to quantum probability Energy Technology Data Exchange (ETDEWEB) Fritz, Tobias 2010-06-25 Chapter 1: On the existence of quantum representations for two dichotomic measurements. Under which conditions do outcome probabilities of measurements possess a quantum-mechanical model? This kind of problem is solved here for the case of two dichotomic von Neumann measurements which can be applied repeatedly to a quantum system with trivial dynamics. The solution uses methods from the theory of operator algebras and the theory of moment problems. The ensuing conditions reveal surprisingly simple relations between certain quantum-mechanical probabilities. It also shown that generally, none of these relations holds in general probabilistic models. This result might facilitate further experimental discrimination between quantum mechanics and other general probabilistic theories. Chapter 2: Possibilistic Physics. I try to outline a framework for fundamental physics where the concept of probability gets replaced by the concept of possibility. Whereas a probabilistic theory assigns a state-dependent probability value to each outcome of each measurement, a possibilistic theory merely assigns one of the state-dependent labels ''possible to occur'' or ''impossible to occur'' to each outcome of each measurement. It is argued that Spekkens' combinatorial toy theory of quantum mechanics is inconsistent in a probabilistic framework, but can be regarded as possibilistic. Then, I introduce the concept of possibilistic local hidden variable models and derive a class of possibilistic Bell inequalities which are violated for the possibilistic Popescu-Rohrlich boxes. The chapter ends with a philosophical discussion on possibilistic vs. probabilistic. It can be argued that, due to better falsifiability properties, a possibilistic theory has higher predictive power than a probabilistic one. Chapter 3: The quantum region for von Neumann measurements with postselection. It is determined under which conditions a probability distribution on a 12. Modeling of absorption data complicated by Fabry endash Perot interference in germanosilicate thin-film waveguides International Nuclear Information System (INIS) Simmons-Potter, K.; Simmons, J.H. 1996-01-01 Complex absorption spectra obtained from thin films at normal incidence can be difficult to interpret owing to the appearance of Fabry endash Perot interference fringes in the data. We describe a technique for modeling such spectra so that true absorption features can be identified and evaluated separately from the overlying fringes. The technique is used to interpret data obtained from photosensitive germanosilicate solgel films on fused-silica substrates but may be easily extended to analysis in other material systems. copyright 1996 Optical Society of America 13. Expected utility with lower probabilities DEFF Research Database (Denmark) Hendon, Ebbe; Jacobsen, Hans Jørgen; Sloth, Birgitte 1994-01-01 An uncertain and not just risky situation may be modeled using so-called belief functions assigning lower probabilities to subsets of outcomes. In this article we extend the von Neumann-Morgenstern expected utility theory from probability measures to belief functions. We use this theory... 14. Evaluating the suitability of wind speed probability distribution models: A case of study of east and southeast parts of Iran International Nuclear Information System (INIS) Alavi, Omid; Mohammadi, Kasra; Mostafaeipour, Ali 2016-01-01 Highlights: • Suitability of different wind speed probability functions is assessed. • 5 stations distributed in east and south-east of Iran are considered as case studies. • Nakagami distribution is tested for first time and compared with 7 other functions. • Due to difference in wind features, best function is not similar for all stations. - Abstract: Precise information of wind speed probability distribution is truly significant for many wind energy applications. The objective of this study is to evaluate the suitability of different probability functions for estimating wind speed distribution at five stations, distributed in the east and southeast of Iran. Nakagami distribution function is utilized for the first time to estimate the distribution of wind speed. The performance of Nakagami function is compared with seven typically used distribution functions. The achieved results reveal that the more effective function is not similar among all stations. Wind speed characteristics, quantity and quality of the recorded wind speed data can be considered as influential parameters on the performance of the distribution functions. Also, the skewness of the recorded wind speed data may have influence on the accuracy of the Nakagami distribution. For Chabahar and Khaf stations the Nakagami distribution shows the highest performance while for Lutak, Rafsanjan and Zabol stations the Gamma, Generalized Extreme Value and Inverse-Gaussian distributions offer the best fits, respectively. Based on the analysis, the Nakagami distribution can generally be considered as an effective distribution since it provides the best fits in 2 stations and ranks 3rd to 5th in the remaining stations; however, due to the close performance of the Nakagami and Weibull distributions and also flexibility of the Weibull function as its widely proven feature, more assessments on the performance of the Nakagami distribution are required. 15. U.S. Nuclear Regulatory Commission Extremely Low Probability of Rupture pilot study: xLPR framework model user's guide International Nuclear Information System (INIS) Kalinich, Donald A.; Sallaberry, Cedric M.; Mattie, Patrick D. 2010-01-01 For the U.S. Nuclear Regulatory Commission (NRC) Extremely Low Probability of Rupture (xLPR) pilot study, Sandia National Laboratories (SNL) was tasked to develop and evaluate a probabilistic framework using a commercial software package for Version 1.0 of the xLPR Code. Version 1.0 of the xLPR code is focused assessing the probability of rupture due to primary water stress corrosion cracking in dissimilar metal welds in pressurizer surge nozzles. Future versions of this framework will expand the capabilities to other cracking mechanisms, and other piping systems for both pressurized water reactors and boiling water reactors. The goal of the pilot study project is to plan the xLPR framework transition from Version 1.0 to Version 2.0; hence the initial Version 1.0 framework and code development will be used to define the requirements for Version 2.0. The software documented in this report has been developed and tested solely for this purpose. This framework and demonstration problem will be used to evaluate the commercial software's capabilities and applicability for use in creating the final version of the xLPR framework. This report details the design, system requirements, and the steps necessary to use the commercial-code based xLPR framework developed by SNL. 16. Quantum probability measures and tomographic probability densities NARCIS (Netherlands) Amosov, GG; Man'ko, [No Value 2004-01-01 Using a simple relation of the Dirac delta-function to generalized the theta-function, the relationship between the tomographic probability approach and the quantum probability measure approach with the description of quantum states is discussed. The quantum state tomogram expressed in terms of the 17. Modelling of oil spill frequency, leak sources and contamination probability in the Caspian Sea using multi-temporal SAR images 2006–2010 and stochastic modelling Directory of Open Access Journals (Sweden) Emil Bayramov 2016-05-01 Full Text Available The main goal of this research was to detect oil spills, to determine the oil spill frequencies and to approximate oil leak sources around the Oil Rocks Settlement, the Chilov and Pirallahi Islands in the Caspian Sea using 136 multi-temporal ENVISAT Advanced Synthetic Aperture Radar Wide Swath Medium Resolution images acquired during 2006–2010. The following oil spill frequencies were observed around the Oil Rocks Settlement, the Chilov and Pirallahi Islands: 2–10 (3471.04 sq km, 11–20 (971.66 sq km, 21–50 (692.44 sq km, 51–128 (191.38 sq km. The most critical oil leak sources with the frequency range of 41–128 were observed at the Oil Rocks Settlement. The exponential regression analysis between wind speeds and oil slick areas detected from 136 multi-temporal ENVISAT images revealed the regression coefficient equal to 63%. The regression model showed that larger oil spill areas were observed with decreasing wind speeds. The spatiotemporal patterns of currents in the Caspian Sea explained the multi-directional spatial distribution of oil spills around Oil Rocks Settlement, the Chilov and Pirallahi Islands. The linear regression analysis between detected oil spill frequencies and predicted oil contamination probability by the stochastic model showed the positive trend with the regression coefficient of 30%. 18. Mathematical model of diabetic encephalopathy in diagnosis of complicated forms of diabetes mellitus Directory of Open Access Journals (Sweden) A.O. Popruga 2017-09-01 Full Text Available Background. The purpose of this research is to optimize the methods for diagnosis of diabetic encephalopathy based on the study of indicators of cerebrovascular hemodynamics, functional state of the brain, metabolic disorders and morphological characteristics of the brain tissue. Materials and methods. A comprehensive survey was carried out in 537 patients with diabetes mellitus (DM, including 342 (63.7 % persons with type 1 DM, and 195 (36.3 % — with type 2 DM. Results. The article presents data on the integrated study of clinical, metabolic and functional indicators as risk factors for diabetic encephalopathy. Their diagnostic significance is argued. On the basis of a comprehensive assessment of the obtained data, which expanded the view on the pathogenesis of diabetic encephalopathy, the priority of metabolic disorders was confirmed. Diagnostic criteria of diabetic encephalopathy were established and its mathematical model was developed. Conclusions. The availa­bility of informative indicators identified will allow the doctor to diagnose diabetic encephalopathy at the early stages or to predict its development and to detect at the preclinical stage. 19. A Model of Left Ventricular Dysfunction Complicated by CAWS Arteritis in DBA/2 Mice Directory of Open Access Journals (Sweden) Naoto Hirata 2012-01-01 Full Text Available It was reported previously that a Candida albicans water-soluble fraction (CAWS, including a mannoprotein and β-glucan complex, has strong potency in inducing fatal necrotizing arteritis in DBA/2 mice. In this study, histopathological changes and cardiac function were investigated in this system. One mg/day of CAWS was given to DBA/2 mice via peritoneal injection for five days. The CAWS-treated DBA/2 mice were induced aortitis and died at an incidence of 100% within several weeks. Histological findings included stenosis in the left ventricular outflow tract (LVOT and severe inflammatory changes of the aortic valve with fibrinoid necrosis. Cardiomegaly was observed and heart weight increased 1.62 fold (<0.01. Echocardiography revealed a severe reduction in contractility and dilatation of the cavity in the left ventricle (LV: LV fractional shortening (LVFS decreased from 71% to 38% (<0.01, and the LV end-diastolic diameter (LVDd increased from 2.21 mm to 3.26 mm (<0.01. The titer of BNP mRNA increased in the CAWS-treated group. Severe inflammatory changes resulting from CAWS brought about lethal LV dysfunction by aortic valve deformation with LVOT stenosis. This system is proposed as an easy and useful experimental model of heart failure because CAWS arteritis can be induced by CAWS injection alone. 20. Emotional Stress and Cardiovascular Complications in Animal Models: A Review of the Influence of Stress Type. Science.gov (United States) Crestani, Carlos C 2016-01-01 Emotional stress has been recognized as a modifiable risk factor for cardiovascular diseases. The impact of stress on physiological and psychological processes is determined by characteristics of the stress stimulus. For example, distinct responses are induced by acute vs. chronic aversive stimuli. Additionally, the magnitude of stress responses has been reported to be inversely related to the degree of predictability of the aversive stimulus. Therefore, the purpose of the present review was to discuss experimental research in animal models describing the influence of stressor stimulus characteristics, such as chronicity and predictability, in cardiovascular dysfunctions induced by emotional stress. Regarding chronicity, the importance of cardiovascular and autonomic adjustments during acute stress sessions and cardiovascular consequences of frequent stress response activation during repeated exposure to aversive threats (i.e., chronic stress) is discussed. Evidence of the cardiovascular and autonomic changes induced by chronic stressors involving daily exposure to the same stressor (predictable) vs. different stressors (unpredictable) is reviewed and discussed in terms of the impact of predictability in cardiovascular dysfunctions induced by stress. 1. An integrated logistic formula for prediction of complications from radiosurgery International Nuclear Information System (INIS) Flickinger, J.C. 1989-01-01 2. [Risk factor analysis of the patients with solitary pulmonary nodules and establishment of a prediction model for the probability of malignancy]. Science.gov (United States) Wang, X; Xu, Y H; Du, Z Y; Qian, Y J; Xu, Z H; Chen, R; Shi, M H 2018-02-23 Objective: This study aims to analyze the relationship among the clinical features, radiologic characteristics and pathological diagnosis in patients with solitary pulmonary nodules, and establish a prediction model for the probability of malignancy. Methods: Clinical data of 372 patients with solitary pulmonary nodules who underwent surgical resection with definite postoperative pathological diagnosis were retrospectively analyzed. In these cases, we collected clinical and radiologic features including gender, age, smoking history, history of tumor, family history of cancer, the location of lesion, ground-glass opacity, maximum diameter, calcification, vessel convergence sign, vacuole sign, pleural indentation, speculation and lobulation. The cases were divided to modeling group (268 cases) and validation group (104 cases). A new prediction model was established by logistic regression analying the data from modeling group. Then the data of validation group was planned to validate the efficiency of the new model, and was compared with three classical models(Mayo model, VA model and LiYun model). With the calculated probability values for each model from validation group, SPSS 22.0 was used to draw the receiver operating characteristic curve, to assess the predictive value of this new model. Results: 112 benign SPNs and 156 malignant SPNs were included in modeling group. Multivariable logistic regression analysis showed that gender, age, history of tumor, ground -glass opacity, maximum diameter, and speculation were independent predictors of malignancy in patients with SPN( P prediction model for the probability of malignancy as follow: p =e(x)/(1+ e(x)), x=-4.8029-0.743×gender+ 0.057×age+ 1.306×history of tumor+ 1.305×ground-glass opacity+ 0.051×maximum diameter+ 1.043×speculation. When the data of validation group was added to the four-mathematical prediction model, The area under the curve of our mathematical prediction model was 0.742, which is greater 3. Planejamento tridimensional para radioterapia de tumores de esôfago: comparação de técnicas de tratamento e análise de probabilidade de complicações Three-dimensional radiotherapy planning system for esophageal tumors: comparison of treatment techniques and analysis of probability of complications Directory of Open Access Journals (Sweden) Pitágoras Báskara Justino 2003-06-01 Full Text Available Para comparar diversas técnicas de irradiação para o câncer de esôfago, foi utilizado sistema de planejamento tridimensional. Em um paciente com carcinoma espinocelular de esôfago médio, foram estudadas as seguintes técnicas de tratamento: dois campos ântero-posteriores e dois campos látero-laterais paralelos e opostos, três campos em "Y" e em "T" e quatro campos em "X". Foram obtidos os histogramas dose-volume, considerando como órgãos de risco medula espinhal e pulmões. Os resultados foram analisados de acordo com as recomendações da Normal Tissue Complication Probability (NTCP e Tumor Control Probability (TCP. Quanto às doses de irradiação em pulmão, a melhor opção foi a técnica em dois campos ântero-posteriores paralelos e opostos. A medula foi mais poupada quando se utilizaram campos látero-laterais. Sugerimos a combinação de pelo menos duas técnicas de tratamento: ântero-posterior e as técnicas com campos em "Y", "T" ou látero-laterais, para o balanceamento das doses em pulmões e medula espinhal. Ou, ainda, a utilização de técnicas de três campos durante todo o tratamento.Radiotherapy techniques for esophageal cancer were compared using a three-dimensional planning system. We studied the following treatment techniques used for a patient with squamous cell carcinoma of the middle third of the esophagus: two antero-posterior and two latero-lateral parallel opposed fields, three fields ("Y" and "T", and four fields ("X". Dose-volume histograms were obtained considering spinal cord and lungs as organs at risk. Analysis was performed comparing doses in these organs as recommended by the Normal Tissue Complication Probability (NTCP and Tumor Control Probability (TCP. When only the lungs were considered the best technique was two antero-posterior parallel opposed fields. The spinal cord was best protected using latero-lateral fields. We suggest the combination of at least two treatment techniques: antero 4. Introduction of an Evaluation Tool to Predict the Probability of Success of Companies: The Innovativeness, Capabilities and Potential Model (ICP Directory of Open Access Journals (Sweden) Michael Lewrick 2009-05-01 Full Text Available Successful innovation requires management and in this paper a model to help manage the innovation process is presented. This model can be used to audit the management capability to innovate and to monitor how sales increase is related to innovativeness. The model was developed from a study of companies in the high technology cluster around Munich and validated using statistical procedures. The model was found to be effective at predicting the success or otherwise of the innovation strategy pursued by the company. The use of this model and how it can be used to identify areas for improvement are documented in this paper. 5. Normal tissue complication probability (NTCP), the clinician,s perspective International Nuclear Information System (INIS) Yeoh, E.K. 2011-01-01 Full text: 3D radiation treatment planning has enabled dose distributions to be related to the volume of normal tissues irradiated. The dose volume histograms thus derived have been utilized to set NTCP dose constraints to facilitate optimization of treatment planning. However, it is not widely appreciated that a number of important variables other than DYH's which determine NTCP in the individual patient. These variables will be discussed under the headings of patient and treatment related as well as tumour related factors. Patient related factors include age, co-morbidities such as connective tissue disease and diabetes mellitus, previous tissue/organ damage, tissue architectural organization (parallel or serial), regional tissue/organ and individual tissue/organ radiosensitivities as well as the development of severe acute toxicity. Treatment related variables which need to be considered include dose per fraction (if not the conventional 1.8012.00 Gy/fraction, particularly for IMRT), number of fractions and total dose, dose rate (particularly if combined with brachytherapy) and concurrent chemotherapy or other biological dose modifiers. Tumour related factors which impact on NTCP include infiltration of normal tissue/organ usually at presentation leading to compromised function but also with recurrent disease after radiation therapy as well as variable tumour radiosensitivities between and within tumour types. Whilst evaluation of DYH data is a useful guide in the choice of treatment plan, the current state of knowledge requires the clinician to make an educated judgement based on a consideration of the other factors. 6. Normal tissue complication models for clinically relevant acute esophagitis (≥ grade 2) in patients treated with dose differentiated accelerated radiotherapy (DART-bid) International Nuclear Information System (INIS) Zehentmayr, Franz; Söhn, Matthias; Exeli, Ann-Katrin; Wurstbauer, Karl; Tröller, Almut; Deutschmann, Heinz; Fastner, Gerd; Fussl, Christoph; Steininger, Philipp; Kranzinger, Manfred; Belka, Claus; Studnicka, Michael; Sedlmayer, Felix 2015-01-01 7. An attempt to model the probability of growth and aflatoxin B1 production of Aspergillus flavus under non-isothermal conditions in pistachio nuts. Science.gov (United States) Aldars-García, Laila; Ramos, Antonio J; Sanchis, Vicente; Marín, Sonia 2015-10-01 Human exposure to aflatoxins in foods is of great concern. The aim of this work was to use predictive mycology as a strategy to mitigate the aflatoxin burden in pistachio nuts postharvest. The probability of growth and aflatoxin B1 (AFB1) production of aflatoxigenic Aspergillus flavus, isolated from pistachio nuts, under static and non-isothermal conditions was studied. Four theoretical temperature scenarios, including temperature levels observed in pistachio nuts during shipping and storage, were used. Two types of inoculum were included: a cocktail of 25 A. flavus isolates and a single isolate inoculum. Initial water activity was adjusted to 0.87. Logistic models, with temperature and time as explanatory variables, were fitted to the probability of growth and AFB1 production under a constant temperature. Subsequently, they were used to predict probabilities under non-isothermal scenarios, with levels of concordance from 90 to 100% in most of the cases. Furthermore, the presence of AFB1 in pistachio nuts could be correctly predicted in 70-81 % of the cases from a growth model developed in pistachio nuts, and in 67-81% of the cases from an AFB1 model developed in pistachio agar. The information obtained in the present work could be used by producers and processors to predict the time for AFB1 production by A. flavus on pistachio nuts during transport and storage. Copyright © 2015 Elsevier Ltd. All rights reserved. 8. A Simple Model to Assess the Probability of Invasion in Ductal Carcinoma In Situ of the Breast Diagnosed by Needle Biopsy Directory of Open Access Journals (Sweden) Oldřich Coufal 2014-01-01 Full Text Available Objectives. The aim of the study was to develop a clinical prediction model for assessing the probability of having invasive cancer in the definitive surgical resection specimen in patients with biopsy diagnosis of ductal carcinoma in situ (DCIS of the breast, to facilitate decision making regarding axillary surgery. Methods. In 349 women with DCIS, predictors of invasion in the definitive resection specimen were identified. A model to predict the probability of invasion was developed and subsequently simplified to divide patients into two risk categories. The model’s performance was validated on another patient population. Results. Multivariate logistic regression revealed four independent predictors of invasion: (i suspicious (microinvasion in the biopsy specimen; (ii visibility of the lesion on ultrasonography; (iii size of the lesion on mammography >30 mm; (iv clinical palpability of the lesion. The actual frequency of invasion in the high-risk patient group in the test and validation population was 52.6% and 48.3%, respectively; in the low-risk group it was 16.8% and 7.1%, respectively. Conclusion. The model proved to have good performance. In patients with a low probability of invasion, an axillary procedure can be omitted without a substantial risk of additional surgery. 9. Theoretical model of the probability of fusion between deuterons within deformed lattices with microcracks at room temperature International Nuclear Information System (INIS) Frisone, Fulvio 2006-01-01 In this work we wish to demonstrate that a reaction path as the following dislocations, deformations due to thermodynamic stress and, finally, microcrack occurrence, can enhance the process of fusion of the deuterons introduced into the lattice by deuterium loading (F. Frisone, Can variations in temperature influence deuteron interaction within crystalline lattices?, Nuovo Cimento D, 18, 1279 (1996)). In fact, calculating the rate of deuteron-plasmon-deuteron fusion within a microcrack, showed, together with an enhancement of the tunneling effect, an increase of at least 2 - 3 orders of magnitude compared to the probability of fusion on the no deformed lattice. In fact, strong electric fields can take place in the microcrack and the deuterons are accelerated to the energy which is enough for the D-D tunnelling (M. Rabinowitz, High temperature superconductivity and cold fusion, Mod. Phys, Lett. B, 4, 233 (1990); J. Price Hirt and J. Lothe, Theory of Dislocation (McGraw Hill); Z. Phys., 457, 156 (1960)). These phenomena open the way to the theoretical hypothesis that a kind of chain reaction, catalyzed by the microcracks produced in the structure as a result of deuterium loading, can favour tho process of deuteron-plasmon fusion (N. W. Ashcroft and N. D. Mermin (Eds.), Solid State Physics, Chapter 25 (Saunders College, Philadelphia, 1972, pp. 492-509) 10. Mathematical analysis and modeling of epidemics of rubber tree root diseases: Probability of infection of an individual tree Energy Technology Data Exchange (ETDEWEB) Chadoeuf, J.; Joannes, H.; Nandris, D.; Pierrat, J.C. 1988-12-01 The spread of root diseases in rubber tree (Hevea brasiliensis) due to Rigidoporus lignosus and Phellinus noxius was investigated epidemiologically using data collected every 6 month during a 6-year survey in a plantation. The aim of the present study is to see what factors could predict whether a given tree would be infested at the following inspection. Using a qualitative regression method we expressed the probability of pathogenic attack on a tree in terms of three factors: the state of health of the surrounding trees, the method used to clear the forest prior to planting, and evolution with time. The effects of each factor were ranked, and the roles of the various classes of neighbors were established and quantified. Variability between successive inspections was small, and the method of forest clearing was important only while primary inocula in the soil were still infectious. The state of health of the immediate neighbors was most significant; more distant neighbors in the same row had some effect; interrow spread was extremely rare. This investigation dealt only with trees as individuals, and further study of the interrelationships of groups of trees is needed. 11. Informing a Tidal Turbine Strike Probability Model through Characterization of Fish Behavioral Response using Multibeam Sonar Output Energy Technology Data Exchange (ETDEWEB) Bevelhimer, Mark S. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Environmental Sciences Division; Colby, Jonathan [Verdant Power, Inc., New York, NY (United States); Adonizio, Mary Ann [Verdant Power, Inc., New York, NY (United States); Tomichek, Christine [Kleinschmidt Associates, Pittsfield, ME (United States); Scherelis, Constantin C. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Environmental Sciences Division 2016-07-31 One of the most important biological questions facing the marine and hydrokinetic (MHK) energy industry is whether fish and marine mammals that encounter MHK devices are likely to be struck by moving components. For hydrokinetic (HK) devices, i.e., those that generate energy from flowing water, this concern is greatest for large organisms because their increased length increases the probability that they will be struck as they pass through the area of blade sweep and because their increased mass means that the force absorbed if struck is greater and potentially more damaging (Amaral et al. 2015). Key to answering this question is understanding whether aquatic organisms change their swimming behavior as they encounter a device in a way that decreases their likelihood of being struck and possibly injured by the device. Whether near-field or far-field behavior results in general avoidance of or attraction to HK devices is a significant factor in the possible risk of physical contact with rotating turbine blades (Cada and Bevelhimer 2011). 12. Probability mapping of contaminants Energy Technology Data Exchange (ETDEWEB) Rautman, C.A.; Kaplan, P.G. [Sandia National Labs., Albuquerque, NM (United States); McGraw, M.A. [Univ. of California, Berkeley, CA (United States); Istok, J.D. [Oregon State Univ., Corvallis, OR (United States); Sigda, J.M. [New Mexico Inst. of Mining and Technology, Socorro, NM (United States) 1994-04-01 Exhaustive characterization of a contaminated site is a physical and practical impossibility. Descriptions of the nature, extent, and level of contamination, as well as decisions regarding proposed remediation activities, must be made in a state of uncertainty based upon limited physical sampling. The probability mapping approach illustrated in this paper appears to offer site operators a reasonable, quantitative methodology for many environmental remediation decisions and allows evaluation of the risk associated with those decisions. For example, output from this approach can be used in quantitative, cost-based decision models for evaluating possible site characterization and/or remediation plans, resulting in selection of the risk-adjusted, least-cost alternative. The methodology is completely general, and the techniques are applicable to a wide variety of environmental restoration projects. The probability-mapping approach is illustrated by application to a contaminated site at the former DOE Feed Materials Production Center near Fernald, Ohio. Soil geochemical data, collected as part of the Uranium-in-Soils Integrated Demonstration Project, have been used to construct a number of geostatistical simulations of potential contamination for parcels approximately the size of a selective remediation unit (the 3-m width of a bulldozer blade). Each such simulation accurately reflects the actual measured sample values, and reproduces the univariate statistics and spatial character of the extant data. Post-processing of a large number of these equally likely statistically similar images produces maps directly showing the probability of exceeding specified levels of contamination (potential clean-up or personnel-hazard thresholds). 13. Probability mapping of contaminants International Nuclear Information System (INIS) Rautman, C.A.; Kaplan, P.G.; McGraw, M.A.; Istok, J.D.; Sigda, J.M. 1994-01-01 Exhaustive characterization of a contaminated site is a physical and practical impossibility. Descriptions of the nature, extent, and level of contamination, as well as decisions regarding proposed remediation activities, must be made in a state of uncertainty based upon limited physical sampling. The probability mapping approach illustrated in this paper appears to offer site operators a reasonable, quantitative methodology for many environmental remediation decisions and allows evaluation of the risk associated with those decisions. For example, output from this approach can be used in quantitative, cost-based decision models for evaluating possible site characterization and/or remediation plans, resulting in selection of the risk-adjusted, least-cost alternative. The methodology is completely general, and the techniques are applicable to a wide variety of environmental restoration projects. The probability-mapping approach is illustrated by application to a contaminated site at the former DOE Feed Materials Production Center near Fernald, Ohio. Soil geochemical data, collected as part of the Uranium-in-Soils Integrated Demonstration Project, have been used to construct a number of geostatistical simulations of potential contamination for parcels approximately the size of a selective remediation unit (the 3-m width of a bulldozer blade). Each such simulation accurately reflects the actual measured sample values, and reproduces the univariate statistics and spatial character of the extant data. Post-processing of a large number of these equally likely statistically similar images produces maps directly showing the probability of exceeding specified levels of contamination (potential clean-up or personnel-hazard thresholds) 14. Biophysical calculations of cell killing probability by the amorphous track structure model for heavy-ion beams International Nuclear Information System (INIS) Kase, Yuki; Matsufuji, Naruhiro; Furusawa, Yoshiya; Kanai, Tatsuaki 2007-01-01 In a treatment planning of heavy-ion radiotherapy, it is necessary to estimate the biological effect of the heavy-ion beams. Physical dose should be associated with the relative biological effectiveness (RBE) at each point. Presently, carbon ion radiotherapy has been carried out at the National Institute Radiological Sciences (NIRS) in Japan and the Gesellschaft fuer Schwerionenforschung mbH (GSI) in Germany. Both facilities take individual approach for the calculation of the RBE value. At NIRS, the classical LQ model has been used while the local effect model (LEM) has been incorporated into the treatment planning system at GSI. The first aim of this study is to explain the RBE model of NIRS by the microdosimetric kinetic model (MKM). In addition, the clarification of similarities and differences between the MKM and the LEM was also investigated. (author) 15. The development and validation of a clinical prediction model to determine the probability of MODY in patients with young-onset diabetes. Science.gov (United States) Shields, B M; McDonald, T J; Ellard, S; Campbell, M J; Hyde, C; Hattersley, A T 2012-05-01 Diagnosing MODY is difficult. To date, selection for molecular genetic testing for MODY has used discrete cut-offs of limited clinical characteristics with varying sensitivity and specificity. We aimed to use multiple, weighted, clinical criteria to determine an individual's probability of having MODY, as a crucial tool for rational genetic testing. We developed prediction models using logistic regression on data from 1,191 patients with MODY (n = 594), type 1 diabetes (n = 278) and type 2 diabetes (n = 319). Model performance was assessed by receiver operating characteristic (ROC) curves, cross-validation and validation in a further 350 patients. The models defined an overall probability of MODY using a weighted combination of the most discriminative characteristics. For MODY, compared with type 1 diabetes, these were: lower HbA(1c), parent with diabetes, female sex and older age at diagnosis. MODY was discriminated from type 2 diabetes by: lower BMI, younger age at diagnosis, female sex, lower HbA(1c), parent with diabetes, and not being treated with oral hypoglycaemic agents or insulin. Both models showed excellent discrimination (c-statistic = 0.95 and 0.98, respectively), low rates of cross-validated misclassification (9.2% and 5.3%), and good performance on the external test dataset (c-statistic = 0.95 and 0.94). Using the optimal cut-offs, the probability models improved the sensitivity (91% vs 72%) and specificity (94% vs 91%) for identifying MODY compared with standard criteria of diagnosis MODY. This allows an improved and more rational approach to determine who should have molecular genetic testing. 16. Incorporating Anthropogenic Influences into Fire Probability Models: Effects of Human Activity and Climate Change on Fire Activity in California. Science.gov (United States) Mann, Michael L; Batllori, Enric; Moritz, Max A; Waller, Eric K; Berck, Peter; Flint, Alan L; Flint, Lorraine E; Dolfi, Emmalee 2016-01-01 The costly interactions between humans and wildfires throughout California demonstrate the need to understand the relationships between them, especially in the face of a changing climate and expanding human communities. Although a number of statistical and process-based wildfire models exist for California, there is enormous uncertainty about the location and number of future fires, with previously published estimates of increases ranging from nine to fifty-three percent by the end of the century. Our goal is to assess the role of climate and anthropogenic influences on the state's fire regimes from 1975 to 2050. We develop an empirical model that integrates estimates of biophysical indicators relevant to plant communities and anthropogenic influences at each forecast time step. Historically, we find that anthropogenic influences account for up to fifty percent of explanatory power in the model. We also find that the total area burned is likely to increase, with burned area expected to increase by 2.2 and 5.0 percent by 2050 under climatic bookends (PCM and GFDL climate models, respectively). Our two climate models show considerable agreement, but due to potential shifts in rainfall patterns, substantial uncertainty remains for the semiarid inland deserts and coastal areas of the south. Given the strength of human-related variables in some regions, however, it is clear that comprehensive projections of future fire activity should include both anthropogenic and biophysical influences. Previous findings of substantially increased numbers of fires and burned area for California may be tied to omitted variable bias from the exclusion of human influences. The omission of anthropogenic variables in our model would overstate the importance of climatic ones by at least 24%. As such, the failure to include anthropogenic effects in many models likely overstates the response of wildfire to climatic change. 17. Incorporating Anthropogenic Influences into Fire Probability Models: Effects of Human Activity and Climate Change on Fire Activity in California. Directory of Open Access Journals (Sweden) Michael L Mann Full Text Available The costly interactions between humans and wildfires throughout California demonstrate the need to understand the relationships between them, especially in the face of a changing climate and expanding human communities. Although a number of statistical and process-based wildfire models exist for California, there is enormous uncertainty about the location and number of future fires, with previously published estimates of increases ranging from nine to fifty-three percent by the end of the century. Our goal is to assess the role of climate and anthropogenic influences on the state's fire regimes from 1975 to 2050. We develop an empirical model that integrates estimates of biophysical indicators relevant to plant communities and anthropogenic influences at each forecast time step. Historically, we find that anthropogenic influences account for up to fifty percent of explanatory power in the model. We also find that the total area burned is likely to increase, with burned area expected to increase by 2.2 and 5.0 percent by 2050 under climatic bookends (PCM and GFDL climate models, respectively. Our two climate models show considerable agreement, but due to potential shifts in rainfall patterns, substantial uncertainty remains for the semiarid inland deserts and coastal areas of the south. Given the strength of human-related variables in some regions, however, it is clear that comprehensive projections of future fire activity should include both anthropogenic and biophysical influences. Previous findings of substantially increased numbers of fires and burned area for California may be tied to omitted variable bias from the exclusion of human influences. The omission of anthropogenic variables in our model would overstate the importance of climatic ones by at least 24%. As such, the failure to include anthropogenic effects in many models likely overstates the response of wildfire to climatic change. 18. Development and application of surrogate model for assessment of ex-vessel debris bed dryout probability - 15157 International Nuclear Information System (INIS) Yakush, S.E.; Lubchenko, N.T.; Kudinov, P. 2015-01-01 In this work we consider a water-cooled power reactor severe accident scenario with pressure vessel failure and subsequent release of molten corium. A surrogate model for prediction of dryout heat flux for ex-vessels debris beds of different shapes is developed. Functional form of dryout heat flux dependence on problem parameters is developed by the analysis of coolability problem in non-dimensional variables. It is shown that for a flat debris bed the dryout heat flux can be represented in terms of three 1-dimensional functions for which approximating formulas are found. For two-dimensional debris beds (cylindrical, conical, Gaussian heap, mound-shaped), an additional function taking into account the bed shape geometry is obtained from numerical simulations using DECOSIM code as a full model. With the surrogate model in hand, risk analysis of debris bed coolability is carried out by Monte Carlo sampling of the input parameters within selected ranges, with assumed distribution functions 19. Early bowel complications following radiotherapy of uterine cervix carcinoma International Nuclear Information System (INIS) Kim, Won Dong 1999-01-01 20. Philosophical theories of probability CERN Document Server Gillies, Donald 2000-01-01 The Twentieth Century has seen a dramatic rise in the use of probability and statistics in almost all fields of research. This has stimulated many new philosophical ideas on probability. Philosophical Theories of Probability is the first book to present a clear, comprehensive and systematic account of these various theories and to explain how they relate to one another. Gillies also offers a distinctive version of the propensity theory of probability, and the intersubjective interpretation, which develops the subjective theory. 1. On complicity theory. Science.gov (United States) Kline, A David 2006-04-01 The received account of whistleblowing, developed over the last quarter century, is identified with the work of Norman Bowie and Richard DeGeorge. Michael Davis has detailed three anomalies for the received view: the paradoxes of burden, missing harm and failure. In addition, he has proposed an alternative account of whistleblowing, viz., the Complicity Theory. This paper examines the Complicity Theory. The supposed anomalies rest on misunderstandings of the received view or misreadings of model cases of whistleblowing, for example, the Challenger disaster and the Ford Pinto. Nevertheless, the Complicity Theory is important for as in science the contrast with alternative competing accounts often helps us better understand the received view. Several aspects of the received view are reviewed and strengthened through comparison with Complicity Theory, including why whistleblowing needs moral justification. Complicity Theory is also critiqued. The fundamental failure of Complicity Theory is its failure to explain why government and the public encourage and protect whistleblowers despite the possibility of considerable harm to the relevant company in reputation, lost jobs, and lost shareholder value. 2. Non-Archimedean Probability NARCIS (Netherlands) Benci, Vieri; Horsten, Leon; Wenmackers, Sylvia We propose an alternative approach to probability theory closely related to the framework of numerosity theory: non-Archimedean probability (NAP). In our approach, unlike in classical probability theory, all subsets of an infinite sample space are measurable and only the empty set gets assigned 3. Modeling the Deep Impact Near-nucleus Observations of H2O and CO2 in Comet 9P/Tempel 1 Using Asymmetric Spherical Coupled Escape Probability Science.gov (United States) Gersch, Alan M.; A’Hearn, Michael F.; Feaga, Lori M. 2018-04-01 We have applied our asymmetric spherical adaptation of Coupled Escape Probability to the modeling of optically thick cometary comae. Expanding on our previously published work, here we present models including asymmetric comae. Near-nucleus observations from the Deep Impact mission have been modeled, including observed coma morphology features. We present results for two primary volatile species of interest, H2O and CO2, for comet 9P/Tempel 1. Production rates calculated using our best-fit models are notably greater than those derived from the Deep Impact data based on the assumption of optically thin conditions, both for H2O and CO2 but more so for CO2, and fall between the Deep Impact values and the global pre-impact production rates measured at other observatories and published by Schleicher et al. (2006), Mumma et al. (2005), and Mäkinen et al. (2007). 4. Joint probabilities and quantum cognition International Nuclear Information System (INIS) Acacio de Barros, J. 2012-01-01 In this paper we discuss the existence of joint probability distributions for quantumlike response computations in the brain. We do so by focusing on a contextual neural-oscillator model shown to reproduce the main features of behavioral stimulus-response theory. We then exhibit a simple example of contextual random variables not having a joint probability distribution, and describe how such variables can be obtained from neural oscillators, but not from a quantum observable algebra. 5. Joint probabilities and quantum cognition Energy Technology Data Exchange (ETDEWEB) Acacio de Barros, J. [Liberal Studies, 1600 Holloway Ave., San Francisco State University, San Francisco, CA 94132 (United States) 2012-12-18 In this paper we discuss the existence of joint probability distributions for quantumlike response computations in the brain. We do so by focusing on a contextual neural-oscillator model shown to reproduce the main features of behavioral stimulus-response theory. We then exhibit a simple example of contextual random variables not having a joint probability distribution, and describe how such variables can be obtained from neural oscillators, but not from a quantum observable algebra. 6. Incorporating anthropogenic influences into fire probability models: Effects of development and climate change on fire activity in California Science.gov (United States) Mann, M.; Moritz, M.; Batllori, E.; Waller, E.; Krawchuk, M.; Berck, P. 2014-12-01 The costly interactions between humans and natural fire regimes throughout California demonstrate the need to understand the uncertainties surrounding wildfire, especially in the face of a changing climate and expanding human communities. Although a number of statistical and process-based wildfire models exist for California, there is enormous uncertainty about the location and number of future fires. Models estimate an increase in fire occurrence between nine and fifty-three percent by the end of the century. Our goal is to assess the role of uncertainty in climate and anthropogenic influences on the state's fire regime from 2000-2050. We develop an empirical model that integrates novel information about the distribution and characteristics of future plant communities without assuming a particular distribution, and improve on previous efforts by integrating dynamic estimates of population density at each forecast time step. Historically, we find that anthropogenic influences account for up to fifty percent of the total fire count, and that further housing development will incite or suppress additional fires according to their intensity. We also find that the total area burned is likely to increase but at a slower than historical rate. Previous findings of substantially increased numbers of fires may be tied to the assumption of static fuel loadings, and the use of proxy variables not relevant to plant community distributions. We also find considerable agreement between GFDL and PCM model A2 runs, with decreasing fire counts expected only in areas of coastal influence below San Francisco and above Los Angeles. Due to potential shifts in rainfall patterns, substantial uncertainty remains for the semiarid deserts of the inland south. The broad shifts of wildfire between California's climatic regions forecast in this study point to dramatic shifts in the pressures plant and human communities will face by midcentury. The information provided by this study reduces the 7. Linear positivity and virtual probability International Nuclear Information System (INIS) Hartle, James B. 2004-01-01 We investigate the quantum theory of closed systems based on the linear positivity decoherence condition of Goldstein and Page. The objective of any quantum theory of a closed system, most generally the universe, is the prediction of probabilities for the individual members of sets of alternative coarse-grained histories of the system. Quantum interference between members of a set of alternative histories is an obstacle to assigning probabilities that are consistent with the rules of probability theory. A quantum theory of closed systems therefore requires two elements: (1) a condition specifying which sets of histories may be assigned probabilities and (2) a rule for those probabilities. The linear positivity condition of Goldstein and Page is the weakest of the general conditions proposed so far. Its general properties relating to exact probability sum rules, time neutrality, and conservation laws are explored. Its inconsistency with the usual notion of independent subsystems in quantum mechanics is reviewed. Its relation to the stronger condition of medium decoherence necessary for classicality is discussed. The linear positivity of histories in a number of simple model systems is investigated with the aim of exhibiting linearly positive sets of histories that are not decoherent. The utility of extending the notion of probability to include values outside the range of 0-1 is described. Alternatives with such virtual probabilities cannot be measured or recorded, but can be used in the intermediate steps of calculations of real probabilities. Extended probabilities give a simple and general way of formulating quantum theory. The various decoherence conditions are compared in terms of their utility for characterizing classicality and the role they might play in further generalizations of quantum mechanics 8. Prediction uncertainty assessment of a systems biology model requires a sample of the full probability distribution of its parameters Directory of Open Access Journals (Sweden) Simon van Mourik 2014-06-01 Full Text Available Multi-parameter models in systems biology are typically ‘sloppy’: some parameters or combinations of parameters may be hard to estimate from data, whereas others are not. One might expect that parameter uncertainty automatically leads to uncertain predictions, but this is not the case. We illustrate this by showing that the prediction uncertainty of each of six sloppy models varies enormously among different predictions. Statistical approximations of parameter uncertainty may lead to dramatic errors in prediction uncertainty estimation. We argue that prediction uncertainty assessment must therefore be performed on a per-prediction basis using a full computational uncertainty analysis. In practice this is feasible by providing a model with a sample or ensemble representing the distribution of its parameters. Within a Bayesian framework, such a sample may be generated by a Markov Chain Monte Carlo (MCMC algorithm that infers the parameter distribution based on experimental data. Matlab code for generating the sample (with the Differential Evolution Markov Chain sampler and the subsequent uncertainty analysis using such a sample, is supplied as Supplemental Information. 9. Survival probability of Baltic larval cod in relation to spatial overlap patterns with their prey obtained from drift model studies DEFF Research Database (Denmark) Hinrichsen, H.H.; Schmidt, J.O.; Petereit, C. 2005-01-01 Temporal mismatch between the occurrence of larvae and their prey potentially affects the spatial overlap and thus the contact rates between predator and prey. This might have important consequences for growth and survival. We performed a case study investigating the influence of circulation......-prey overlap, dependent on the hatching time of cod larvae. By performing model runs for the years 1979-1998 investigated the intra- and interannual variability of potential spatial overlap between predator and prey. Assuming uniform prey distributions, we generally found the overlap to have decreased since... 10. Unusual sugar specificity of banana lectin from Musa paradisiaca and its probable evolutionary origin. Crystallographic and modelling studies. Science.gov (United States) Singh, D D; Saikrishnan, K; Kumar, Prashant; Surolia, A; Sekar, K; Vijayan, M 2005-10-01 The crystal structure of a complex of methyl-alpha-D-mannoside with banana lectin from Musa paradisiaca reveals two primary binding sites in the lectin, unlike in other lectins with beta-prism I fold which essentially consists of three Greek key motifs. It has been suggested that the fold evolved through successive gene duplication and fusion of an ancestral Greek key motif. In other lectins, all from dicots, the primary binding site exists on one of the three motifs in the three-fold symmetric molecule. Banana is a monocot, and the three motifs have not diverged enough to obliterate sequence similarity among them. Two Greek key motifs in it carry one primary binding site each. A common secondary binding site exists on the third Greek key. Modelling shows that both the primary sites can support 1-2, 1-3, and 1-6 linked mannosides with the second residue interacting in each case primarily with the secondary binding site. Modelling also readily leads to a bound branched mannopentose with the nonreducing ends of the two branches anchored at the two primary binding sites, providing a structural explanation for the lectin's specificity for branched alpha-mannans. A comparison of the dimeric banana lectin with other beta-prism I fold lectins, provides interesting insights into the variability in their quaternary structure. 11. Probabilities and energies to obtain the counting efficiency of electron-capture nuclides. KLMN model; Probabilidades y energias de reestructuracion atomica subsiguientes a la captura electronica. Modelo KLMN Energy Technology Data Exchange (ETDEWEB) Galiano, G.; Grau, A. 1994-07-01 An intelligent computer program has been developed to obtain the mathematical formulae to compute the probabilities and reduced energies of the different atomic rearrangement pathways following electron-capture decay. Creation and annihilation operators for Auger and X processes have been introduced. Taking into account the symmetries associated with each process, 262 different pathways were obtained. This model allows us to obtain the influence of the M-electro capture in the counting efficiency when the atomic number of the nuclide is high. (Author) 12. Application of the huff model of shopping probability in the selected stores in Prešov (Prešov, the Slovak Republic Directory of Open Access Journals (Sweden) Mitríková Jana 2015-01-01 Full Text Available The main objective of this article is the calculation of the Huff Model and the comparison of obtained data with the questionnaire survey results. Afterwards, we will try to assess the validity of this theoretical model of shopping probability for practical use as well as its overall objectivity. Calculation of the Huff Model of Probability will be calculated for nine large-scale retail stores in fifteen local regions within the area of the third largest town in the Slovak Republic - Prešov. These results will be compared to results of the questionnaire research which was conducted during the period of November 2013 until March 2014 on a sample of 1,096 respondents. The questionnaire was addressing customers' preference for various large-scale stores as well as their place of residence. Based on the calculation results and the results obtained from the questionnaire survey, cartographic outputs were created, pointing out overall consistency as well as differences in the already mentioned results. 13. A scan statistic for binary outcome based on hypergeometric probability model, with an application to detecting spatial clusters of Japanese encephalitis. Science.gov (United States) Zhao, Xing; Zhou, Xiao-Hua; Feng, Zijian; Guo, Pengfei; He, Hongyan; Zhang, Tao; Duan, Lei; Li, Xiaosong 2013-01-01 As a useful tool for geographical cluster detection of events, the spatial scan statistic is widely applied in many fields and plays an increasingly important role. The classic version of the spatial scan statistic for the binary outcome is developed by Kulldorff, based on the Bernoulli or the Poisson probability model. In this paper, we apply the Hypergeometric probability model to construct the likelihood function under the null hypothesis. Compared with existing methods, the likelihood function under the null hypothesis is an alternative and indirect method to identify the potential cluster, and the test statistic is the extreme value of the likelihood function. Similar with Kulldorff's methods, we adopt Monte Carlo test for the test of significance. Both methods are applied for detecting spatial clusters of Japanese encephalitis in Sichuan province, China, in 2009, and the detected clusters are identical. Through a simulation to independent benchmark data, it is indicated that the test statistic based on the Hypergeometric model outweighs Kulldorff's statistics for clusters of high population density or large size; otherwise Kulldorff's statistics are superior. 14. Modelling effective diagnosis of risk complications in gestational diabetes mellitus: an e-diabetic expert system for pregnant women Science.gov (United States) Sreedevi, E.; Vijaya Lakshmi, K.; Chaitanya Krishna, E.; Padmavathamma, M. 2012-04-01 Diabetes is a chronic illness that requires continuous medical care and patient self-management education to prevent acute complications and to reduce the risk of long-term complications. This paper deals with study and development of algorithm to develop an initial stage expert system to provide diagnosis to the pregnant women who are suffering from Gestational Diabetes Mellitus (GDM) by means of Oral Glucose Tolerance Test (OGTT). 15. Optimized O'Neill/Glaser Model for Human Population of Space and its Impact on Survival Probabilities
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7134274840354919, "perplexity": 3250.727003340259}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-43/segments/1539583512421.5/warc/CC-MAIN-20181019170918-20181019192418-00536.warc.gz"}
https://www.vedantu.com/question-answer/if-aniline-is-treated-with-conc-h2so4-and-heated-class-12-chemistry-cbse-5fd6f2d7aef1aa270cdd876b
If aniline is treated with conc. ${H_2}S{O_4}$ and heated at ${200^\circ }C$, what would be the product?A) Anilinium sulphateB) Benzenesulfonic acidC) m-aminobenzenesulfonic acidD) sulphanilic acid Verified 119.1k+ views Hint:As we know that aromatic amines undergo typical electrophilic substitution reactions of aromatic ring and the amine group present on the ring is electron releasing and strongly ring activating and thereby increasing the electron density on ortho and para positions. We also know that aniline undergoes sulfonation reaction where it reacts with concentrated sulphuric acid resulting in the formation of anilinium hydrogen sulphate as an intermediate which on further heating with sulphuric acid at a temperature of about $453 - 473K$ or simply at ${200^\circ }C$, produces a para-aminobenzene sulfonic acid which is commonly called as sulphanilic acid as the major product where $- S{O_3}H$ group gets attached at the para position of the benzene ring. We can show this equation with the help of a chemical reaction as given below: Note:Always remember that $- N{H_2}$ group present on the benzene ring is electron releasing and thus ortho (-o) and para (-p) directing towards electrophilic aromatic substitution reaction.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6399272084236145, "perplexity": 5078.823422961064}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320305423.58/warc/CC-MAIN-20220128074016-20220128104016-00576.warc.gz"}
https://www.llvm.org/doxygen/classllvm_1_1Error.html
LLVM  14.0.0git llvm::Error Class Reference Lightweight error class with error context and mandatory checking. More... #include "llvm/Support/Error.h" Inheritance diagram for llvm::Error: [legend] ## Public Member Functions Error (const Error &Other)=delete Error (Error &&Other) Move-construct an error value. More... Create an error value. More... Erroroperator= (const Error &Other)=delete Erroroperator= (Error &&Other) Move-assign an error value. More... ~Error () Destroy a Error. More... operator bool () Bool conversion. More... template<typename ErrT > bool isA () const Check whether one error is a subclass of another. More... const void * dynamicClassID () const Returns the dynamic class id of this error, or null if this is a success value. More... ## Static Public Member Functions static ErrorSuccess success () Create a success value. More... ## Protected Member Functions Error () Create a success value. Prefer using 'Error::success()' for readability. More... ## Friends class ErrorList template<typename T > class Expected template<typename... HandlerTs> Error handleErrors (Error E, HandlerTs &&... Handlers) Pass the ErrorInfo(s) contained in E to their respective handlers. More... LLVMErrorRef wrap (Error) Conversion from Error to LLVMErrorRef for C error bindings. More... raw_ostreamoperator<< (raw_ostream &OS, const Error &E) ## Detailed Description Lightweight error class with error context and mandatory checking. Instances of this class wrap a ErrorInfoBase pointer. Failure states are represented by setting the pointer to a ErrorInfoBase subclass instance containing information describing the failure. Success is represented by a null pointer value. Instances of Error also contains a 'Checked' flag, which must be set before the destructor is called, otherwise the destructor will trigger a runtime error. This enforces at runtime the requirement that all Error instances be checked or returned to the caller. There are two ways to set the checked flag, depending on what state the Error instance is in. For Error instances indicating success, it is sufficient to invoke the boolean conversion operator. E.g.: Error foo(<...>); if (auto E = foo(<...>)) return E; // <- Return E if it is in the error state. // We have verified that E was in the success state. It can now be safely // destroyed. A success value can not be dropped. For example, just calling 'foo(<...>)' without testing the return value will raise a runtime error, even if foo returns success. For Error instances representing failure, you must use either the handleErrors or handleAllErrors function with a typed handler. E.g.: class MyErrorInfo : public ErrorInfo<MyErrorInfo> { // Custom error info. }; Error foo(<...>) { return make_error<MyErrorInfo>(...); } auto E = foo(<...>); // <- foo returns failure with MyErrorInfo. auto NewE = [](const MyErrorInfo &M) { // Deal with the error. }, [](std::unique_ptr<OtherError> M) -> Error { if (canHandle(*M)) { // handle error. return Error::success(); } // Couldn't handle this error instance. Pass it up the stack. return Error(std::move(M)); ); // Note - we must check or return NewE in case any of the handlers // returned a new error. The handleAllErrors function is identical to handleErrors, except that it has a void return type, and requires all errors to be handled and no new errors be returned. It prevents errors (assuming they can all be handled) from having to be bubbled all the way to the top-level. All Error instances must be checked before destruction, even if they're moved-assigned or constructed from Success values that have already been checked. This enforces checking through all levels of the call stack. Definition at line 157 of file Error.h. ## ◆ Error() [1/4] llvm::Error::Error ( ) inlineprotected Create a success value. Prefer using 'Error::success()' for readability. Definition at line 176 of file Error.h. ## ◆ Error() [2/4] llvm::Error::Error ( const Error & Other ) delete ## ◆ Error() [3/4] llvm::Error::Error ( Error && Other ) inline Move-construct an error value. The newly constructed error is considered unchecked, even if the source error had been checked. The original error becomes a checked Success value, regardless of its original state. Definition at line 191 of file Error.h. References move, and Other. ## ◆ Error() [4/4] llvm::Error::Error ( std::unique_ptr< ErrorInfoBase > Payload ) inline Create an error value. Prefer using the 'make_error' function, but this constructor can be useful when "re-throwing" errors from handlers. Definition at line 198 of file Error.h. ## ◆ ~Error() llvm::Error::~Error ( ) inline Destroy a Error. Fails with a call to abort() if the error is unchecked. Definition at line 227 of file Error.h. References getPtr(). Referenced by llvm::JITSymbol::~JITSymbol(). ## ◆ dynamicClassID() const void* llvm::Error::dynamicClassID ( ) const inline Returns the dynamic class id of this error, or null if this is a success value. Definition at line 247 of file Error.h. References getPtr(). ## ◆ isA() template<typename ErrT > bool llvm::Error::isA ( ) const inline Check whether one error is a subclass of another. Definition at line 241 of file Error.h. References getPtr(). ## ◆ operator bool() llvm::Error::operator bool ( ) inlineexplicit Bool conversion. Returns true if this Error is in a failure state, and false if it is in an accept state. If the error is in a Success state it will be considered checked. Definition at line 235 of file Error.h. References getPtr(). ## ◆ operator=() [1/2] Error& llvm::Error::operator= ( const Error & Other ) delete ## ◆ operator=() [2/2] Error& llvm::Error::operator= ( Error && Other ) inline Move-assign an error value. The current error must represent success, you you cannot overwrite an unhandled error. The current error is then considered unchecked. The source error becomes a checked success value, regardless of its original state. Definition at line 210 of file Error.h. References Other. ## ◆ ErrorList friend class ErrorList friend Definition at line 161 of file Error.h. ## ◆ Expected template<typename T > friend class Expected friend Definition at line 169 of file Error.h. ## ◆ handleErrors template<typename... HandlerTs> Error handleErrors ( Error E, HandlerTs &&... Handlers ) friend Pass the ErrorInfo(s) contained in E to their respective handlers. Any unhandled errors (or Errors returned by handlers) are re-concatenated and returned. Because this function returns an error, its result must also be checked or returned. If you intend to handle all errors use handleAllErrors (which returns void, and will abort() on unhandled errors) instead. Definition at line 935 of file Error.h. ## ◆ operator<< raw_ostream& operator<< ( raw_ostream & OS, const Error & E ) friend Definition at line 315 of file Error.h. ## ◆ wrap LLVMErrorRef wrap ( Error Err ) friend Conversion from Error to LLVMErrorRef for C error bindings. Definition at line 1381 of file Error.h. The documentation for this class was generated from the following file: M We currently emits eax Perhaps this is what we really should generate is Is imull three or four cycles eax eax The current instruction priority is based on pattern complexity The former is more complex because it folds a load so the latter will not be emitted Perhaps we should use AddedComplexity to give LEA32r a higher priority We should always try to match LEA first since the LEA matching code does some estimate to determine whether the match is profitable if we care more about code then imull is better It s two bytes shorter than movl leal On a Pentium M llvm::Error::handleErrors friend Error handleErrors(Error E, HandlerTs &&... Handlers) Pass the ErrorInfo(s) contained in E to their respective handlers. Definition: Error.h:935 llvm::Error::success static ErrorSuccess success() Create a success value. Definition: Error.h:331 E static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC") foo < i32 > tmp foo
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.18009833991527557, "perplexity": 12511.561088507347}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046154798.45/warc/CC-MAIN-20210804080449-20210804110449-00126.warc.gz"}
http://nrich.maths.org/1026/note?nomenu=1
## Pumpkin Pie Problem Peter, the pumpkin eater, wanted to make two pies for a party. His mother, a professional pie maker, had a recipe for him to use. However, she always made $80$ pies at a time. She used: $10$ dozen eggs $27$ litres of condensed milk $480$ tablespoons of sugar $100$ teaspoons of cinnamon $140$ cups of pumpkin Peter looked in the cupboard and found: $4$ cups of pumpkin $2$ eggs $1 \frac{1}{2}$ teaspoons of cinnamon $\frac{2}{3}$ of a litre of condensed milk $15$ tablespoons of sugar Did Peter have enough ingredients to make two pumpkin pies for the party or did he need to buy more? ### Why do this problem? Ratio is a notoriously difficult topic and many teachers avoid it! There are many possible ways of solving this problem which make it ideal for class or group discussion, and offer opportunities to assess your children's understanding and misconceptions. You will need to allow time for a reasonable discussion. ### Possible approach Working in pairs gives the children opportunities to clarify their thinking. Give each pair a large piece of paper on which to record - many may want to draw pictures and working on small pieces of paper isn't helpful. Having explained the problem, provide each pair with the two lists (you can download a printable copy here). Give a little time for the children to 'get into' the problem. When appropriate, draw the group or class together and ask what they have found useful to begin. Give more time for them to use some of these ideas, or continue with their own. As each pair comes to a solution, ask them to prepare a new piece of paper which will help them to explain their working. Choose a confident pair to start and then invite each pair in turn to explain what they have done. The rest of the group can ask questions if clarification is needed. Focus the children's attention on the different ways pairs have solved the problem but how all involve scaling amounts up and down - ratio and proportion. ### Key questions Can you paraphrase the question (ask it in a different way) to make it easier to understand? What do we need to find out? ### Possible extension Most children will have focused on the eggs first as these are easier to calculate than litres, teaspoons etc. The question is then answered as no matter how much of everything else he has, he doesn't have enough eggs. Children who find this quite easy could be challenged to find out what additional other ingredients would have to be bought for two pies. What is the largest number of pies he could make without buying more of each ingredient? ### Possible support This is probably not a suitable question for children who struggle with the idea of doubling, halving and other simple ratio.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.32089775800704956, "perplexity": 1361.6465590029443}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1409535919886.18/warc/CC-MAIN-20140909051335-00256-ip-10-180-136-8.ec2.internal.warc.gz"}
http://clay6.com/qa/32527/ch-equiv-ch-x-2ch-3cooh-y-rightarrow-a-quad-underrightarrow-quad-b-c-b-can-
Browse Questions # $CH\equiv CH(X)+2CH_3COOH(Y)\rightarrow A\quad \underrightarrow{\Delta}\quad B+C$.B can be obtained from X by hydration and c can be obtained from Y by heating with $P_2O_5$.Hence A is Hence (b) is the correct answer.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8087372183799744, "perplexity": 1011.7361790294702}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128323870.46/warc/CC-MAIN-20170629051817-20170629071817-00505.warc.gz"}
http://math.stackexchange.com/questions/180949/help-with-understanding-definition-of-definite-integral
# Help with understanding definition of Definite Integral I am having some difficulty in understanding notation/meaning in the definition of definite integral. Can you guys help clarify/correct and fill in the gaps in my understanding. Following is from wikipedia entry for the Darboux integral which is similar to what is in my textbook. My notes are inline. A partition of an interval $[a,b]$ is a finite sequence of values $x_i$ such that, $a = x_0 < x_1 < ... < x_n = b$ Partition to integrate over Each interval $[x_i−1,x_i]$ is called a subinterval of the partition. Let $ƒ:[a,b]→R$ be a bounded function, and let $P = (x_0, ..., x_n)$ be a partition of $[a,b]$. Let Defining rectangles that would be used to approximate area under curve $M_i = \sup\limits_{x \in [x_{i-1}, x_i]} f(x)$ $m_i = \inf\limits_{x \in [x_{i-1}, x_i]} f(x)$ Height of rectangle when above the curve or below the curve respectively The upper Darboux sum of ƒ with respect to P is $U_{f,P} = \sum_{i=1}^n (x_i - x_{i-1})M_i$ sum when using supremum for height The lower Darboux sum of ƒ with respect to P is $L_{f,P} = \sum_{i=1}^n (x_i - x_{i-1})m_i$ sum when using infimum for height The upper Darboux integral of ƒ is $U_f = \inf \{U_{f,P} : \mbox{P is a partition of} [a,b]\}$ The lower Darboux integral of ƒ is $L_f = \sup \{L_{f,P} : \mbox{P is a partition of} [a,b]\}$ This is the part that I am unable to understand. What do the infimum and supremum imply here? I understand the definite integral as area under curve. Wouldn't the sums $U_{f,P}$ and $L_{f,P}$ then be the integrals themselves? Why do we need to take the infimum/supremum of these areas? If $U_ƒ = L_ƒ$, then we say that $ƒ$ is Darboux-integrable and set $\int_a^b f(t)dt = U_f = L_f$ this parts seems to intuitively suggest both areas would be nearly equal - The upper and lower Darboux sums only approximate the area under the curve $(x, f(x))$. There is no single partition $P$ of $[a, b]$ such that $U_{f,P}$ is the exact area, but the approximation keeps getting better as $P$ gets finer. An equivalent way to say that $f$ is integrable is for every $\epsilon > 0$, there is a partition $P$ of $[a, b]$ such that $U_{f, P} - L_{f, P} < \epsilon$. – Ink Aug 10 '12 at 7:32 Thank you, I understand now. – mathguy80 Aug 10 '12 at 7:39 We take infimum and supermum with respect to partitions. Note that $U_{f,P}$ and $L_{f,P}$ depend on $P$. For each $P$, we know that $L_{f,P} \le U_{f,P}$. As $P$ gets finer, $U_{f,P}$ may decrease and $L_{f,P}$ may increase. If there exists a sequence of partition $P_1, P_2, \ldots$ that makes $U_{f,P_n}$ and $L_{f,P_n}$ converge to the same number, that number is taken as the integral. We require that the sequence of partitions must be "increasing" in the sense that all grid points in $P_n$ must be in $P_{n+1}$. However, there are infinitely many choices of increasing sequences of partitions. If we find a sequence that "doesn't work", that doesn't mean there doesn't exist a sequence that works. To take into account ALL partitions at once, $\sup$ and $\inf$ are used. Note that not only we know that $L_{f,P} \le U_{f,P}$, but we actually know that $L_{f,P} \le U_{f,Q}$ for any partitions $P$ and $Q$. That means every number in the set $\{L_{f,P}|P\text{ is a partition}\}$ is smaller than or equal to every number in the set $\{U_{f,P}|P\text{ is a partition}\}$. If the two set has a common accumulation point, then we can create a sequence of partitions $P_1, P_2, \ldots$ such that $U_{f,P_n}$ and $L_{f,P_n}$ converge to the same number. (Pick once sequence for $U_f$ and one sequence for $L_f$, then create a new partition sequence by taking the union of partitions from both sequences.) $U_{f,P}$ is the area of the upper rectangles and is, generally, bigger than the area under $f$; $L_{f,P}$ is the area of the lower rectangles and is, generally, less than the area under $f$. $U_f$ is the infimum of $U_{f,P}$ over all possible partitions, and $L_f$ is the supremum of $L_{f,P}$ over all possible partitions. If they are equal, then they must equal the integral.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9879967570304871, "perplexity": 153.41779427897097}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398461113.77/warc/CC-MAIN-20151124205421-00059-ip-10-71-132-137.ec2.internal.warc.gz"}
http://openstudy.com/updates/4ed41fb5e4b0bcd98c9f99ca
## Got Homework? ### Connect with other students for help. It's a free community. • across Online now • laura* Helped 1,000 students Online now • Hero College Math Guru Online now Here's the question you clicked on: 55 members online • 0 viewing ## Sariah Group Title The point P on the circle x^2 + y^2 = r^2 that is also on the terminal side of an angle \theta in standard position is given. Find the indicated trigonometric function. (-2, -1) Find \csc \theta . Seleccione una respuesta. a. -2 b. - \sqrt{5} c. -5 d. \sqrt{5} 2 years ago 2 years ago Edit Question Delete Cancel Submit • This Question is Closed • Attachments: ## See more questions >>> ##### spraguer (Moderator) 5→ View Detailed Profile 23 • Teamwork 19 Teammate • Problem Solving 19 Hero • You have blocked this person. • ✔ You're a fan Checking fan status... Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9548763036727905, "perplexity": 11306.47507871568}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500835822.36/warc/CC-MAIN-20140820021355-00201-ip-10-180-136-8.ec2.internal.warc.gz"}
https://topdrawer.aamt.edu.au/Reasoning/Assessment/Designing-tasks-for-assessing-reasoning/Levels-of-proof/Triangular-numbers
Triangular numbers A teacher asked year 9 students to consider the following statement: The sum of any two consecutive triangular numbers is a square number. The teacher asked, "Is this true, and if so, can you prove it?" The teacher observed and listened to students explaining their approaches as they worked and during a reporting back session. The teacher was able to identify various strategies and hence levels of reasoning. • Trying particular cases "Well, I thought of the triangular numbers, like 3, 6, 10. Three plus 6 equalled 9 and that is a square number. Six and 10 made 16, a square number. And I kept going like that. They all worked, so it is true." • Trying extreme cases "I tried a few examples and they worked, but that does not prove it is always true. I looked up bigger triangular numbers on the web and the ones I tried all worked, too. I tried –3 and –6 but that was +9. Maybe they do not all work, but I could not find any that didn't — they all made square numbers. So I reckon it’s true." • Proof, using modelling "I drew a triangular number, 6, like a right-angled triangle. Then I drew the next one, 10. I could see that the 6 and 10 would fit together like this, and make a square. Proof by modelling. "Then I could see that it would always work, because no matter how many lines you have got, you have also got the smaller number that is one row less. Like 21 has 1 + 2 + 3 + 4 + 5 + 6, and the next triangular number down (15) is the same without the 6, so it will make a square. And that would be the case even with bigger numbers. "Then I realised you could also draw it as an isosceles triangle and you could just put the smaller number underneath. I did not draw it but I can imagine it." • Proof, using algebra "First I developed a rule for the nth triangular number: multiply n by n + 1 then divide by 2, where n is the number of rows. Then I used it to write the rule for the (n + 1)th triangular number. I added these two rules together, then I used algebra to show that it is a square number. I found out something else, too: that the square number is the square of the difference between the two triangular numbers. I am working on a proof of that now." Curriculum links Year 6: Identify and describe properties of prime, composite, square and triangular numbers Year 6: Continue and create sequences involving whole numbers, fractions and decimals. Describe the rule used to create the sequence Year 10: Apply logical reasoning, including the use of congruence and similarity, to proofs and numerical exercises involving plane shapes Source
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8566007018089294, "perplexity": 413.0284342904314}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202781.83/warc/CC-MAIN-20190323101107-20190323123107-00171.warc.gz"}
http://mathhelpforum.com/algebra/18125-solved-what-s-original-value-percentage.html
# Thread: [SOLVED] What's the original value of this percentage 1. ## [SOLVED] What's the original value of this percentage Hi guys, I need to find the original value of X for a particular percentage of Y. Example If 65% of X = 100(being Y), what is X (for this example it's apx 155) but I need exact numbers. Cheers. 2. Originally Posted by kps_uk Hi guys, I need to find the original value of X for a particular percentage of Y. Example If 65% of X = 100(being Y), what is X (for this example it's apx 155) but I need exact numbers. Cheers. first we need some vocabulary (vocabulary in math ?! yes, vocabulary in math). "Percent" is a fancy word for "out of 100" or "divided by 100." so 65% means 65/100 in math, "of" means "multiply." so 65% of means "65/100 x" so now, rewrite the equation: we have: $\frac {65}{100}x = 100$ Now solve for $x$ 3. Originally Posted by kps_uk Hi guys, I need to find the original value of X for a particular percentage of Y. Example If 65% of X = 100(being Y), what is X (for this example it's apx 155) but I need exact numbers. Cheers. In general, $x$% of $y = z$ ( $x$% = $x$ per cent = $x$ per $100$ = $\frac{x}{100}$) $\frac{x}{100} * y = z$ $y = z*\frac{100}{x}$ Hope this helps Mahurshi Akilla
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 11, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8442296385765076, "perplexity": 3837.2370528798556}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-40/segments/1474738661289.57/warc/CC-MAIN-20160924173741-00042-ip-10-143-35-109.ec2.internal.warc.gz"}
https://tf2maps.net/threads/difficulty-generating-pop-files.32390/
# Difficulty Generating .pop Files Discussion in 'Mapping Questions & Discussion' started by Erk, Apr 9, 2017. 1. ### Server StaffErkerk Messages: 323 Positive Ratings: 798 I decided to start an mvm map for the contest, and I've copied the entities required from the mvm_example.vmf. However when I start the map up, the wave number stays at 0, even though I've copied the mvm_example.pop file and renamed it to my map name. I've used nav_generate, and nav_mark_walkable. I've got the pop file in C:\Program Files (x86)\Steam\steamapps\common\Team Fortress 2\tf\scripts\population folder. I've got no errors regarding the actual map, just the pop file. Any idea how to get this to work? Super fast edit: http://i.imgur.com/eXNQbC3.png I forgot to add that the info_player_teamspawns are above brushwork. 2. ### HydrogenMvM Maniac Messages: 84 Positive Ratings: 196 What does your console say when you load your map? (Error messages.) Try punching in nav_edit 1 and see if you can see your navigation mesh on the map? Also, I'm not sure if the example pop file is actually syntactically valid... try renaming mvm_coaltown.pop to your map name and try again? (You can find mvm_coaltown.pop from the tf2_misc_dir.vpk file.) I don't have any other ideas off the top of my head on why it doesn't work. Send me a PM with the files you tried and I'll experiment around with it to see if I can get it to load. • Thanks x 1 3. ### Server StaffErkerk Messages: 323 Positive Ratings: 798 Got this error message: "Can't open scripts/population/mvm_erkentry_a1.pop. Parse Failed in CPopulationManager::Initialize for scripts/population/mvm_erkentry_a1.pop" 4. ### HydrogenMvM Maniac Messages: 84 Positive Ratings: 196 Are you sure that the mvm_erkentry_a1.pop is located in a folder like this? C:\Program Files (x86)\Steam\steamapps\common\Team Fortress 2\tf\scripts\population\mvm_erkentry_a1.pop By default TF2 does not come with a "population" folder inside tf/scripts (there should be just "items" only). Make the "population" folder (make sure it's spelled correctly) in there and then put the mvm_erkentry_a1.pop file in it. EDIT: I just read the bottom of your post and you said you put it in there. Send me your map file and I'll toy with it. 5. ### Server StaffErkerk Messages: 323 Positive Ratings: 798 My population folder is called "Population folder". It should be just named "population", right? 6. ### HydrogenMvM Maniac Messages: 84 Positive Ratings: 196 Yes, just "population" only. • Thanks x 1 7. ### Server StaffErkerk Messages: 323 Positive Ratings: 798 Im an idiot. Edit: Thanks for the help! • Like x 1 • Funny x 1 Last edited: Apr 9, 2017 Tags:
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8192239999771118, "perplexity": 7836.804302984372}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376829140.81/warc/CC-MAIN-20181218102019-20181218124019-00095.warc.gz"}
https://handwiki.org/wiki/Fluxion
# Fluxion Short description: Historical mathematical concept; form of derivative Newton's introduction of the notions "fluent" and "fluxion" in his 1736 book A fluxion is the instantaneous rate of change, or gradient, of a fluent (a time-varying quantity, or function) at a given point.[1] Fluxions were introduced by Isaac Newton to describe his form of a time derivative (a derivative with respect to time). Newton introduced the concept in 1665 and detailed them in his mathematical treatise, Method of Fluxions.[2] Fluxions and fluents made up Newton's early calculus.[3] ## History Fluxions were central to the Leibniz–Newton calculus controversy, when Newton sent a letter to Gottfried Wilhelm Leibniz explaining them, but concealing his words in code due to his suspicion. He wrote:[4] I cannot proceed with the explanations of the fluxions now, I have preferred to conceal it thus: 6accdæ13eff7i3l9n4o4qrr4s8t12vx. The gibberish string was in fact a hash code (by denoting the frequency of each letter) of the Latin phrase Data æqvatione qvotcvnqve flventes qvantitates involvente, flvxiones invenire: et vice versa, meaning: "Given an equation that consists of any number of flowing quantities, to find the fluxions: and vice versa".[5] ## Example If the fluent $\displaystyle{ y }$ is defined as $\displaystyle{ y=t^2 }$ (where $\displaystyle{ t }$ is time) the fluxion (derivative) at $\displaystyle{ t=2 }$ is: $\displaystyle{ \dot y = \frac{\Delta y}{\Delta t} = \frac{(2+o)^2-2^2}{(2+o)-2} = \frac{4+4o+o^2-4}{2+o-2} = \frac{4o+o^2}{o} }$ Here $\displaystyle{ o }$ is an infinitely small amount of time.[6] So, the term $\displaystyle{ o^2 }$ is second order infinite small term and according to Newton, we can now ignore $\displaystyle{ o^2 }$ because of its second order infinite smallness comparing to first order infinite smallness of $\displaystyle{ o }$.[7] So, the final equation gets the form: $\displaystyle{ \dot y = \frac{\Delta y}{\Delta t}=\frac{4o}{o}=4 }$ He justified the use of $\displaystyle{ o }$ as a non-zero quantity by stating that fluxions were a consequence of movement by an object. ## Criticism Bishop George Berkeley, a prominent philosopher of the time, denounced Newton's fluxions in his essay The Analyst, published in 1734.[8] Berkeley refused to believe that they were accurate because of the use of the infinitesimal $\displaystyle{ o }$. He did not believe it could be ignored and pointed out that if it was zero, the consequence would be division by zero. Berkeley referred to them as "ghosts of departed quantities", a statement which unnerved mathematicians of the time and led to the eventual disuse of infinitesimals in calculus. Towards the end of his life Newton revised his interpretation of $\displaystyle{ o }$ as infinitely small, preferring to define it as approaching zero, using a similar definition to the concept of limit.[9] He believed this put fluxions back on safe ground. By this time, Leibniz's derivative (and his notation) had largely replaced Newton's fluxions and fluents, and remains in use today.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9128999710083008, "perplexity": 1965.5859291645047}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296948609.41/warc/CC-MAIN-20230327060940-20230327090940-00170.warc.gz"}
https://dkfzsearch.kobv.de/advancedSearch.do?f1=author&v1=LEE%2C+Y.+S.&index=internal&plv=2
An error occurred while sending the email. Please try again. Proceed reservation? Export • 1 Electronic Resource Springer Abdominal imaging 25 (2000), S. 255-258 ISSN: 1432-0509 Keywords: Key words: Hepatic veins, thrombosis—Liver neoplasms—Liver neoplasms, CT—Liver neoplasms, MR. Source: Springer Online Journal Archives 1860-2000 Topics: Medicine Notes: Abstract We report the imaging findings of spiral computed tomography (CT), magnetic resonance (MR) imaging, and MR angiography in a patient with nodular regenerative hyperplasia of the liver associated with Budd–Chiari syndrome. Spiral CT showed multiple enhancing nodules during the hepatic arterial and portal venous phases. MR images showed multiple hyperintense nodules on T1-weighted images and hypointense or isointense nodules on T2-weighted images. MR angiography showed thrombotic occlusion of three hepatic veins, suggesting Budd–Chiari syndrome. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 2 Electronic Resource Springer Abdominal imaging 24 (1999), S. 125-128 ISSN: 1432-0509 Keywords: Key words: Gastrointestinal edema—Edema cirrhosis—Colon edema—CT. Source: Springer Online Journal Archives 1860-2000 Topics: Medicine Notes: Abstract Background: To evaluate the computed tomographic (CT) features of colonic wall thickening in cirrhotic patients and to determine their prognostic value. Methods: We retrospectively reviewed 28 cirrhotic patients with colonic wall thickening (≥10 mm) on CT. Twenty-six of the 28 patients had hepatocellular carcinoma. The severity of hepatic dysfunction was determined by using the Child–Pugh classification. We analyzed the patterns of bowel wall thickening and degree of portal hypertension on CT and the survival periods after initial CT detection of colonic wall thickening. Results: The involved segment of the colon was diffusely thickened with either scalloped or nodular circumferential configuration. In all patients, the thickened colonic wall enhanced poorly. Although the ascending colon was involved in all patients, the transverse (n= 14) or descending (n= 5) colon was also simultaneously involved. Most patients exhibited an advanced stage of portal hypertension on CT. The median survival period of 25 patients who expired was 34 days, and 21 patients (84%) expired within 3 months. Conclusions: Colonic wall thickening on CT can be used as one of the indicators of poor prognosis in cirrhotic patients. Advanced liver cirrhosis with significantly severe dysfunction is the likely cause of mortality. Therefore, a less aggressive therapeutic approach is recommended if hepatocellular carcinoma is coexistent in these patients. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 3 Electronic Resource Oxford, UK : Blackwell Science Ltd Journal of oral rehabilitation 30 (2003), S. 0 ISSN: 1365-2842 Source: Blackwell Publishing Journal Backfiles 1879-2005 Topics: Medicine Notes: summary  The aim of this study was to subjectively determine the distribution of anaesthesia by mapping areas of sensory loss following inferior alveolar nerve block. Fifty healthy dental students were the subjects of this study (men 32, women 18). They were asked to draw the anaesthetized area on a diagram of the face and tongue 20 min after inferior alveolar nerve block. They evaluated the degree of anaesthesia by touching their faces and moving their tongues. All of the 50 subjects reported anaesthesia in the facial area. Of these, 21 (42%) reported the cutaneous distribution of anaesthesia on mental nerve territory only. Seventeen subjects (34%) reported anaesthesia on mental and buccal nerve territory. Nine subjects (18%) reported anaesthesia on mental, buccal, and auriculotemporal nerve territory. Two subjects (4%) reported anaesthesia on mental and auriculotemporal nerve territory and one subject (2%) on mental, buccal and infra-orbital nerve territory. Forty-seven of the 50 subjects (94%) reported anaesthesia of the tongue with the various degree of anaesthesia according to the area. Of these, 17 subjects (34%) reported strong anaesthesia on the anterior area and weak anaesthesia on the middle part of the tongue. Nineteen subjects (38%) reported strong anaesthesia of the lateral area and weak anaesthesia on the medial area, and 11 subjects (22%) reported anaesthesia on only the lateral side of the tongue. Three subjects (6%) reported no anaesthesia of the tongue. The distribution of anaesthesia of the facial and glossal regions determined subjectively after inferior alveolar nerve block, varies significantly between individuals. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 4 Electronic Resource Oxford, UK : Blackwell Publishing Ltd ISSN: 1365-2230 Source: Blackwell Publishing Journal Backfiles 1879-2005 Topics: Medicine Notes: Sera from 3528 patients with autoimmune disease, and non-autoimmune disease, and 500 normal individuals were studied For the presence of anticentromere antibodies (ACA) by indirect immunofluorescence on HEP-2 cells. Sixty-seven specimens were identified showing discrete speckled staining: 55 (82.1%), 11 (16.4%), and one (1.5%), were from patients with autoimmune disease, non-autoimmune disease and normal control subjects, respectively. These ACA were present frequently in CREST syndrome (55%), Raynaud's disease (29.6%) and primary biliary cirrhosis (30%). Only 16.4% of the antibody positive patients carried a clinical diagnosis of CREST, which means that ACA are not specific for CREST syndrome. High antibody titre persisted irrespective of whether or not the patients had active disease. The ACA were present infrequently in Sjögren's syndrome, systemic lupus erythematosus, rheumatoid arthritis, immune thrombocytopenic purpura, Graves' disease, immune haemolytic anaemia, and vitiligo. Sera from 107 patients with various other autoimmune diseases were negative for ACA. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 5 Electronic Resource Springer ISSN: 1588-2780 Source: Springer Online Journal Archives 1860-2000 Topics: Chemistry and Pharmacology , Energy, Environment Protection, Nuclear Power Engineering Notes: Abstract Radioaactive samples in cylindrical beakers have been measured using a high purity Ge spectrometer. Self-attenuation effects at various densities of the radioactive samples filled in cylindrical beakers were studied by experimenal and theoretical methods. Coincidence summing effects for specific nuclides were also determined from the measurements of full-energy peak efficiencies in a given source-to-detector geometry. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 6 Electronic Resource Springer Bioprocess engineering 21 (1999), S. 293-298 ISSN: 0178-515X Source: Springer Online Journal Archives 1860-2000 Topics: Process Engineering, Biotechnology, Nutrition Technology Notes: Abstract In order to increase the synthesis of bovine growth hormone (bGH) using T7 promoter system in E. coli, the artificial AT-rich block was introduced into the upstream region of a consensus Shine-Dalgarno (SD) sequence and the spacer region (between SD and ATG codon) was enriched with A and T nucleotides. The cells harboring pTAJ plasmids with AT-rich block produced bGH in the range of 3% to 25% and the cells harboring pTBJ plasmids with AT-rich sequence in the spacer region from 0.8% to 20% of total cell proteins. This result suggests that AT rich block and AT nucleotides in the spacer region destabilize mRNA secondary structure, depending on the downstream coding information of bGH gene and also, implying that the disruption of mRNA secondary structure might be a major factor for regulating bGH expression in the translational initiation process. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 7 Electronic Resource Springer Molecular genetics and genomics 262 (1999), S. 448-452 ISSN: 1617-4623 Keywords: Key words Replication initiation ; iciA ; Phosphate regulon ; PhoB ; Escherichia coli Source: Springer Online Journal Archives 1860-2000 Topics: Biology Notes: Abstract The IciA protein from Escherichia coli has been shown specifically to inhibit the in vitro initiation of chromosomal DNA replication. However, the in vivo role of IciA has not yet been established. In order to investigate the in vivo function of this protein, expression of the iciA gene was studied by monitoring the β-galactosidase activity specified by an iciA promoter-lacZ fusion inserted into the chromosome. Among the conditions tested (carbon starvation, the stringent response, phosphate starvation, and the SOS response), only phosphate depletion increased iciA expression. Supplementation of phosphate-depleted cultures with inorganic phosphate reduced the β-galactosidase activity to basal levels. Enhanced expression of iciA-lacZ was dependent upon the PhoB protein. PhoB is known to be a transcriptional activator of the Pho regulon, expression of which is activated during phosphate starvation. It was also found that the iciA promoter contains a PhoB protein-binding sequence, termed the Pho box, which is necessary for the activation of genes of the Pho regulon. These results suggest that the iciA gene is a member of the Pho regulon. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 8 Electronic Resource Springer Acta mechanica 49 (1983), S. 95-109 ISSN: 1619-6937 Source: Springer Online Journal Archives 1860-2000 Topics: Mechanical Engineering, Materials Science, Production Engineering, Mining and Metallurgy, Traffic Engineering, Precision Mechanics , Physics Notes: Summary A stress solution for a steady state creep material described by the power law is obtained by applying the pseudo-complex stress function which satisfies the biharmonic equation. The relation between the complex stress and the pseudo-stress function is established. The stress function or the stress components can be obtained from the known pseudo stress function. Two specific cases are investigated: 1) the stress field is presumed to contian σ x , σ y and σ xy and 2) the stress field contains only shearing stress, σ xz and σ yz . Both cases presume plane strain, incompressibility and an isothermal condition. Using the solutions obtained from the analysis two boundary value problems are illustrated: 1) the creep continuum is dislocated in the radial direction with constant velocity and 2) the media is dislocated in the axial direction with a constant velocity. If the constant velocity term in the analysis presented herein is replaced with a constant displacement term, then these linear elastic continuum problems are known as Volterra problems. If the problem is reduced to the problem for a perfectly viscous material (m=1.0), the solution for the stress expression is shown to be identical to existing classical solutions. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 9 Electronic Resource Springer Acta mechanica 79 (1989), S. 183-205 ISSN: 1619-6937 Source: Springer Online Journal Archives 1860-2000 Topics: Mechanical Engineering, Materials Science, Production Engineering, Mining and Metallurgy, Traffic Engineering, Precision Mechanics , Physics Notes: Summary Three stress solutions for a cylindrical particle on an interface bonded with two identical materials are obtained by using the pseudo-stress function developed previously. The stress solutions correspond to three different bonding mechanisms of a particle interface: 1) complete bonding of a rigid cylindrical particle, 2) radial bonding of the cylindrical rigid particle, and 3) complete bonding with deformable cylindrical particle in a perfectly viscous material. Using stress analysis results, the steady-state vacancy flux distribution at the interface of two materials (material interface) may be expressed as an in terms of the second derivative of normal stress on the interface with respect to two-dimensional space. The behavior of a vacancy sink or a vacancy source is found to be dependent on particle size, relative magnitude of material vs. particle properties, and bonding mechanisms, and loading conditions. In the case of the third stress solution, decohesion at the interface may be prevented by controlling a dimensionless parameter Ω1, which is a function of $${{\sigma _0 ^p } \mathord{\left/ {\vphantom {{\sigma _0 ^p } {\sigma _0 }}} \right. \kern-\nulldelimiterspace} {\sigma _0 }}$$ and the external stress and strain rate hardening exponent, where $$\sigma _0 ^p$$ and σ0 are associated with the particle and matrix constitutive equation $$\bar \sigma = \sigma _0 \dot \bar \varepsilon ^m$$ respectively. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... • 10 Electronic Resource Woodbury, NY : American Institute of Physics (AIP) Applied Physics Letters 77 (2000), S. 1244-1246 ISSN: 1077-3118 Source: AIP Digital Archive Topics: Physics Notes: Femtosecond optical pulses are used to generate narrow-band terahertz wave forms via optical rectification in a periodically poled lithium niobate crystal. By cooling the crystal to reduce losses due to phonon absorption, we are able to obtain bandwidths as narrow as 18 GHz at a carrier frequency of 1.8 THz. Temperature-dependent measurements show insignificant bandwidth broadening between 10 and 120 K, although the terahertz power substantially decreases as the temperature increases. Absolute power measurements indicate a conversion efficiency of at least 10−5. © 2000 American Institute of Physics. Type of Medium: Electronic Resource Signatur Availability Others were also interested in ... Close ⊗
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5725104212760925, "perplexity": 11656.501641293054}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496671239.99/warc/CC-MAIN-20191122042047-20191122070047-00305.warc.gz"}
https://docs.helium.com/devblog/2020/11/06/blockchain-release-rak-miner-updates/
# RAK Hotspot Miner: Firmware Update Issues The team has noticed issues where a small number of RAK Hotspot Miners are not applying firmware updates containing bug fixes. After investigation, we found that these Miners were booting from a secondary, fallback image that is used as backup on the device. The Miner is fooled into thinking a reset button was pressed which doesn't exist on the RAK Hotspot Miner. The code is in place to support Helium Hotspots. Because the RAK Hotspot Miners continue to think the "reset button" is being pushed, they will always boot from their older, fallback image. This issue only affects a very small number of RAK Hotspot Miners #### Identifying the Problem​ • Is it a Helium Hotspot? -> If yes, you do not have the issue • Is it actively mining? -> If yes, you do not have the issue • Does it have the latest firmware 2020.10.31.1? -> If yes, you do not have the issue If you find that you do not have the latest firmware, unplug it from power, wait ~30 minutes, and plug it back in. Then check if it has the latest firmware. If it doesn’t update, you have an affected unit. #### Interim Solution​ Since we know OTA updates continue to be applied to the primary image, the fix is to copy the primary image to the fallback image, effectively making the image the same. This allows the RAK Hotspot Miners to boot the fallback image which is the latest version. #### Permanent Solution​ The permanent solution requires two steps: 1. An OTA with a permanent fix that will be applied to the primary image. 2. Manually copying the primary image to the fallback image and reboot. Once rebooted, the RAK Hotspot miner will download and apply new firmware updates to both locations, so there is no difference between locations. ### Plan​ The short term fix can be applied at any time. If your RAK Hotspot Miner does not have the latest update 2020.10.31.1, please contact support@helium.com. You can find out by running a diagnostic report on the Miner. The permanent fix needs additional testing. The plan is to beta the next firmware on Tuesday, November 10th, 2020 with a scheduled GA on Wednesday, November 11th, 2020.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.17493490874767303, "perplexity": 4112.853138630745}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499758.83/warc/CC-MAIN-20230129180008-20230129210008-00389.warc.gz"}
https://collaborate.princeton.edu/en/publications/measurement-of-the-top-quark-mass-using-proton-proton-data-at-s7-
# Measurement of the top quark mass using proton-proton data at (s)=7 and 8 TeV MEASUREMENT of the TOP QUARK MASS USING ⋯ V. KHACHATRYAN et al. CMS Collaboration Research output: Contribution to journalArticlepeer-review 66 Scopus citations ## Abstract A new set of measurements of the top quark mass are presented, based on the proton-proton data recorded by the CMS experiment at the LHC at s=8 TeV corresponding to a luminosity of 19.7 fb-1. The top quark mass is measured using the lepton+jets, all-jets and dilepton decay channels, giving values of 172.35±0.16(stat)±0.48(syst) GeV, 172.32±0.25(stat)±0.59(syst) GeV, and 172.82±0.19(stat)±1.22(syst) GeV, respectively. When combined with the published CMS results at s=7 TeV, they provide a top quark mass measurement of 172.44±0.13(stat)±0.47(syst) GeV. The top quark mass is also studied as a function of the event kinematical properties in the lepton+jets decay channel. No indications of a kinematic bias are observed and the collision data are consistent with a range of predictions from current theoretical models of ttproduction. Original language English (US) 072004 Physical Review D 93 7 https://doi.org/10.1103/PhysRevD.93.072004 Published - Apr 7 2016 ## All Science Journal Classification (ASJC) codes • Nuclear and High Energy Physics ## Fingerprint Dive into the research topics of 'Measurement of the top quark mass using proton-proton data at (s)=7 and 8 TeV MEASUREMENT of the TOP QUARK MASS USING ⋯ V. KHACHATRYAN et al.'. Together they form a unique fingerprint.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9880256056785583, "perplexity": 3409.526282162679}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337524.47/warc/CC-MAIN-20221004184523-20221004214523-00101.warc.gz"}
http://mathhelpforum.com/algebra/168266-dvt-word-problem-help-needed.html
# Math Help - DVT word problem help needed 1. ## DVT word problem help needed A airplane that travels from one city to another city against the wind takes it 4 hours to get there, the voyage back only takes it 2 hours. If the distance between both city's is 600km, what is the speed of the plane? So far I have: D = V x T Go 600 4 Return 600 2 It's a table. So, 600/4=V 600/2=V Is this right so far? 2. Originally Posted by Methodd A airplane that travels from one city to another city against the wind takes it 4 hours to get there, the voyage back only takes it 2 hours. If the distance between both city's is 600km, what is the speed of the plane? So far I have: D = V x T Go 600 4 Return 600 2 It's a table. So, 600/4=V 600/2=V Is this right so far? $600 = (v - u)(4)$ .... (1) $600 = (v + u)(2)$ .... (2) Equate equations (1) and (2) and simplify: v = 3u. Substitute v = 3u into either equation (1) or (2) and solve for u. Therefore get v. I will let you think about which of u or v is the speed of the plane. 3. Originally Posted by mr fantastic $600 = (v - u)(4)$ .... (1) $600 = (v + u)(2)$ .... (2) Equate equations (1) and (2) and simplify: v = 3u. Substitute v = 3u into either equation (1) or (2) and solve for u. Therefore get v. I will let you think about which of u or v is the speed of the plane. Ah yes, its coming back now. Thank you very much my good sir! 4. Originally Posted by Methodd A airplane that travels from one city to another city against the wind takes it 4 hours to get there, the voyage back only takes it 2 hours. If the distance between both city's is 600km, what is the speed of the plane? So far I have: D = V x T Go 600 4 Return 600 2 It's a table. So, 600/4=V 600/2=V Is this right so far? Okay, so the net speed going (against the wind) was 600/4= 150 mph and returning (with the wind) was 600/2= 300 mph. Taking v to be "indicated air speed" (speed of the plane relative to the wind) and w to be the wind speed, we have v- w= 150 and v+ w= 300. Add those two equations to solve for v and subtract them to solve for w.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 4, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8042819499969482, "perplexity": 1421.5853997457311}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500822407.51/warc/CC-MAIN-20140820021342-00434-ip-10-180-136-8.ec2.internal.warc.gz"}
https://www.edaboard.com/members/adun.43371/recent-content
1. ### a question about Maxwell equation thanks lagrange, i just realized my math flaw when i am taking a shower:) hail to the maxwell equations 2. ### Gain differences between HFSS simulation and test results how do you manage to simulate a transistor device (LNA) in HFSS? 3. ### a question about Maxwell equation But i think the ambiguity here is between (a) & (b) (a) Physically, change of coordinate system should NOT change how fast the charge accumulates in a point. (b) Mathematically, change of coordinate system does seem to change how fast the charge accumulates in a point. 4. ### a question about Maxwell equation please refer to the attached picture. thanks:) 5. ### How to calculate/estimate the resonate frequency in Microstrip Resonator? Hi, If i have a microstrip line with dimension W and L, (L>W). The substrate is of height H and dielectric constatnt Er ( Hence an effective dielectric constant Eeff can be calculated). This segment of microstrip line should resonate at certain frequency right? So I have 2 questions: (1) How... 6. ### HFSS interpolating sweep problem i saw 'interpolating' sweep missed resonate frequency for several times. but i never saw it reported a false resonate freq. 7. ### compact phase shifter at 1MHz Hello, I am trying to build a compact phase shifter at low frequency, say 1MHz, with a phase shift of ±140~150 degree. The phase shift does not have to be tunable. Could you please show me some directions on it? Thanks a lot! p.s. i have done some literature review and found out that phase... 8. ### HFSS to ADS design Export: Help Plz to export gerber file from an HFSS file decribing a multi-layered structure (say 6 metal layers) you need to define 6 offset coordinates, whose origin lie on eatch of the 6 layers. Then you export them one by one to gerber format. 9. ### Measurement using UTF (Universal Test Fixture) universal test fixture Hi there, I have fabricated stripline bandpass filter on LTCC. A pair of extra vias and pads are added to elevate the buried striline port up to the surface so that measurement can be done by UTF. The performance of the filter looks fine under HFSS simulation. However... 10. ### [Matrix Manipulation] Help plz! A,B,C,D (hence P,Q) and X are all unknown... 11. ### [Matrix Manipulation] Help plz! Hi Louis the 'X' is in the equation $P\times X\times Q=T2$ 12. ### [Matrix Manipulation] Help plz! Assuming $P=\left[\begin{array}{cc}A & B \\C & D\end{array}\right]$ and $Q=\left[\begin{array}{cc}D & B \\C & A\end{array}\right]$ where A, B, C, D are unknown. It's also known that $P\times M\times Q=T1$ $P\times X\times Q=T2$ where X is unknown M, T1 and T2 are known. All the... 13. ### How to synthesize lumped model from s-param of inductor? this is a huge area of research... which often leads to a phd work so you would not find any easy to use software. you may like to search inductor modelling on IEEE 14. ### PML thickness in HFSS did you mean the 'largest frequency'? 15. ### [Modelling] Good capacitor model? hi i am trying to find a better Metal-Insulator-Metal capacitor model. This is what I have right now: C=Er*E0*(a+f)*(b+f)/h where a, b are the width and length of the two rectangular plates; h is this distance between the two plates, (or thickness of the insulator) and f represents effective...
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8325661420822144, "perplexity": 2705.289350236562}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320306181.43/warc/CC-MAIN-20220129122405-20220129152405-00387.warc.gz"}
https://dgtal.org/doc/0.9.3/testMeasure_8cpp.html
DGtal  0.9.3 testMeasure.cpp File Reference #include <cstdio> #include <cmath> #include <iostream> #include <fstream> #include <vector> #include "DGtal/base/Common.h" #include "DGtal/math/MeasureOfStraightLines.h" Include dependency graph for testMeasure.cpp: Go to the source code of this file. ## Functions void testUnitSquare () void testUnitSquareCentroid () int main (int argc, char **argv) ## Detailed Description This program is free software: you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see http://www.gnu.org/licenses/. Date 2011/06/25 Functions for testing class Measure. This file is part of the DGtal library. This program is free software: you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see http://www.gnu.org/licenses/. Date 2010/03/03 This file is part of the DGtal library Definition in file testMeasure.cpp. ## ◆ main() int main ( int argc, char ** argv ) Definition at line 100 of file testMeasure.cpp. 101 { 102  trace.beginBlock ( "Testing class MeasureOfStraightLines" ); 103  trace.info() << "Args:"; 104  for ( int i = 0; i < argc; ++i ) 105  trace.info() << " " << argv[ i ]; 106  trace.info() << endl; 107 108  testUnitSquare(); 110 111  trace.endBlock(); 112 113  return 0; 114 } void beginBlock(const std::string &keyword="") Trace trace Definition: Common.h:137 double endBlock() void testUnitSquareCentroid() Definition: testMeasure.cpp:78 void testUnitSquare() Definition: testMeasure.cpp:51 std::ostream & info() ## ◆ testUnitSquare() void testUnitSquare ( ) Compute the measure of the unit square [0,1]x[0,1] Expected value : sqrt{2}/2 Definition at line 51 of file testMeasure.cpp. Referenced by main(). 52 { 53  vector<double> a; 54  vector<double> b; 55 56  a.push_back(0); 57  b.push_back(0); 58  a.push_back(1); 59  b.push_back(0); 60  a.push_back(1); 61  b.push_back(1); 62  a.push_back(0); 63  b.push_back(1); 64 65  MeasureOfStraightLines measure; 66 67 68  trace.info() << "Measure of the Straight of Lines of the unit square = " << measure.computeMeasure(a,b)<< std::endl; 69  trace.emphase() <<"Expected value = 0.707107 (sqrt(2)/2)"<<endl; 70 } Trace trace Definition: Common.h:137 The aim of this class is to compute the measure in the Lebesgues sense of the set of straight lines a... double computeMeasure(const std::vector< double > &a, const std::vector< double > &b) std::ostream & emphase() std::ostream & info() ## ◆ testUnitSquareCentroid() void testUnitSquareCentroid ( ) Compute the measure of the unit square [0,1]x[0,1] Expected value : sqrt{2}/2 Definition at line 78 of file testMeasure.cpp. Referenced by main(). 79 { 80  vector<double> a; 81  vector<double> b; 82 83  a.push_back(0); 84  b.push_back(0); 85  a.push_back(1); 86  b.push_back(0); 87  a.push_back(1); 88  b.push_back(1); 89  a.push_back(0); 90  b.push_back(1); 91 92  MeasureOfStraightLines measure; 93 94  trace.info() << "Centroid measure of the unit square = (" << measure.computeCentroidA(a,b) 95  << ","<<measure.computeCentroidB(a,b)<<")"<<std::endl; 96  trace.emphase() <<"Expected value = (0.4142,0.5)"<<endl; 97 } Trace trace Definition: Common.h:137 The aim of this class is to compute the measure in the Lebesgues sense of the set of straight lines a... double computeCentroidA(const std::vector< double > &a, const std::vector< double > &b) std::ostream & emphase() double computeCentroidB(const std::vector< double > &a, const std::vector< double > &b) std::ostream & info()
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7924845814704895, "perplexity": 6862.616915977475}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618038101485.44/warc/CC-MAIN-20210417041730-20210417071730-00616.warc.gz"}
https://internetdo.com/2023/02/solving-sbt-lesson-4-multiplication-and-division-of-one-variable-polynomials-c7-sbt-math-7-horizons-math-book/
## Solving SBT Lesson 4: Multiplication and division of one-variable polynomials (C7 SBT Math 7 Horizons) – Math Book Solving SBT Lesson 4: Multiplication and division of one-variable polynomials (C7 SBT Math 7 Horizons) =========== ### Solve problem 1 page 32 SBT Math 7 Creative horizon episode 2 – CTST Multiply $$\left( {7x – 2} \right)\left( { – 2x + 5} \right)$$. Detailed instructions for solving Lesson 1 Solution method Understand the one-variable polynomial rule: To multiply a polynomial by a polynomial, we multiply each term of this polynomial by each term of the other polynomial and then add the products together. Detailed explanation $$\left( {7x – 2} \right)\left( { – 2x + 5} \right) = 7x\left( { – 2x} \right) + 7x.5 + \left( { – 2} \ right)\left( { – 2x} \right) + \left( { – 2} \right).5 = – 14{x^2} + 39x – 10$$ –> — ***** ### Solution 2 page 32 SBT Math 7 Creative horizon episode 2 – CTST Multiply $$\left( {3x – 4} \right)\left( { – 2{x^2} + 7x + 4} \right)$$. Detailed instructions for solving Lesson 2 Solution method Understand the rule of one-variable polynomials: To multiply a polynomial by a polynomial, we multiply each term of one polynomial by each term of the other polynomial and then add the products together. Detailed explanation $$\begin{array}{l}\left( {3x – 4} \right)\left( { – 2{x^2} + 7x + 4} \right)\\ = 3x\left( { – 2 {x^2}} \right) + 3x.7x + 3x.4 + \left( { – 4} \right).\left( { – 2{x^2}} \right) + \left( { – 4} \right).7x + \left( { – 4} \right).4\\ = – 6{x^3} + 21{x^2} + 12x + 8{x^2} – 28x – 16 \\ = – 6{x^3} + 29{x^2} – 16x – 16\end{array}$$ –> — ***** ### Solution 3 page 32 SBT Math 7 Creative horizon episode 2 – CTST Multiply $$\left( {4{x^2} – 2x + 1} \right)\left( { – 2{x^2} + 5x + 3} \right)$$. Detailed instructions for solving Lesson 3 Solution method Understand the rule of one-variable polynomials: To multiply a polynomial by a polynomial, we multiply each term of one polynomial by each term of the other polynomial and then add the products together. Detailed explanation $$\begin{array}{l}\left( {4{x^2} – 2x + 1} \right)\left( { – 2{x^2} + 5x + 3} \right)\\ = 4{x^2}\left( { – 2{x^2}} \right) + 4{x^2}.5x + 4{x^2}.3 – 2x.\left( { – 2{x ^2}} \right) – 2x.5x – 2x.3 + \left( { – 2{x^2}} \right) + 5x + 3\\ = – 8{x^4} + 20{x^ 3} + 12{x^2} + 4{x^3} – 10{x^2} – 6x – 2{x^2} + 5x + 3\\ = – 8{x^4} + 24{x ^3} – x + 3\end{array}$$ –> — ***** ### Solution 4 page 33 SBT Math 7 Creative horizon episode 2 – CTST Make a polynomial expression in the variable $$x$$ representing the area of ​​the shaded part in Figure 1. Detailed instructions for solving Lesson 4 Solution method Use the learned area formulas to write the expression. Area of ​​a rectangle is the product of length and width Detailed explanation The area of ​​the large rectangle is $$\left( {3x + 2} \right)\left( {2x + 4} \right) = 2x.3x + 2x.2 + 4.3x + 4.2 = 6{x^2 } + 16x + 8$$. The area of ​​the small rectangle is $$x\left( {x + 1} \right) = {x^2} + x$$. The required area is $$6{x^2} + 16x + 8 – \left( {{x^2} + x} \right) = 5{x^2} + 15x + 8$$. –> — ***** ### Solve problem 5 page 33 SBT Math 7 Creative horizon episode 2 – CTST Do the division $$\left( {9{x^5} – 15{x^4} + 27{x^3} – 12{x^2}} \right):3{x^2}$$ . Detailed instructions for solving Lesson 5 Solution method Polynomial division by monomial: Divide each term of the polynomial by the monomial and add the quotients together. Detailed explanation $$\begin{array}{l}\left( {9{x^5} – 15{x^4} + 27{x^3} – 12{x^2}} \right):3{x^ 2}\\ = 9{x^5}:3{x^2} + \left( { – 15{x^4}} \right):3{x^2} + 27{x^3}:3 {x^2} + \left( { – 12{x^2}} \right):3{x^2}\\ = 3{x^3} – 5{x^2} + 9x – 4\end {array}$$ –> — ***** ### Solution 6 page 33 SBT Math 7 Creative horizon episode 2 – CTST Do the division $$\left( {2{x^2} – 5x + 3} \right):\left( {2x – 3} \right)$$. Detailed instructions for solving Lesson 6 Solution method Divide polynomials divide polynomials by setting the division calculation. Detailed explanation So $$\left( {2{x^2} – 5x + 3} \right):\left( {2x – 3} \right) = x – 1$$ –> — ***** ### Solution 7 page 33 SBT Math 7 Creative horizon episode 2 – CTST Do the division $$\left( {4{x^2} – 5} \right):\left( {x – 2} \right)$$. Detailed instructions for solving Lesson 7 Solution method Divide polynomials divide polynomials by setting the division calculation. Detailed explanation We perform polynomial division: So $$\frac{{4{x^2} – 5}}{{x – 2}} = 4x + 8 + \frac{{11}}{{x – 2}}$$. –> — ***** ### Solve problem 8 page 33 SBT Math 7 Creative horizon episode 2 – CTST Do the division $$\left( {4{x^3} – 7x + 2} \right):\left( {2{x^2} – 3} \right)$$. Detailed instructions for solving lesson 8 Solution method Divide polynomials divide polynomials by setting the division calculation. Detailed explanation So $$\frac{{4{x^3} – 7x + 2}}{{2{x^2} – 3}} = 2x + \frac{{ – x + 2}}{{2{x^) 2} – 3}}$$ –> — ***** ### Solve lesson 9 page 33 SBT Math 7 Creative horizon episode 2 – CTST Calculate the length of a rectangle whose area is $$4{y^2} + 4y – 3$$ (cm2) and the width is equal to $$\left( {2y – 1} \right)$$(cm). Detailed instructions for solving Lesson 9 Solution method The length of the rectangle is the quotient of the area divided by the width. Detailed explanation We have So the length of the given rectangle is $$2y + 3$$ cm. –> — ***** ### Solution 10 page 33 SBT Math 7 Creative horizon episode 2 – CTST Given a rectangular box with volume $$V = 3{x^3} + 8{x^2} – 45x – 50$$ (cm3), length is equal to $$\left( {x + 5} \right)$$ cm and height $$\left( {x + 1} \right)$$ cm. Calculate the width of the rectangular box. Detailed instructions for solving Lesson 10 Solution method The volume of a rectangular box is the product of the three dimensions. Width equals volume divided by product of length and height Detailed explanation We have the width calculated by the calculation: $$\left( {3{x^3} + 8{x^2} – 45x – 50} \right):\left[ {\left( {x + 5} \right)\left( {x + 1} \right)} \right] = \left( {3{x^3} + 8{x^2} – 45x – 50} \right):\left( {{x^2} + 6x + 5} \right)$$ We have So the width of the rectangle is $$3x – 10$$ cm. –> — *****
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5662564039230347, "perplexity": 3990.0966506483464}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943589.10/warc/CC-MAIN-20230321002050-20230321032050-00735.warc.gz"}
https://asmedigitalcollection.asme.org/GT/proceedings/GT1993/78903/V03AT15A018/236385
Turbulence measurements were conducted inside a porous-wall tube insert placed coaxially in a solid-wall outer tube. The outer tube diameter was 8 inches. The diameter of the inner tube was 6.5 inches. Three different porosities of 7.5, 15 and 22.5 percent were used for the inner tube. The average velocity of the flow in the outer tube was 55 ft/sec. The nominal Reynolds Number, based on diameter and mean velocity of the flow in the outer tube, is 2 × 105. Results are presented for the three rms fluctuations, the Reynolds stress and the turbulence kinetic energy. The nature of the turbulence activity was examined by using the quadrant splitting scheme. This content is only available via PDF.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9658714532852173, "perplexity": 800.1898264776456}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875144498.68/warc/CC-MAIN-20200220005045-20200220035045-00387.warc.gz"}
https://www.physicsforums.com/threads/sterling-numbers-of-the-first-kind-combinatorial-proof.396049/
# Homework Help: Sterling numbers of the First kind Combinatorial Proof 1. Apr 17, 2010 ### silvermane 1. The problem statement, all variables and given/known data Say we have sterling numbers of the first kind where we're given s(n, n-2) = 2(nC3) + 3(nC4) for n greater than or equal to 4. 3. The attempt at a solution So, for the left side, we have n people, and we wish to seat them at n-2 circular tables, where if we first seat n-2 people, we'll have 2 remaining people to sit down at any of the n-2 circular tables. I just want to know if I'm on the right track here; thank you so much for your time! 2. Apr 19, 2012
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8080754280090332, "perplexity": 500.9980802421694}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267160620.68/warc/CC-MAIN-20180924165426-20180924185826-00197.warc.gz"}
http://mathhelpforum.com/calculus/70038-finding-limits.html
# Math Help - Finding limits... 1. ## Finding limits... $ \lim_{x \to 0} {{tan3x}} / {{sin8x}} $ Can I take the 3/8 out and just multiply it at the end? In any case, I can't find the limit! I know $ \lim{x /to 0} {{sinx}} / {{x}} = 1 $ but I can't figure out how to get it to work out in this problem... Also, where can I learn the proper math notation (ie \lim_{x \to 0} {{tan3x}} )? 2. Originally Posted by tom ato $ \lim_{x \to 0} {{tan3x}} / {{sin8x}} $ Can I take the 3/8 out and just multiply it at the end? In any case, I can't find the limit! I know $ \lim{x /to 0} {{sinx}} / {{x}} = 1 $ but I can't figure out how to get it to work out in this problem... Also, where can I learn the proper math notation (ie \lim_{x \to 0} {{tan3x}} )? I'll help you to a certain extent... $\lim_{x\to0}\frac{\tan(3x)}{\sin(8x)}=\lim_{x\to0} \frac{\displaystyle\frac{\sin(3x)}{\cos(3x)}}{\sin (8x)}=\lim_{x\to0}\frac{\sin(3x)}{\sin(8x)}\cdot\l im_{x\to0}\frac{1}{\cos(3x)}=\lim_{x\to0}\frac{\si n(3x)}{\sin(8x)}$ Now, try to incorporate $\lim_{x\to0}\frac{\sin x}{x}$ into this limit. Can you take it from here??
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 6, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.995366096496582, "perplexity": 491.27732464967585}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-42/segments/1414637899132.1/warc/CC-MAIN-20141030025819-00138-ip-10-16-133-185.ec2.internal.warc.gz"}
http://www.quantandfinancial.com/2013/
## Saturday, August 31 ### Portfolio Optimization II : Black-Litterman model In the previous post, we have been discussing conventional approach to the portfolio optimization, where assets' expected returns, variances and covariances were estimated from historical data. Since these parameters affect optimal portfolio allocation, it is important to get their estimates right. This article illustrates how to achieve this goal using Black-Litterman model and the technique of reverse optimization. All examples in this post are build around the case study implemented in Python. Instability of asset returns Empirical studies show that expected asset returns explain vast majority of optimal portfolio weightings. However, extrapolation of past returns into the future doesn't work well due to a stochastic nature of financial markets. Nevertheless, our goal is to achieve optimal and stable asset allocation, rather than trying to predict future market returns. The world of efficient markets To maximize investor's utility function, Modern Portfolio Theory suggests holding the combination of a risk-free asset and optimal risky portfolio (ORP) lying on a tangency of the efficient frontier and capital market line. Modern Portfolio Theory also suggests that optimal risky portfolio is a market portfolio (e.g. capitalization-weighted index). If we assume markets are fully efficient and all assets are fairly priced, we don't have any reason to deviate from the market portfolio in our asset allocation. In such case, we don't even need to know equilibrium asset returns nor perform any kind of portfolio optimization, as outlined in the previous article. An optimization based on equilibrium asset returns would lead back to the same market portfolio anyway. An active investor's view The different situation is when investor believes the market as a whole is efficient, but has concerns about the performance of specific assets or asset classes due to the possession of material non-public information, resulting for example from superior fundamental analysis. If investor with distinct views on a performance of few specific securities doesn't want to completely abolish the idea of mean-variance optimization, he may still use quantitative approach to incorporate such view into the MVO framework, further referred to as the Black-Litterman model. Black-Litterman model In the previous article, we have been discussing classic Mean-Variance optimization process. The process solves for asset weights which maximize risk-return trade-off of the ORP portfolio, given historical asset returns and covariances: On the other hand, the general workflow of the Black-Litterman optimization is illustrated on the diagram below. The forward-optimization part in this model is the same as in the classic MVO process (boxes bgijkl). However, the thing which differs is a way how we are observing expected asset returns, which is one of the inputs into the forward optimizer (g). Previously we have used historical asset returns as a proxy. Now, as we can see in the diagram, equilibrium asset returns are used instead (d or g). Equilibrium asset returns (d) are returns implied from the market capitalization weights of individual assets or asset classes (a) and historical asset covariances (b) in a process known as reverse optimization (c). Forward (h) and reverse (c) optimizations are mutually inverse functions. Therefore, forward-optimizing equilibrium asset returns (d) would yield to the optimal risky portfolio (j) with exactly the same weights as the weights observed from assets' capitalization (a). However, the beauty of the Black-Litterman model comes with the ability to adjust equilibrium market returns (f) by incorporating views into it and therefore to get optimal risky portfolio (j) reflecting those views. This ORP portfolio may be therefore different to the initial market cap weights (a). Case Study We will build upon the case study introduced in the previously published post Mean-Variance portfolio optimization. In that post, we have used forward optimizer to calculate weights of an optimal risky portfolio $$\mathbf(W_{n \times 1})$$ given historical returns $$\mathbf(R_{n \times 1})$$ and covariances $$\mathbf(C_{n \times n})$$  Now, we will use the same case study to illustrate the principle of the Black-Litterman model. Before doing so, let us define some basic symbols: $$\mathbf{R}$$ Vector of historical asset returns $$\pi$$ Vector of implied equilibrium returns $${\pi}'$$ Vector of view-adjusted implied equilibrium returns $$\mathbf{C}$$ Matrix of historical return covariances $$\mathbf{W}$$ Vector of market capitalization weights $$\mathbf{W}'$$ Vector of optimal risky portfolio weights $$\lambda$$ Risk-return trade-off $$\mathbf{Q}$$ Vector of views about asset returns $$\mathbf{P}$$ Matrix linking views to the portfolio assets Reverse optimization of equilibrium returns Equilibrium market returns are those implied by the capitalization weights of market constituents. They represent more reasonable estimates than those derived from historical performance. The another advantage is that equilibrium returns (whether or not adjusted) will yield into the portfolio weights similar than weights of the capitalization weighted index, thus representing more intuitive and investable portfolios. Equilibrium excess returns $$\pi$$ are derived in Black-Litterman using this formula: $$\lambda = \frac{r_p - r_f}{\sigma_p^2}$$ $$\pi=\lambda \cdot \mathbf{C} \cdot \mathbf{W}$$ In our example, corresponding code is represented by the following snippet. Please note that excess returns don't include risk free rate. # Calculate portfolio historical return and variance mean, var = port_mean_var(W, R, C) # Black-litterman reverse optimization lmb = (mean - rf) / var # Calculate return/risk trade-off Pi = dot(dot(lmb, C), W) # Calculate equilibrium excess returns As we can see, equilibrium returns implied from capitalization weights represent more conservative estimates than those calculated from historical performance. For example, implied return on Google $$\pi_6$$ is "only" 16.5% given its capitalization weight of 11.6%, as opposed to unjustified historical performance $$\mathbf{R}_6$$ of 33.2%. Name $$\mathbf{W}$$ $$\pi$$ $$\mathbf{R}$$ XOM 16.0% 15.6% 7.3% AAPL 15.6% 17.9% 13.0% MSFT 11.2% 15.2% 17.2% JNJ 9.7% 9.9% 14.4% GE 9.4% 17.9% 14.0% GOOG 11.6% 16.5% 33.2% CVX 9.2% 17.3% 9.2% PG 8.5% 9.4% 11.2% WFC 8.7% 21.8% 26.9% Efficient frontier based on historical returns is charted in red, while implied returns are in green: Incorporating custom views to equilibrium returns As we have already mentioned, the whole idea behind the reverse and forward optimization in Black-Litterman model is a process of incorporating custom views to the returns. The good news is that the whole process can be expressed as mathematical formula, while the bad news is that the steps to derive this formula are beyond the scope of this article. Given the views and links matrices $$\mathbf{Q}$$ and $$\mathbf{P}$$, we calculate view-adjusted excess returns $${\pi}'$$ as follows: $$\tau = 0.025$$ $$\omega = \tau \cdot \mathbf{P} \cdot \mathbf{C} \cdot \mathbf{P}^T$$ $${\pi}' = [(\tau \mathbf{C})^{-1} + \mathbf{P}^T \omega^{-1} \mathbf{P}]^{-1} \cdot [(\tau \mathbf{C})^{-1} \pi + \mathbf{P}^T \omega^{-1} \mathbf{Q})]$$ , where $$\tau$$ is scaling factor, $$\omega$$ is uncertainty matrix and $${\pi}'$$ is vector of view-adjusted equilibrium excess returns. Please note that rather than specifying uncertainty matrix of views $$\mathbf{P}$$ explicitly by the investor, we derive it from the formula above. It's because we believe that volatility of assets represented by covariance matrix $$\mathbf{C}$$ is affecting certainty of both returns and views in a similar way. Going back to our example, we have the vector of nine excess returns $$\pi$$, to which we want to apply our views. Let's our views to be as below: • Microsoft (MSFT) will outperform General Electric (GE) by 2% • Apple (AAPL) will under-perform Johnson & Johnson (JNJ) by 2% Of course these views are created randomly for the sake of our example, but in a real-world scenario they may be the result of some kind of observation based on statistical arbitrage. These two views will be mapped to a portfolio of nine assets. Therefore, the corresponding views and link matrices will have the following dimensions and content: $$\mathbf{Q}_{1 \times 2} = \begin{pmatrix}0.02 & 0.02\end{pmatrix}$$ $$\mathbf{P}_{2 \times 9} = \begin{pmatrix}0 & 0 & 1 & 0 &-1 & 0 & 0 & 0 & 0\\ 0 &-1 & 0 & 1 & 0 & 0 & 0 & 0 & 0\end{pmatrix}$$ , where the first row in a link matrix $$\mathbf{P}$$ has a positive sign +1 for MSFT and negative -1 for GE. By applying these views to an original vector of excess equilibrium returns $$\pi$$, we will get the view-adjusted vector $${\pi}'$$, later used to forward-optimize optimal risky portfolio weights $$\mathbf{W}'$$: Name $$\mathbf{W}$$ $${\pi}$$ $${\pi}'$$ $$\mathbf{W}'$$ XOM 16.0% 15.6% 14.9% 15.3% AAPL 15.6% 17.9% 13.1% 3.3% MSFT 11.2% 15.2% 15.7% 22.6% JNJ 9.7% 9.9% 10.1% 21.9% GE 9.4% 17.9% 16.1% -0.0% GOOG 11.6% 16.5% 15.2% 11.6% CVX 9.2% 17.3% 16.4% 8.8% PG 8.5% 9.4% 9.2% 8.4% WFC 8.7% 21.8% 20.5% 8.2% These views are incorporated into equilibrium returns $$\pi$$ also with respect to the uncertainty matrix $$\omega$$. Therefore, for example, the stated view that MSFT will outperform GE by 2% is not fully incorporated into a difference in their implied returns, which is just 16.1% - 15.7% = 0.4%. Once the view-adjusted equilibrium returns are ready, the final step is forward mean-variance optimization of the market weights $$\mathbf{W}'$$ with respect to these returns $${\pi}'$$ and historical covariances $$\mathbf{C}$$, pretty much as described in the previous post: Mean-Variance Portfolio Optimization Graphically, the efficient frontier constructed using equilibrium returns before and after incorporating views looks as follows. Conclusion Black-Litterman model is based on an assumption that asset returns have greatest impact to portfolio weightings in mean-variance optimization. It is therefore attempting to reverse-engineer these returns from index constituents rather than relying on historical data. This results into several advantages including but not limited to: • Intuitive portfolios with weightings not too much different from benchmark indices. • Ability to incorporate custom views • Lower tracking error and reliance on historical data • Greater stability of efficient frontier and better diversification ## Saturday, July 20 ### Mean-Variance Portfolio Optimization Calculation of assets' weights, returns and covariances Our example starts with the calculation of vector of asset weights $$\mathbf{W}_{n \times 1}$$, expected annual returns $$\mathbf{R}_{n \times 1}$$ and covariances $$\mathbf{C}_{n \times n}$$. Top nine S&P companies sorted by market capitalization are used as asset classes for the purpose of this article: XOM, AAPL, MSFT,  JNJ, GE, GOOG, CVX, PG and WFC, as of 1 Jul 2013. The time frame of historical data should depend on planned investment horizon. For the purpose of our analysis, we will consider daily prices for recent two years. The function call from previously mentioned example implementing the loader is as follows: # Load names, prices, capitalizations from yahoo finance # Estimate assets's expected return and covariances names, W, R, C = assets_meanvar(names, prices, caps) rf = .015 # Define Risk-free rate The data downloaded in this example will yield to the historical returns, deviations and capitalization-based weights as shown below. In the next section, we will use these figures to calculate portfolio risk/return characteristics and to optimize its asset weights. Name       Weight Return    Dev XOM         16.0%   7.3%  19.8% AAPL        15.6%  13.0%  30.3% MSFT        11.3%  17.2%  22.6% JNJ          9.7%  14.4%  13.9% GE           9.4%  14.0%  24.3% GOOG        11.6%  33.2%  25.9% CVX          9.2%   9.2%  22.6% PG           8.5%  11.2%  15.9% WFC          8.7%  26.9%  30.0% In the investment management process, almost every risky asset class exhibits some degree of trade-off between the returns and uncertainty of these returns. Mean returns are quantitatively measured by geometrically averaging series of daily or weekly returns over a certain period of time, while the uncertainty is expressed by the variance or standard deviation of such series. Naturally more volatile the asset is, higher is also the expected return. This relationship is determined by the supply and demand forces on a capital market and may be mathematically expressed by one of the capital pricing models, such as APT, CAPM or others. Most of these pricing models are based on an assumption that only systematic risk is reflected in capital prices and therefore investors shouldn't expect additional risk premium by holding poorly diversified or standalone investments. Assuming normal distribution of asset returns, we can quantitatively measure this diversification benefit by calculating correlation between two price series, which is equal or less than one. Let's consider a simple example of portfolio containing two assets, 50% of asset A and 50% of asset B, where asset returns are $$r_{A}=4\%$$, $$r_{B} = 5\%$$ and variances are $$\sigma_{A}^{2}=7\%^{2}$$ and $$\sigma_{B}^{2}=8\%^{2}$$. In case there is no diversification benefit, portfolio expected return and variance will be just linear combination of the asset weights, in this case $$r_{p}=4.5\%$$, $$\sigma_{p}^{2}=7.5\%^{2}$$. However, in case the correlation between these two assets is less than one and therefore diversification potential is available, portfolio variance will be less than linear combination of weights: $$\sigma_{p}^{2}<7.5\%^{2}$$. The mean-variance trade-offs for different levels of diversification are shown on the Figure 1. Portfolio mean-variance calculation The generalized formula for calculating portfolio return $$r_{p}$$ and variance $$\sigma_{p}$$ consisting of $$n$$ different assets is then defined as: $$r_{p} = \mathbf{W} \cdot \mathbf{R}$$ , where $$\mathbf{W}_{n\times 1}$$ is the vector of asset weights and $$\mathbf{R}_{n\times 1}$$ vector of corresponding asset returns. Portfolio variance is then defined as: $$\sigma_{p}^{2} = \mathbf{W} \cdot \mathbf{C} \cdot \mathbf{W}$$ , where $$\mathbf{C}_{n\times n}$$ is the covariance matrix of asset returns. The corresponding code in our python example: # Calculate portfolio historical return and variance mean, var = port_mean_var(W, R, C) Portfolio Optimization Considering the starting vector of weights $$\mathbf(W_{n \times 1})$$, the optimization process is tailored towards maximizing some kind of mean-variance utility function, such as Sharpe ratio: $$s=\frac{r_{p} - r_{f} }{\sigma_{p}}$$ , because we know that optimal risky portfolio with highest sharpe ratio $$s$$ lies on a tangency of efficient frontier with the capital market line (CML). Given the historical asset returns $$\mathbf{R}_{n \times 1}$$ and covariances $$\mathbf{C}_{n \times n}$$, the optimization code using SciPy's SLSQP method may look for example as follows: # Given risk-free rate, assets returns and covariances, this function calculates # weights of tangency portfolio with respect to sharpe ratio maximization def solve_weights(R, C, rf): def fitness(W, R, C, rf): # calculate mean/variance of the portfolio mean, var = port_mean_var(W, R, C) util = (mean - rf) / sqrt(var) # utility = Sharpe ratio return 1/util # maximize the utility n = len(R) W = ones([n])/n # start with equal weights b_ = [(0.,1.) for i in range(n)] # weights between 0%..100%. # No leverage, no shorting c_ = ({'type':'eq', 'fun': lambda W: sum(W)-1. }) # Sum of weights = 100% optimized = scipy.optimize.minimize(fitness, W, (R, C, rf), method='SLSQP', constraints=c_, bounds=b_) if not optimized.success: raise BaseException(optimized.message) return optimized.x # Return optimized weights The function returns vector of optimized weights, which together with original vector of returns and deviations looks as follows: # optimize W = solve_weights(R, C, rf) mean, var = port_mean_var(W, R, C) # calculate tangency portfolio front_mean, front_var = solve_frontier(R, C, rf) # calculate min-var frontier We can see that this optimization results in a concentrated portfolio consisting just of two constituents: Johnson&Johnson and Google: Name       Weight Return    Dev XOM          0.0%   7.3%  19.8% AAPL        -0.0%  13.0%  30.3% MSFT        -0.0%  17.2%  22.6% JNJ         45.1%  14.4%  13.9% GE           0.0%  14.0%  24.3% GOOG        52.8%  33.2%  25.9% CVX          0.0%   9.2%  22.6% PG          -0.0%  11.2%  15.9% WFC          2.1%  26.9%  30.0% It is not surprising that these assets have one of the lowest correlations among all pairs (0.431). Companies Apple and Exxon were not selected despite the fact that they have even lower mutual correlation. This is because their stand-alone risk/return trade-off is not that great indeed. Correlation between assets Calculation of minimum variance frontier In the previous section, we have used optimization technique to find the best combination of weights in order to maximize the risk/return profile (Sharpe ratio) of the portfolio. This resulted into a single optimal risky portfolio represented by a single point in the mean-variance graph. Although the utility function is clear, to maximize the Sharpe ratio, it has two degrees of freedom - the mean and the variance. In order to calculate the minimum variance frontier, we need to iterate through all levels of investor's risk aversity, represented by required return (y-axis) and use the optimization algorithm in order to minimize the portfolio variance. In each iteration, the required return is hold constant for the purpose of optimization, reducing the problem to one degree of freedom. The output of such optimization will be a vector of minimum variances, one value for each level of risk aversity, also called the minimum variance frontier. The corresponding Python function may look for example as follows: # Given risk-free rate, assets returns and covariances, this function calculates # min-var frontier and returns its [x,y] points in two arrays def solve_frontier(R, C, rf): def fitness(W, R, C, r): # For given level of return r, find weights which minimizes # portfolio variance. mean, var = port_mean_var(W, R, C) # Big penalty for not meeting stated portfolio return # effectively serves as optimization constraint penalty = 100*abs(mean-r) return var + penalty frontier_mean, frontier_var = [], [] n = len(R) # Number of assets in the portfolio # Iterate through the range of returns on Y axis for r in linspace(min(R), max(R), num=20): W = ones([n])/n # start optimization with equal weights b_ = [(0,1) for i in range(n)] c_ = ({'type':'eq', 'fun': lambda W: sum(W)-1. }) optimized = scipy.optimize.minimize(fitness, W, (R, C, r), method='SLSQP', constraints=c_, bounds=b_) if not optimized.success: raise BaseException(optimized.message) # add point to the min-var frontier [x,y] = [optimized.x, r] frontier_mean.append(r) frontier_var.append(port_var(optimized.x, C)) return array(frontier_mean), array(frontier_var) Again, we have chosen Sequential Least Square Programming method (SLSQP), which allows us to perform the constrained optimization based on the following constraints: • Sum of weights must be 0 • Weights must be in range (0,1) - no short-selling or leveraged holdings. • Portfolio required return is given The first constraint is provided using the lambda function lambda W: sum(W)-1., which is equality constraint saying that the whole expression must be as close to zero as possible. The second constraint are effectively search bounds passed to the optimization function and third constraint is implemented in a fitness function itself. It imposes a big penalty for portfolio mean return not meeting the target return "r". After solving minimum variance portfolios for all twenty levels of risk aversity, we will get the following minimum variance frontier with optimal risky portfolio represented by dot (o) and individual constituents represented by cross (x): Minimum Variance Frontier Conclusion The major drawback of mean-variance optimization based on historical returns is that such optimization leads to undiversified portfolios, as seen in our example. The reason behind this observation is that market prices are often far from their equilibriums on a risk-adjusted basis, as modeled by the Capital Market Pricing Model. Moreover, extrapolated historical returns are notoriously bad proxies to be used instead of expected mean returns, which are direct inputs into the mean-variance optimization algorithm. In the next post, we will illustrate how to use Black-Litterman model to overcome this problem by deriving returns from the market capitalization weights in the process known as reverse optimization.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 2, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6489433646202087, "perplexity": 3091.355715770933}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917118963.4/warc/CC-MAIN-20170423031158-00273-ip-10-145-167-34.ec2.internal.warc.gz"}
http://math.stackexchange.com/questions/89191/nonsingular-bilinear-map
# Nonsingular bilinear map This is related to a previous question I asked. But I realize that my logic there is total bonkers. And it would be great if someone could help me out a bit. $B:V\times V\to F$ is a bilinear form where $V$ is a finite-dimensional vector space. $X\leq V$ is a subspace which is also the annihilator of another subspace $Y\leq V$ wrt $B$. I am given that $B|_X$ is nonsingular. I wish to show that it follows that $B$ itself is nondegenerate. Thanks. - The term you should use is nondegenerate, as far as I know; nonsingular applies to linear maps (and certainly one doesn't want to say $B$ vanishes only on $(\vec0,\vec0)$). Being degenerate means some nonzero vector annihilates every other vector. Now how would such a vector in $V$, if it existed, be related to $X$? –  Marc van Leeuwen Dec 7 '11 at 10:33 @MarcvanLeeuwen: Thanks, I have edited the word. :) If there is a vector $x\in X$ that kills off all the others wrt $B$ then we have $B(x,X)=0$, right? But I still don't quite see how to prove the statement. –  Zachmann Dec 7 '11 at 10:37 What you are trying to prove is the nonexistence of a vector **in the whole space $V$** that kills off everyone, knowing the nonexistence of a vector $x\in X$ like the one you describe. Given the negative character of this statement, you seem to begin "at the wrong end" of the argument. –  Marc van Leeuwen Dec 7 '11 at 11:13 @MarcvanLeeuwen: Yes, your interpretation of the question is exactly right. –  Zachmann Dec 7 '11 at 11:16 If $B$ is not symmetric, then in general a subspace has both a "left annihilator" and a "right annihilator". You didn't specify, so I'll just pick one of the two: For $W \subset V$, $W^{\perp} = \{v \in V \ | \ B(v,W) = 0 \}$. Now let $0 \neq v \in V$. We want to find $v' \in V$ with $B(v,v') \neq 0$. Suppose first that $v \in X$. Then by nonsingularity of $X$, there exists $v' \in X$ such that $B(v,v') \neq 0$. Now suppose $v \in V \setminus X$. Since $X = Y^{\perp}$, $v \notin Y^{\perp}$, i.e., there exists $v' \in Y$ with $B(v,v') \neq 0$. - Thank you, Pete. –  Zachmann Dec 7 '11 at 12:07 I suppose you've figured it out by now but just in case... Assume $X$ is given to be the left annihilator of $Y$: $$X=Y^{\perp,\mathrm{left}} = \{v \in V \ | \ B(v,Y) = 0 \}.$$ Now if there existed a nonzero $v\in V$ with $\forall w\in V: B(v,w)=0$, then certainly $v\in Y^{\perp,\mathrm{left}}=X$, and certainly $B(v,x)=0$ for all $x\in X$, contradicting the nondegeneracy of $B|_X$. - Thank you, Marc. –  Zachmann Dec 7 '11 at 12:07
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9750890731811523, "perplexity": 291.98581820821073}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440644066266.26/warc/CC-MAIN-20150827025426-00152-ip-10-171-96-226.ec2.internal.warc.gz"}
https://www.forrs.de/news/FORRS_EOY.php
# Hit By Year-End First, maybe surprisingly, but the end of the calendar year is not necessarily the relevant End-of-year date for every market participant, for a variety of reasons: 1. First choice could be Dec 31st. Being Saturday or Sunday with no trading activity and consequently, no market data is available as of this date. 2. But also if Dec 31st is a weekday: Typically, there is comparably low liquidity on this day and some markets are even closed. Given these reasons, some market participants use the last business day before Dec 31st. 3. Some even don't use this one because it is perceived that market makers use their competitive advantage and the poor liquidity at EOY to quote market parameters in their favour. 4. Regardless of the relevant date (and – by the way – also time!) used to fix market parameters, some institutions – despite using market data from a day before Dec 31st – do apply Dec 31st as valuation date (in order to reflect respective accrual or Theta effects), others do not. Consequently, valuations for identical financial instruments can vary across different institutions. However, if no big market movements can be observed at EOY (this has been the case in the recent past), the respective impact is negligible. And from a single market participant’s view, it is only a question in which fiscal year the P&L in question is disclosed, there is no money made or lost due to the choice of a different EOY date. However, actual losses (or profits) caused by EOY procedures did happen in the past. Let’s look at a very simple example. Imagine a EUR bank founded on Jan 1st, 2020 with 1 million Euro equity and no debt. Imagine further that on day one, the whole equity amount has been invested in the trading book to buy Microsoft stocks on Nasdaq. Funding of the USD amount was done via an FX Swap maturing in one year. Following market data are observed: $FX \space Rate \space EUR/USD \space = \space 1$ $Yield \space curves \space in \space EUR \space and \space USD \space flat \space @ \space 0\%$ $Stock \space price \space Microsoft \space at \space purchase \space = \space 200 \space USD$ Risk management is done in the bank’s front office system. All relevant transactions are recorded there and one would expect the following cash flow structure: At year end, the bank’s book is closed. Microsoft’s share price has moved up to 220, FX and interest markets didn’t move at all. So we would observe the following figures: Quite obviously, the bank earned a profit – arising from the value change of the shares. Since the reporting currency of the bank is EUR, all amounts in foreign currency must be translated into EUR and the bank would report a profit of $USD \space 100.000 \space ^* \space 1 {EUR \over USD} \space = \space 100.000 \space EUR.$ The responsible trader might be compensated for his good performance with a bonus in the upcoming spring. Later in 2021 at the bank’s annual general meeting it might be decided that dividends are paid out. However, it is important to state that in our example there is no actual cash flow at year-end. More general, cash flows simply do not happen out of the blue solely because year-end is reached. This is a fact. However, some banks do ignore this within the scope of their end-of-year procedures: While pursuing the target of resetting the P&L framework, they are manipulating cash balances with the effect that front office system positions do not reflect reality anymore. Let’s go back to our example in order to get a better understanding: One way to reset P&L is to insert a cash booking in the amount of P&L in currency times -1 (let’s call this Alternative 1). In our case this would imply a cash booking of -100.000 USD. Right after year-end with no markets moving, P&L would be 0 USD and hence, 0 EUR (as expected). More generally, Year-to-Date P&L would be calculated in the Front Office System as follows: \begin{align*} \\P\&L_{YtD} &= Present \space Value \space (in \space CCY) \space ^* \space FX \space Rate \space ({EUR \over CCY})\\ \\&+ Cash \space Balance \space (in \space CCY) \space ^* \space FX \space Rate \space ({EUR \over CCY})\\ \\&- YearEndCorrection \space (in \space CCY) \space ^* \space FX \space Rate \space ({EUR \over CCY})\\ \\&= [Present \space value \space (in \space CCY) \space + \space Cash \space Balance \space (in \space CCY)\\ \\&- \space YearEndCorrection \space (in \space CCY)] \space ^* FX \space Rate \space ({EUR \over CCY})\\ \\&= 0 \end{align*} We can see from the equation that, if PV and Cash Balance would not change after year end, P&L figures would also not fluctuate. As stated before, Alternative 1 is a procedure commonly offered by front office systems. Traders might like this because they can start the new year with a “clean position” in every currency. However, this is not in line with accounting procedures and – in particular – reality. This EOY correction cash flow of -100.000 USD did not really occur and leads to a misrepresentation of risk: Since the reporting currency of the bank is EUR, every non-EUR cash flow bears FX risk exposure. If an artificial cash flow in foreign currency is entered, the FX risk position available from the Front Office System is wrong. In our example, assuming that PV and CB stay constant in the new year and FX markets moving, the artificial -100.000 USD EOY correction would keep P&L figures still constantly at zero. Accounting figures would show different results because the -100.000 USD correction cash flow would not have been booked in accounting systems. Within reconciliation of accounting and economic P&L it might be detected eventually that there is a difference. In the past it happened more than once that until then significant losses (or profits) already have been accumulated. We have seen so far that Alternative 1 might not be a good idea. So how about Alternative 2: In order to avoid the FX risk misstatement, a cash booking in the amount of P&L in accounting currency (here EUR) is inserted. In our example, we would book a cash flow of -100.000 EUR. This would reset the EOY profit of 100.000 USD to zero. And, since no FX cash flow is involved, the FX risk position apparently would stay as it was before. In terms of P&L fluctuation, future moves in the EUR/USD FX rate would impact the P&L because the EOY FX rate is implicitly fixed within the EOY adjustment in currency EUR. So if EUR/USD would move to 1,5 while PV’s and CB’s in currency remain constant, P&L would move accordingly: \begin{align*} P\&L_{YtD} &= Present \space Value \space (in \space CCY) \space ^* \space FX \space Rate \space ({EUR \over CCY})\\ \\&+ Cash \space Balance \space (in \space CCY) \space ^* \space FX \space Rate \space ({EUR \over CCY})\\ \\&- YearEndCorrection \space (in \space EUR)\\ \\&= 1.100.000 \space USD \space ^* \space 1,5{EUR \over USD} - 100.000 \space EUR\\ \\&= 50.000 \space EUR \end{align*} Thus, if Alternative 2 is applied, the risk position available from the Front Office System would be correct. At the same time we know also that there is no real cash flow in the amount of -100.000 EUR at year end. Does this create any issues? Not from a market risk perspective. However, the Cash Balance in reporting currency is misstated. Consequently, EOY correction cash flows should not be considered for calculation of funding needs. Furthermore, in case theoretical funding costs apply within the P&L calculation, the calculation basis would be distorted. Assuming a constant funding rate of 1% p.a. in our example, the EOY correction cash flow would produce an artificial and unjustified -1.000 EUR loss from funding costs. This problem could be avoided by Alternative 3, a slight modification of Alternative 2: Here, all EOY correction cash flows are input using an artificial currency, say “EOY” for which a constant exchange rate of 1 is applied versus the reporting currency. Cash flows in currency EOY are excluded from funding positions and are not subject to theoretical funding costs. FX risk is represented correctly. This is maybe the cleanest way to begin the new year. So far theory explained by a very simple example. Reality is by far more complex: Front office systems contain up to millions of trades in different instruments, often more than one front office system is involved, of course there is a portfolio structure containing a lot of desks with individual mandates, positions are mirrored, special processes have evolved for historical reasons and workarounds do exist. For correct EOY treatment, it is of paramount importance to reflect this complexity resp. reality. While application of Alternative 3 would put you on the safe side, you might get through also with Alternative 2 (in case of funding tickets setting all cash balances to zero on a daily basis) or even with Alternative 1 (in case all FX P&L is – actually – sold down to reporting currency on year end). We have seen that there is an operational risk exposure created by methodically incorrect EOY procedures. Additional operational risk exposure can arise by failure or delays of these processes. If traders have no access to correct risk exposures, proper risk management is not possible. Accordingly, strong governance concerning EOY methods and processes help mitigating this risk. More concretely, we recommend to have a look on the following aspects: • Start EOY planning punctually in order to have a timely reconciled (and where appropriate changed) procedure. • Make sure that your EOY method (still) reflects reality, in particular perform respective reviews with regard to changes made in your application and process ecosystem since last year end. • Determine the relevant EOY date and market data used long before, reconcile this with all internal stakeholders as well as external auditors. • Perform comprehensive test runs following a well-documented runbook and get approval of all stakeholders in order to minimize operational risk and prevent EOY losses. As a conclusion we can state that each year end is for sure not comparable with Y2K, nevertheless it is reasonable to invest in proper procedures to avoid losses from respective operational risks. Follow FORRS on LinkedIn for recent insights into financial and energy markets. Stay engaged - with FORRSight!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6222845315933228, "perplexity": 3793.6804731295165}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107876768.45/warc/CC-MAIN-20201021151342-20201021181342-00123.warc.gz"}
https://mathoverflow.net/questions/275282/is-the-surface-minor-ordering-of-plane-graphs-a-well-quasi-ordering
# Is the “surface-minor” ordering of plane graphs a well-quasi-ordering? A plane graph is a finite simple graph with a fixed embedding into the two-sphere. The embedding induces an embedding on a minors of a plane graph (i.e. a graph obtained by successive removal of vertices, removal of edges, and contraction of edges). In other words, we may consider minors of plane graphs to be plane graphs themselves. These minors are called surface minors. The surface-minor relation is an ordering on plane graphs, which is finer than the minor relation: a plane graph may be a minor of another plane graph, without being a surface minor (indeed, if a graph has several non-isomorphic embeddings into the two-sphere, then these are examples of that behavior). The question is: Is the surface-minor ordering of plane graphs a well-quasi-ordering? That is to say, is there among any infinite collection of plane graphs a pair of two plane graphs, one of which is a surface minor of the other? This seems a very natural question to me, yet I couldn't find an answer in the literature. One would assume that this question must be answered in one of Robertson and Seymour's papers, since both the notion of well-quasi-orderings and the notion of surface minors are pretty central in them. If one restricts oneself to plane trees, the answer is positive: this is a version of Kruskal's Tree Theorem. The answer is also positive if one restricts oneself to 3-connected plane graphs, since Whitney's Theorem says that they have a unique embedding, and so the surface-minor and the minor relation coincide. Of course, one can ask this question not only about the two-sphere, but about any other fixed (not necessarily orientable) closed surface. Please note this question is not a direct corollary of the Robertson-Seymour theorem that the minor ordering on finite graphs is a well-quasi-ordering; and it does not have a direct relation to Kuratowski's Theorem, which gives the two forbidden minors for planar graphs (not to be confused with the plane graphs this question is about). This is a partial answer, for the case when the given sequence $G_1,G_2,\dots$ of plane graphs has unbounded treewidth. In such a case, for every $n$ there is an $i$ such that $G_i$ contains the $n\times n$ grid as a minor, and thus also as a plane minor. The rest follows from the simple fact that $G_1$ is a plane minor of a sufficiently large plane grid.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.857545018196106, "perplexity": 129.3343593403241}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107893011.54/warc/CC-MAIN-20201027023251-20201027053251-00627.warc.gz"}
https://hpmuseum.org/forum/thread-16303-post-142813.html
(34C) (11C) Summation of Infinite Alternating Series 02-10-2021, 06:04 PM (This post was last modified: 02-13-2021 05:22 PM by Valentin Albillo.) Post: #1 Valentin Albillo Senior Member Posts: 887 Joined: Feb 2015 (34C) (11C) Summation of Infinite Alternating Series Extracts from HP Program VA342 - HP-34C Summation of Alternating Series Abstract SUMALT is a program written in 1979 for the HP-34C programmable calculator to quickly and accurately find the sum of infinite alternating series, even divergent ones (Euler sum). Three worked examples are included. Keywords: sum, infinite alternating series, Euler Transformation, Euler sum, differences, programmable calculator, RPN, HP-34C Introduction SUMALT is a short (84 steps) RPN program that I wrote in 1979 for the HP-34C calculator (will also run as-is or with minor modifications in many RPN models, such as the HP-11C) which, given an infinite alternating series (i.e.: consecutive terms alternate signs) whose general term is defined by the user, it will compute its sum very quickly using the Euler Transformation up to 7th-order differences. [...] V. All My Articles & other Materials here:  Valentin Albillo's HP Collection 02-12-2021, 03:04 PM Post: #2 Albert Chan Senior Member Posts: 2,066 Joined: Jul 2018 RE: (34C) Summation of Infinite Alternating Series The code is amazing ! I did spot a few typos in the pdf, however. S = Σ((-1)^i * y(i), i = 0 .. inf) = y(0)/2 - Δ(y(0))/4 + Δ²(y(0))/8 - ... S is splitted to 2 parts: S = S' + S'' S' = Σ((-1)^i * y(i), i = 0 .. n) S'' = y(n+1)/2 - Δ(y(n+1))/4 + Δ²(y(n+1))/8 - ... There are 2 issues with this. 1. S' actually summed n+1 terms, not n terms, as stated in the pdf 2. if n is even, S = S' - S'', not S' + S'' With these corrections, I was able to confirm example 1, S = 1 - 1/2 + 1/3 - 1/4 + ... Code: function euler_transform(t) -- S = t[1] - t[2] + t[3] - t[4] + ...     local n, k = #t, 0.5     for i = 2,n do          -- {1, D, D^2, ...} of t[1]         for j = n,i,-1 do t[j] = t[j] - t[j-1] end     end     for i = 1,n do t[i] = t[i]*k; k = -0.5*k; end end                         -- S = t[1] + t[2] + t[3] + t[4] + ... lua> n, d, t = 10, 7, {} lua> for i=n+1,n+d+1 do t[i-n] = 1/(i+1) end -- terms to estimate s2 lua> euler_transform(t) lua> table.foreachi(t, print) 1      0.041666666666666664 2      0.0016025641025641003 3      0.00011446886446886233 4      1.1446886446884671e-005 5      1.4308608058594997e-006 6      2.1042070674336805e-007 7      3.50701177901638e-008 8      6.4602848558431396e-009 lua> s1, s2 = 0, 0 lua> for i=0,n do s1 = s1 + (-1)^i/(i+1) end -- n+1 terms lua> for i=1,d+1 do s2 = s2 + t[i] end          -- d+1 terms lua> s1, s2 0.7365440115440116       0.043396829332061765 lua> s1 - s2       -- n is even, thus the minus sign 0.6931471822119498 We may apply Aitken Extrapolation, to slight improve estimate of s2 lua> s1 - (s2 - t[d+1]^2/(t[d+1] - t[d])) 0.6931471807531758 lua> log(2) 0.6931471805599453 02-13-2021, 05:51 PM Post: #3 Valentin Albillo Senior Member Posts: 887 Joined: Feb 2015 RE: (34C) (11C) Summation of Infinite Alternating Series Hi, Albert Chan: (02-12-2021 03:04 PM)Albert Chan Wrote: The code is amazing ! Thank you for your kind words, much appreciated. Quote:I did spot a few typos in the pdf, however. Nah, not exactly typos as in "I typed something incorrectly" but simply that I was being somewhat informal (originally it was a "Long Live"-Series article, not a formal math paper), not counting term 0 or caring to specify the precise alternating sign for the correction terms. However, rest assured that the RPN program code of course does take these details into account and is perfectly correct, as you may check if you run the examples featured or any others of your own. Thanks again and keep your eyes peeled for my next post tomorrow 14th, to commemorate my worldwide-known saint patron's festivity. I'll welcome any inputs from you but please, no Lua, no Xcas, I know you'll be itching to use them but please refrain from doing so. If you want to contribute something, I'd appreciate it if you would please post code or results for some HP calc using some HP-calc language and we'll remain in good terms, Ok ? Regards V. All My Articles & other Materials here:  Valentin Albillo's HP Collection 02-13-2021, 06:15 PM Post: #4 Albert Chan Senior Member Posts: 2,066 Joined: Jul 2018 RE: (34C) (11C) Summation of Infinite Alternating Series (02-13-2021 05:51 PM)Valentin Albillo Wrote:  Thanks again and keep your eyes peeled for my next post tomorrow 14th, to commemorate my worldwide-known saint patron's festivity. Looking forward to it Quote: I'll welcome any inputs from you but please, no Lua, no Xcas ... Noted. I just wanted to show how Euler transformation converted slowly convering sum, into a fast one. I should have just post the Euler transformed partial sum table ... 02-13-2021, 10:34 PM Post: #5 robve Senior Member Posts: 342 Joined: Sep 2020 RE: (34C) (11C) Summation of Infinite Alternating Series I have worked through this nice paper and algorithm too, many thanks Valentin! (02-12-2021 03:04 PM)Albert Chan Wrote:  We may apply Aitken Extrapolation, to slight improve estimate of s2 lua> s1 - (s2 - t[d+1]^2/(t[d+1] - t[d])) 0.6931471807531758 lua> log(2) 0.6931471805599453 Very clever to think of that. I have one question. I would think that t[d+1]^2/(t[d+1] - t[d]) and (x[n+1]-x[n])^2/((x[n+1]-x[n])-(x[n]-x[n-1])) are related if terms x[] are extrapolated, but that is not the case? Shouldn't the formula be (t[d]+t[d+1])^2/(t[d+1]) since when d=2 for example t[2]=x[n+2]-x[n+1] and t[3]=x[n+3]-2x[n+2]+x[n+1]? How did you derive your formula? - Rob "I count on old friends" -- HP 71B,Prime|Ti VOY200,Nspire CXII CAS|Casio fx-CG50...|Sharp PC-G850,E500,2500,1500,14xx,13xx,12xx... 02-13-2021, 11:30 PM Post: #6 Albert Chan Senior Member Posts: 2,066 Joined: Jul 2018 RE: (34C) (11C) Summation of Infinite Alternating Series Hi, robve Let the last 3 cumulative sum of s2 be a, b, c a + t[d] = b b + t[d+1] = c = s2 Aitken(a,b,c) = c - (c-b)^2/((c-b)-(b-a)) = s2 - t[d+1]^2 / (t[d+1] - t[d]) Or, Secant's method, from 2 points: (x1, y1) = (b, t[d]), (x2, y2) = (c, t[d+1]) Extrapolate for (x, t[∞] = 0). Both methods are equivalent. x = x2 - y2 * (x2-x1)/(y2-y1) = s2 - t[d+1]^2 / (t[d+1] - t[d]) 02-14-2021, 11:30 AM Post: #7 Albert Chan Senior Member Posts: 2,066 Joined: Jul 2018 RE: (34C) (11C) Summation of Infinite Alternating Series (02-13-2021 11:30 PM)Albert Chan Wrote:  Or, Secant's method, from 2 points: (x1, y1) = (b, t[d]), (x2, y2) = (c, t[d+1]) Selection of points is arbitrary, as long as it is consistent. Example, we can use mean of 2 cumulative sum for x's, estimated error gap for y's (x1, y1) = ((b+a)/2, (b-a)/2) (x2, y2) = ((c+b)/2, (c-b)/2) Again, Secant's method, extrapolate for (x, 0) x = (c+b)/2    - (c-b)/2 * ((c+b)/2 - (b+a)/2) / ((c-b)/2 - (b-a)/2)   = c - (c-b)/2 - (c-b)/2 * (c-a) / (c-2b+a)   = c - (c-b)/2 * ((c-2b+a) + (c-a)) / (c-2b+a)   = c - (c-b)^2 / (c-2b+a)   = Aitken(a, b, c) 02-16-2021, 01:57 AM Post: #8 robve Senior Member Posts: 342 Joined: Sep 2020 RE: (34C) (11C) Summation of Infinite Alternating Series (02-14-2021 11:30 AM)Albert Chan Wrote:  Again, Secant's method, extrapolate for (x, 0) x = (c+b)/2    - (c-b)/2 * ((c+b)/2 - (b+a)/2) / ((c-b)/2 - (b-a)/2)   = c - (c-b)/2 - (c-b)/2 * (c-a) / (c-2b+a)   = c - (c-b)/2 * ((c-2b+a) + (c-a)) / (c-2b+a)   = c - (c-b)^2 / (c-2b+a)   = Aitken(a, b, c) Thanks, that explains it. I misunderstood the extrapolation was for S2, not the infinite alternating series (which couldn't be much of an improvement). Note that in general one should run the summation loop for S2 in Euler_transform backwards to accumulate fewer roundoff errors because the summand terms t[i] diminish quickly. Though I can't see how it really improves the final S that is already an approximation. - Rob "I count on old friends" -- HP 71B,Prime|Ti VOY200,Nspire CXII CAS|Casio fx-CG50...|Sharp PC-G850,E500,2500,1500,14xx,13xx,12xx... 03-02-2022, 03:17 PM Post: #9 Thomas Klemm Senior Member Posts: 1,804 Joined: Dec 2013 RE: (34C) (11C) Summation of Infinite Alternating Series This has been here for a while: (11C) Summation of infinite, alternating series « Next Oldest | Next Newest » User(s) browsing this thread: 1 Guest(s)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.490526407957077, "perplexity": 15040.49284038093}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499744.74/warc/CC-MAIN-20230129144110-20230129174110-00867.warc.gz"}
https://wiki.geogebra.org/s/en/index.php?title=SampleSD_Command&diff=next&oldid=11679
# Difference between revisions of "SampleSD Command" ##### Command Categories (All commands) SampleSD[List of Numbers] Returns sample standard deviation of given list of numbers. Example: SampleSD[{1,2,3}] yields 1. ## CAS Syntax SampleSD[List of Numbers] Returns sample standard deviation of given list of numbers. If the list contains undefined variables, it yields a formula for the sample standard deviation. Example: SampleSD[{1,2,a}] yields \frac{\sqrt{a²-3a+3}}{\sqrt{3}}. • GeoGebra • Help • Partners
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8943498730659485, "perplexity": 12477.172607507538}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780056578.5/warc/CC-MAIN-20210918214805-20210919004805-00044.warc.gz"}
https://twodee.org/blog/16739
# teaching machines ## The Midiometer This post is part of a series of notes and exercises for a summer camp on making musical instruments with Arduino and Pure Data. ### Waves vs. MIDI When we generate sound by altering the frequencies of waves, we are operating at a pretty low level. Most musicians don’t think about physical laws like we have been doing. They think about notes and timing and volume. So that we can operate at a higher level, we will abandon the wave representation of sound. Instead, we will consider the MIDI representation. MIDI was designed in the early 1980s by several companies that made electronic musical instruments. They wanted to agree on a standard for representing music. Before MIDI, each company had their own way of representing music that was incompatible with the other companies. Trying to connect instruments together was painful. MIDI defined a universal language that all instruments could “speak,” making it much simpler for one device to communicate with another. Each note in a MIDI representation is by two numbers. First is the note number. The note numbers for the standard musical notes are shown in the following table: octave C C♯ D E♭ E F F♯ G A♭ A B♭ B 0 21 22 23 1 24 25 26 27 28 29 30 31 32 33 34 35 2 36 37 38 39 40 41 42 43 44 45 46 47 3 48 49 50 51 52 53 54 55 56 57 58 59 4 60 61 62 63 64 65 66 67 68 69 70 71 5 72 73 74 75 76 77 78 79 80 81 82 83 6 84 85 86 87 88 89 90 91 92 93 94 95 7 96 97 98 99 100 101 102 103 104 105 106 107 Check that you can interpret this table with these questions: • What is the note number for C in the 7th octave? It’s 96. • What is the note number for D in the 4th octave? It’s 62. • What is the note number for F♯ in the 5th octave? It’s 78. The second number is the velocity, which determines how loud the note is. 0 means silent, and 127 is as loud as can be. That’s all the background we need to start generating MIDI notes. Now we’re going to rework our tiangle to use MIDI instead of physical waves. Yes. ### Midiometer To produce this new instrument, which we’ll call the midiometer, start by copying your Pure Data patch for the tiangle instrument into a new patch. Tweak it to look like this: The makenote object wants three pieces of information: 1. The note number, which comes from the Arduino. 2. The velocity, which is fixed at 127. 3. The length of each note in milliseconds. (There are 1000 milliseconds in a second.) It sends the note along to the noteout object, which sends a message to whatever MIDI instrument is listening. Pure Data itself doesn’t play MIDI sounds; we need a MIDI instrument. For that, we will use a program called Virtual MIDI Piano Keyboard (VMPK). Follow these steps to get the programs talking to each other: • Hit Command-Space and open Audio MIDI Setup. • Click Window / Show MIDI Studio. • In the MIDI Studio window, double-click on IAC Driver. • Check Device is online. • Quit Audio MIDI Setup. • Hit Command-Space and open VMPK. • Click Edit / MIDI Connections. • Set MIDI IN Driver to Core MIDI. • Check Enable MIDI Thru on MIDI Output. • Set Input MIDI Connection to IAC Driver Bus 1. • Leave VMPK open. • Back in Pure Data, click Media / MIDI Settings. • Set the Output Device to IAC Driver Bus 1. • Leave Edit Mode and crank the potentiometer. You should hear something that sounds like a piano. But something probably went wrong. Feel free to ask for help! Recall that our Arduino spits out numbers in [0, 255]. But the MIDI note numbers are limited to [21, 107]. Let’s remap the potentiometer readings to the note number range. Hit close 1 in Pure Data. In your Arduino sketch, you probably have some code that looks like this: int reading = analogRead(A0); Map the potentiometer’s raw [0, 1023] range to [21, 107] with the map function: int reading = analogRead(A0); 3. Add a number box and connect it to the middle inlet of makenote. Leave Edit Mode and drag on it. What does it do?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2772478759288788, "perplexity": 1619.6257013115226}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247495001.53/warc/CC-MAIN-20190220125916-20190220151916-00284.warc.gz"}
https://matheducators.stackexchange.com/questions/11368/negative-denominator-in-fractions-importance-and-applications
# Negative Denominator in Fractions; Importance and Applications Why do we need fractions such as $\dfrac{3}{-5}$? I need a convincing answer suitable for 8th grade students. Here's what I've already thought of (which don't fully satisfy me!): • Solving the equation $-5x+3=6$, we get to $x=\dfrac{3}{-5}$. • In the formula relating distances of image and object from the mirror and the its focal length, sometimes we need to put negative numbers for $f$ or $q$: $$\dfrac{1}{p}+\dfrac{1}{q}=\dfrac{1}{f}$$ do you have any other suggestion? • Since 3/-5 is -3/5, I don't see how we really need them at all. It is ultimately a matter of convention if you want to allow or forbid fractions to be written with a negative sign in the denominator. Having said that, there is no good reason to forbid it and is clearly more flexible to allow it. – John Coleman Sep 2 '16 at 18:56 • @JohnColeman: with the same argument, one never really need $3/6$ nor $\sqrt{4}$. Rejecting non-normalized expressions is deeply flawed: we have to work with them, and teach to simplify (or normalize) them. The first given argument in the question seems very compelling to me. – Benoît Kloeckner Sep 5 '16 at 19:34 • @BenoîtKloeckner Of course you are correct -- but I was addressing the part of the question when OP asked why we needed expressions with a negative denominator. Of course we don't need them (or things like 3/6 for that matter), but notation should be geared to convenience rather than strict necessity. – John Coleman Sep 5 '16 at 23:44 • @JohnColeman: well, if we are lead to write these expressions at some point, then we do need them at that point. I don't quite get how it could be otherwise, even if we decide that the ultimate result of a computation should avoid them. – Benoît Kloeckner Sep 6 '16 at 15:19 I would say you're doing your student a disservice if you were to seriously disallow a negative denominator. A fraction is simply a ratio of two integers (where the denominator is not allowed to be zero). I disagree with @yoniLavi that we never need such fractions. Since division by negative numbers makes sense, such a fraction with a negative denominator also does. (Though without any further context, I would generally say it's bad form to leave the negative in the denominator; however, in certain real world applications, it might be more desirable to leave it there in the denominator.) Such fractions come up all the time in algebra and calculus. For example, in the definition of slope. We just say given $(x_1,y_1)$ and $(x_2,y_2)$, then $$m= \frac{y_2 - y_1}{x_2 - x_1}$$ we don't define it as $$m= \frac{y_2 - y_1}{x_2 - x_1} \text{ }\text{ if x_2-x_1 > 0}$$ and $$m= \frac{y_1 - y_2}{x_1 - x_2} \text{ }\text{ if x_2-x_1 < 0}$$ Moreover, in application problems, a negative number can easily arise in the denominator. As for how to convince an 8th grader? Well, I would start by considering the student's background, interests, and (mathematical) proclivities. If the student has a high interest in the mathematical motivation for allowing a negative denominator, I would start with what I said in the first paragraph above. Regardless of whether the student had a strong interest in mathematics, I would also give a concrete example that the student could relate to. I would remind the student that fractions can be used not only to express ratios as in the comparison on lengths (like the relative lengths of the sides of a triangle--they'll see this as soon as they're in high school geometry), but that fractions can also be used to measure rates of change: how one quantity varies with respect to the variation of another quantity. This is where knowing something about the student's interests and background are relevant. Come up with something the student is interested in where you're comparing the changes of two quantities and where the second quantity can decrease (ie., it's negative). I would think that that should be enough. Having worked with the tail-end of this age group quite extensively the past 8 years, I would be sure to feed their curiosity, congratulate them on the interesting question, and perhaps encourage them to come up with a real life application which makes sense to them where a negative denominator makes sense. • The example of the slope calculation perfectly illustrates why fractions with a negative denominator are needed: they naturally arise in the middle of calculations, even if at the end you can rewrite the result so denominators are not negative. And actually you can't always be sure about the denominator's sign. In algebra we may want to consider the function $1/(x^2-1)$, or $1/(x^2-a^2)$ for all $x$, and sometimes the denominator is negative. Or in trigonometry consider $\tan x = (\sin x)/(\cos x)$. For many $x$-intervals the denominator is negative. How awkward to try to avoid that! – KCd Sep 4 '16 at 16:39 There are numbers that one (or many) may never use, is this a reason to eliminate them. Is there a reason to have 1 to any non-negative integer power? When responding to questions of this type "Why do we need fractions such as...? " It is not a question of popularity of use but is a question of "Does this numeral name a well define number"? Then depending upon the the response you may want to elaborate on expected or agreed form. Do we need a VII or are we forced to use '7' or like the Maya :| Give the student the definition of a fraction and ask "does this example follow the definition. • The point is, if we never ever encounter a weird fraction with negative denominator, why should we even think about them? – Behzad Sep 6 '16 at 4:18 • One reason is because the student asked about it. If the student had asked why do we need 18/12 we never use it. You would respond, yes we do. It has meaning but is most often seen as 3/2 or 1.5. The same type of response should be given to the question re. 3/-5. Sometimes we write it as -3/5 It also asks the question, "What is 3 divided by -5? – Sid Hollander Sep 7 '16 at 15:24 • As a matter of fact, there's a straightforward answer to the question " why do we need $\dfrac{18}{12}$?": we need it to be able to calculate $\dfrac{3}{2}+\dfrac{1}{12}$. – Behzad Sep 7 '16 at 15:38 • Some might say 3/2 + 1/12 = 1.5 + 0.08.3333333333333. – Sid Hollander Sep 8 '16 at 20:42 • @Behzad There's nothing to calculate with $\frac{3}{2}+\frac{1}{12}$; it's already a well-defined and exact number. Your point is well taken, questions about negative denominators are much like questions about unreduced fractions. These things exist and we should know how to work with them and perhaps even understand their value. Reduced isn't always better. Slice a pizza into 16 pieces, there's definitely a sense in which it's much easier to manage to eat $\frac{8}{16}$ of that pizza than $\frac{1}{2}$ of a pizza only sliced in half. Same quantity of food, different approach to eating. – A.Ellett Sep 30 '16 at 1:33 I'm actually on the student's side here and would say that we never need fractions such as $\dfrac{3}{-5}$. Like @JohnColeman said, it is more flexible to allow examples such as your first one, but I don't remember seeing anyone use these beyond middle school, except to demonstrate a specific point in a calculation. And I think your second example is misleading, since it's confusing two issues. When we have a rational algebraic expression, the denominator can surely turn out negative, but when you plug in the numbers and write it as a numeric fraction, you wouldn't usually write it with the minus in the denominator, just like you wouldn't usually write the fraction in a non-reduced form. So, to sum up, in my opinion, there is never a need for these fractions except to direct the readers' gaze to something meta-mathematical. I think a similar example is with writing a decimal with trailing zeros in the sciences (e.g. 13.20mL), where the trailing zero has no mathematical meaning, but is used to demonstrate the precision of the measurement. • As an applied mathematician who's taught at the high school and college level, and tutored students from 7th grade through graduate school, the point of a trailing zero is noteworthy (and of mathematical interest). If we write $13.2mL$ this could be the result of rounding $13.24mL$ to one decimal place. But, $13.20mL$ could never be the result of such rounding. And, in high school and earlier, this is an important matter for the students to be able to comprehend. I have asked on quizzes for my students to explain why $0.5$ (inexact answer) and $1/2$ (exact answer) may not be equal. – A.Ellett Sep 2 '16 at 22:16 • @A.Ellett, I think that your comment is actually saying exactly what I meant but with a disagreement of semantics. I think we both agree that 0.5 and 1/2 are the same exact number, but the measurements they represent might be different - you say that the difference is of "mathematical interest", while I consider it to be only of meta-mathematical interest. – yoniLavi Sep 3 '16 at 20:53 • (Part 1) Not to quibble, but I disagree that $0.5$ and $1/2$ are the same exact number. Firstly, a decimal is never an exact numerical representation for the simple reason that we do not know how we got it, whether it was rounded or computed. A fraction is an exact number; its value is unquestionable. Philosophically I'm a bit of a constructivist. While that may seem extremely ivory tower, it's quite relevant in the context of students who think decimals and the numbers their calculators spit out are "exact values", which they aren't. – A.Ellett Sep 4 '16 at 4:29 • (Part 2) The "mathematical interest" is in the context of high school where students should be acquiring an understanding of the significance of the different ways we represent numerical quantities. Students struggle to understand that a letter used to represent a given quantity is not therefore a variable, but a constant. Students often fail to grasp that, if $k$ is a fixed value, then $k^2+2k+1$ is not an expression of unknown value which still must be determined, rather it's a well-defined number. – A.Ellett Sep 4 '16 at 4:49
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8819946646690369, "perplexity": 457.79510404086363}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623488544264.91/warc/CC-MAIN-20210623225535-20210624015535-00490.warc.gz"}
http://mathforum.org/mathimages/index.php/Talk:Parametric_Equations
# Talk:Parametric Equations ## Contents I think that the same applet is on this page in two different places. Nordhr 16:56 30 June 2011 I really like the pictures on this page. I think that they look a lot nicer than if they had been made in a math modeling program like matlab. One thing I wasn't sure about was the red explanitory text in the middle of the more mathematical explanation. Is that supposed to be there? Nordhr 18:17 27 June 2011 Smaurer1 6/21/11. Since there are comments from 2009 (on the original work by Brendan John) and 2011 (modifications by Dayo), it is really important to put a full date on all comments. I just had to study the history really hard to see what was what. I have added some information just now so others don't have to study the history page. Fortunately, all the old comments are at the bottom, beginning with Abram's comments of 7/10/09. Also, Dayo, this page is the place to raise the detailed questions you have been putting on S11. • (Becky writes, from mid June 2011) I think your section is a very good addition to the page! • Nice use of images and short paragraphs- it makes the page much less intimidating. • "parametric equations" should not be capitalized every single time it appears on the page. • I think that if you go into more depth in the basic description, the rest of the page will be less complicated. • I think there's a lot more that can be done here. I know you're handwriting a bunch of stuff, but it'd be cool to see it up online! Richard 6/17 ## Basic Description • Define function with a mouseover instead of parameter. parameter is more important to the page and function is more of an outside-knowledge kind of topic • Delete "including(but not limited to) conic sections and spheres." Parametric equations can really define anything. • I think you might even want to go more basic fr the basic description. I think of this section as the place to "spell it out" without math. I'd go into what a parameter is and what it represents, and include a description of why parametric equations are helpful/maybe even when to use them. Nowhere do you say that a parametric equation puts a function or relation in terms of an entirely new "variables/s" or parameter. Richard 6/17 ## Parametrized Curves • I think its pretty standard to define sine like this instead. Adding more symbols makes me have to think. $\sin (t) = \frac{\text{opposite}}{\text{hypotenuse}}$ Richard 6/17 • I think you should briefly explain why circles cannot be made by cartesian equations • "If a parameter (t) is used like to represent an..." you don't need the "like" • I think your first paragraph is a little unclear. It would be helpful to emphasize rather than defining y in terms of an x value (like a familiar function y=f(x)), you are defining both x and y but some third variable, t. This is why parametric equations are in the form x= f(t) and y= h(t). • Rather than saying that a point t represents a physical quantity in space (which I think is a little vague), I would say something like "Each value of t defines a point, (x,y) = (f(t), h(t)). The points that are generated by plugging in all possible values of t and plotting all the coordinates (x,y) that are produced gives us the graph of the parametric equations." Then you can talk about the specific case of the circle like you do. Clarification: You should do this where you say "a point t represents a physical quantity in space". Becky 6/20 • "the components of the vector that" should have the first "the" capitalized. Same with the next sentence: "using trigonometric ratios." • I don't think you can just say sin(t) = opposite/hypotenuse, rather than using the angle t. Same for the cosine equations. Rebecca 02:04, 21 June 2011 (UTC) I was saying that I think it should be $\sin (t) = \frac{opposite (t)}{hypotenuse} = \frac{y}{1}$ not $\sin (t) = \frac{opposite \angle (t)}{hypotenuse} = \frac{y}{1}$ I think Richard is saying the same thing. • You don't need the arrow before the equation y = rsin(t). Also, you haven't introduced r at all. I'm assuming it is the radius, but thus far you've just used 1 as the radius, so I would be consistent and just say y = sin(t). • Same for the next part- you don't need an arrow and you can remove the r. • For the sentence "thus, t generates physical points..." you should capitalize "thus." Also, the sentence is too long... try to break it up if you can. ### Abram 7/10/09 Really good page. Everything about the circle should maybe be moved to the more mathematical section. I think we don't want the basic description to ever require trig, although it is fairly self-contained in this intro, so it may not be a huge detail. Also, the word "complex" in "complex function" should be replaced with "complicated" or even dropped completely. ### Chris Taranta 7/5/09 Beautiful image with a nice intro. Basic Description: Sentence 2: the word “depend” is not grammatical. Sentence 3: “equal to an equation of an independent variable.” How about something like: “Some complex functions are best described by having the coordinates be described using an equation of a separate independent variable, known as a parameter.” Sentence 4: “Changing the value…” How about: “Changing the value of the parameter then changes the value of each variable in the equation.” Parametrized Circle: How is it being drawn parametrically, rather than just being drawn? Can you show the math equations simultaneously? Also, the image behaves strangely for me using Safari. It traces ¾ of the cirde, then flashes the full circle three times before untracing the ¾ originally traced. A More Mathematical Explanation: Your first sentence beginning with “Sometimes curves…” is terrific. It’s so good, in fact, that I think it would be better used in the Basic Description section. Parametrized Surfaces and Manifolds: The butterfly curve image is to the right of the description, so it appears to be related to that paragraph. The paragraph belongs below the image to avoid confusion. Separate out the surfaces and manifolds sections. An image showing the sphere would be very helpful. You state that “objects of more than two dimensions will require more than two parameters.” Didn’t you just parametrize a sphere using two parameters? Please clarify. Can you give an example of a manifold? Can you show a visualization? If not, think about what the purpose is for including this. ### Anna 6/26/09 Can you change your formatting around so that when you say that the equations are below, they actually are right below? I have that big picture in the middle. Also, how about saying that it is often very useful to write things in parametric form. Sometimes it's actually counter productive and makes things a lot more complicated.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 3, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7960303425788879, "perplexity": 893.0573438504366}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1410657139314.4/warc/CC-MAIN-20140914011219-00257-ip-10-234-18-248.ec2.internal.warc.gz"}
http://physics.stackexchange.com/users/8110/fabee
# fabee less info reputation 4 bio website location age member for 1 year, 2 months seen Apr 6 at 6:32 profile views 2 # 3 Questions 1 Simulator for electrostatics 1 confusion about finite line charge 0 Is there a material that changes local conductivity # 117 Reputation +5 Simulator for electrostatics +5 confusion about finite line charge This user has not answered any questions # 5 Tags 0 electricity × 3 0 simulation 0 electromagnetism × 2 0 electrostatics 0 software # 10 Accounts Cross Validated 610 rep 18 Mathematics 458 rep 17 TeX - LaTeX 163 rep 6 Area 51 156 rep 2 Personal Productivity 121 rep 3
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9514129161834717, "perplexity": 11386.74973298657}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368701314683/warc/CC-MAIN-20130516104834-00080-ip-10-60-113-184.ec2.internal.warc.gz"}
https://web2.0calc.com/questions/help-please-thanks_2
+0 0 100 2 1) For what value of the constant a does the system of equations below have infinitely many solutions? \begin{align*} 3x + 2y &= 8,\\ 6x &= 2a - 7 - 4y \end{align*} I tried over and over again and don't seem to get the right answer. I got 23/6(or sothing like that) for the first time and now 37/6 ... 2) How many numbers between 1 and 2005 are integer multiples of 3 or 4 but not 12? What trick can I use to solve this quickly? 3) Using the letters X and Y, the following two-letter code words can be formed: XX, XY, YY, YX. Using the letters X, Y, and Z, how many different 3 letter code words can be formed? Same here... Sep 5, 2018 #1 +1 1)  Try a = 23/2. With this value, equation 1 multiplied by 2 is exactly the same as equation 2. 2) Try dividing 2005/3 + 2005/4 - 2005/6. Ignore the fractional parts. 3) If you are allowed to repeat the letters, then you should have: 3^3 = 27 Sep 5, 2018 edited by Guest  Sep 5, 2018 edited by Guest  Sep 5, 2018 #2 +1 Yeah! That was what I got for no.1 and I thought I was wrong... :( THANK SO MUCH! Guest Sep 5, 2018
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9772565364837646, "perplexity": 697.528508750744}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547584328678.85/warc/CC-MAIN-20190123085337-20190123111337-00460.warc.gz"}
https://www.physicsforums.com/threads/phase-transition-problems.709837/
# Phase transition problems 1. Sep 11, 2013 1. The problem statement, all variables and given/known data http://icho2013.chem.msu.ru/materials/IChO-2013_Theoretical_Official_English_Version.pdf Looking at problem 1. 2. Relevant equations The integrated form of the Clausius-Clapeyron equation: $$ln(\frac{P_{T2}}{P_{T1}}) = -\frac{\Delta H}{R} \cdot (\frac{1}{T_2} - \frac{1}{T_1})$$ 3. The attempt at a solution Part 1 is irrelevant, do not need to read. Part 2 is fine, Part 3 is fine. For Part 4, once you correctly guess the temperature, it's ok, because we rely on the same reaction as before, decomposition of methane hydrate into ice, and looking for minimum temperature at which it will not occur; we can use either of the 2 possible P,T sets (one given, one ascertained from Part 3) to calculate the P for Part 4. However, Part 5 causes trouble because only using one of the 3 P,T sets of data we have now can we find the correct T from the P in Part 5; the other two give wrong readings. Why is this, and how should I tell which sets of P,T data can be used for future P, T calculations from the situation and which cannot? 2. Sep 17, 2013 ### DrDu Could you explicitly state the p, T sets you are referring too? Which values did you get? 3. Sep 26, 2013 I'll write out the problem again if that helps: Real methane hydrate has a non-stoichiometric composition close to CH4•6H2O. At atmospheric pressure, methane hydrate decomposes at –81 °C. However, under high pressures (e.g. on the ocean floor) it is stable at much higher temperatures. Decomposition of methane hydrate produces gaseous methane and solid or liquid water depending on temperature. 2. Write down the equation of decomposition of 1 mole of CH4•6H2O producing solid water (ice) H2O(s). The enthalpy of this process equals 17.47 kJ•mol-1. Assume that the enthalpies do not depend on temperature and pressure, the volume change upon decomposition of hydrate is equal to the volume of released methane, and methane is an ideal gas. 3. At what external pressure does decomposition of methane hydrate into methane and ice take place at –5 °C? 4. What is the minimum possible depth of pure liquid water at which methane hydrates can be stable? To answer this question, you should first deduce at which minimum temperature methane hydrate can coexist with liquid water, by choosing from 3 possibilities: 272.9 K, 273.15 K, 273.4 K. Large methane hydrate stocks on the floor of Baikal lake, the largest freshwater lake in Russia and in the world, have been discovered in July 2009 by the crew of a deep-submergence vehicle «Mir-2». During the ascent from the depth of 1400 m methane hydrate samples started to decompose at the depth of 372 m. 5. Determine the temperature in Baikal lake at the depth of 372 m. The enthalpy of fusion of ice is 6.01 kJ•mol-1. ------ 3 is straightforward because it is asking for the conditions required to make a reaction happen, given that the reaction was previously spontaneous at a given p,T of (192.15 K, 101300 Pa). But the confusion starts at 4. I really don't understand the scenario being proposed, and how it relates with the CH4•6H2O -> CH4 + H2O (s) reaction for which the two p,T sets we have (the initial given p,T of 192.15 K, 101300 Pa, and the solution to part 3, at 268.15 K) apply. And the same confusion grows with 5. I think the problem is a fundamental questioning of whether one understands multi-component phase diagrams, how the Clausius-Clapeyron equation can be used for spontaneity conditions in any reactions and with various different conditions (e.g. 4 proposes that liquid water must also be present), etc. Any help on these topics and understanding them better would be much appreciated. (I have seen the solutions, but they do not clear up my haze about how the Clausius-Clapeyron equation is to be applied in a case like this, and how I should understand and analyse each scenario in the problems, which is the sort of understanding I am trying to build.) 4. Sep 30, 2013 ### DrDu I think in 4 what is meant is that the minimum depth where the hydrate is stable is determined in a two step process: For a given depth (=pressure), the hydrate is most stable at the lowest possible temperature, i.e. the temperature where liquid water is stable and does not freeze. So you would have to find the freezing point curve T(P) of water and plug this temperature into the Gibbs Duhem relation in a second step. Apparently the authors of the question did consider this to be too difficult, and let you chose between 3 temperatures. As the volume of water decreases on melting, the melting point decreases with increasing pressure, so the correct T must be 272.9 K (as 273.15 is at normal pressure). In question 5 you are given the depth (=pressure) where decomposition occurs. Apparently, lake Baikal is somewhat warmer than the freezing point at that depth. So the hydrate decomposes into methane and liquid water and you have to take into account melting enthalpy in calculating the change in equilibrium. 5. Oct 5, 2013 An understanding of the rest of the question is beginning to shape in my head but I will save those questions for later. First I'd like to know what the Gibbs Duhem relation is by which we could have calculated 272.9 K? How do we do this - and if all the needed information for predicting this value is given then why would the question-writer not have expected us to do this? 6. Oct 8, 2013 ### DrDu I think the problem is already very hard the way it is formulated, and giving the temperature was meant to make it more tractable. Gibbs-Duhem: ndμ=VdP-SdT Along the solidus line $\mu_l=\mu_s$ hence $V_l/n_l dP-S_l/n_l dT=V_s/n_s dP -S_s/n_s dT$. V/n and S/n are the molar volume and entropy, respectively. We can write this as $\Delta V_m dP=\Delta S_m dT$ with $\Delta S_m =\Delta H/T$, where $\Delta H$ is the enthalpy of melting we get $dT/dP=T\Delta V_m/\Delta H$ which is the change of melting temperature with pressure. As melting costs heat $\Delta H>0$, but volume decreases $\Delta V_m<0$, the melting temperature decreases with pressure. 7. Oct 12, 2013 Is there any way to understand parts 4 and 5 in terms of points of coexistence of various different species on a multi-component phase diagram? I'm still trying to visualize the scenario but it's not really coming together in an obvious way. 8. Oct 12, 2013 ### DrDu Did you find an answer for questions 4 and 5 and are only looking for a deeper insight?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8451226353645325, "perplexity": 819.9924916244047}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187824899.43/warc/CC-MAIN-20171021205648-20171021225648-00439.warc.gz"}
https://www.edaboard.com/threads/virtex-4-sx-development-boards.52553/
# virtex-4 sx development boards Status Not open for further replies. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 netlist is empty after trimming useless hardware the ML402 evaluation platform does not include the development tools and cable. i can find the price for the system generator, usd $995. wat else do i need? the requirement for system generator is ISE v7.1 and MathWorks R14.1/.2/.3 it does not mean the ISE foundation rite? coz it cost usd$2495. the video starter kit has the system generator included, do i need to get the ISE v7.1 or is it included? and what about the cable? the XtremeDSP development board comes with the system generator at a higher price, but does it include the xtremedsp evaluation software, coz it have the ISE v7.1 .. or can i purchase the xtremedsp evaluation software separately for use with the ML402 evaluation platform? https://www.xilinx.com/dsp/eval_software.htm #### echo47 Joined Apr 7, 2002 Messages 3,933 Helped 638 Reputation 1,274 Reaction score 88 Trophy points 1,328 Location USA Activity points 33,176 system generator netlist is empty The Xilinx web pages are confusing, and I know only some of your answers. I have an ML402 and ML403. Neither board included any software or documentation. System Generator (about $1000 US) requires ISE Foundation (or the discontinued ISE Alliance), and it's not included. ISE Foundation costs about$2500 US. If you also want the EDK, that's another $500 US. I don't use System Generator very much. I mostly use only ISE Foundation. I don't know what is xtremedsp evaluation software. I use the Xilinx Parallel Cable IV (about$100 US). It works with the ML40x boards, and the Spartan-3 Starter Kit, and probably most other Xilinx boards: ISE Foundation (or the free ISE WebPACK) includes iMPACT, the cable download software. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 thanks echo. really cleared my doubts alot. btw, im doing at matlab level and not rtl level so i need the system generator. the website also said that ISE alliance is sufficient, but there dun seem be a link to direct me to purchase the alliance seprately. the first link stated that ise webpack do not support system generator but the 2nd link said that ise webpack 8.1i support it. lol .. think i have to wait for the release, but not sure how soon. both system generator and ise foundation for evaluation comes with full features rite, or the evaluation comes with limited features? that way i can drag some time for ise webpack 8.1i and its free. ^.^ #### samcheetah Joined May 25, 2004 Messages 645 Helped 39 Reputation 78 Reaction score 10 Trophy points 1,298 Location Pakistan Activity points 6,916 echo47 said: I don't use System Generator very much. I mostly use only ISE Foundation do you think that sysgen is a useful tool. does it really simplify the MATLAB->RTL conversion process. and the sysgen page at xilinx.com says that it has the microblaze core. so if i have sysgen, will that mean that i dont need EDK? #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 i've juz called the person in charge. he told me that system generator cannot convert m-codes to hdl. but here is a link that xilinx provide. does this mean i need to create my own m-code blocks in order for system generator to use? https://www.xilinx.com/products/software/sysgen/app_docs/user_guide_Chapter_7_Section_3.htm juz found another link. does that mean if i wan to convert m-codes directly to FPGA, i need accelchip dsp synthesis tool also? #### echo47 Joined Apr 7, 2002 Messages 3,933 Helped 638 Reputation 1,274 Reaction score 88 Trophy points 1,328 Location USA Activity points 33,176 You did the right thing by asking your Xilinx sales rep. Their web page info is too skimpy. System Generator doesn't compile M code. Instead, you draw a Simulink block diagram using the Xilinx blockset, and then press a button to generate a bitstream that you can download to your FPGA. That works fine if you use only high-level signal processing blocks (I think each block simply calls Xilinx CORE Generator), but I found it excruciatingly painful for creating sequencers and other miscellaneous logic. It's too much like schematic capture. A played with it for a day or two, and then went back to using HDL. System Generator includes a MicroBlaze block symbol, but it doesn't seem to include any EDK functionality. That makes sense, because System Generator is about 20 megabytes, and EDK is gigabytes. I haven't tried AccelChip, but I'm suspicious of any tool that claims to convert a high-level language to FPGA. The results are usually greatly inferior to hand-coded HDL. However, that may be acceptable if you have a small project and a big fast FPGA. Xilinx discontinued ISE Alliance about a year ago. Customers with support agreements were automatically upgraded to ISE Foundation. #### samcheetah Joined May 25, 2004 Messages 645 Helped 39 Reputation 78 Reaction score 10 Trophy points 1,298 Location Pakistan Activity points 6,916 echo47 said: System Generator doesn't compile M code. Instead, you draw a Simulink block diagram using the Xilinx blockset, and then press a button to generate a bitstream that you can download to your FPGA. That works fine if you use only high-level signal processing blocks (I think each block simply calls Xilinx CORE Generator), but I found it excruciatingly painful for creating sequencers and other miscellaneous logic. It's too much like schematic capture. A played with it for a day or two, and then went back to using HDL. im shocked!!!!!!!!!!!! thanx for the info echo47 by the way, what design flow do you follow?? MATLAB -> C++ -> HDL or MATLAB -> HDL i heard someone say that the first design flow is better but i dont see any benefit of first converting it into C++ and then into HDL. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 found that my boss's boss have the ise foundation 7.1i, -,-" .. now trying out the system generator evaluation software. they have a mcode block which uses the function in m-file. now im only trying out one function in one m-file, so my m-code block have the same i/o as my function. im not sure wat would happen if there are multiple function in one m-file. now i need to input my dummy code into the fpga and see if i would get the same output .. thanks echo .. hehe .. need more help will post here .. ^.^ good luck to u samcheetah .. #### echo47 Joined Apr 7, 2002 Messages 3,933 Helped 638 Reputation 1,274 Reaction score 88 Trophy points 1,328 Location USA Activity points 33,176 System Generator is not my cup of tea, but some folks like it. The evaluation is free, so give it a try. I consider it one tool in my toolbox. Someday it may save me a day of work, and it will pay for itself. When I design signal processing systems, I do the high-level concept simulation in MATLAB m-code or C. MATLAB is easier unless I'm doing a lot of tricky binary manipulations, and that's easier in C. I don't use Simulink. My high-level simulation also spits out coefficient tables (in HDL format) for initializing block RAMs. Finally, I code the whole system in Verilog, test it with ModelSim, and compile to FPGA. chinsin83, please let us know if that System Generator m-code block does anything useful for you. I guess I overlooked that key feature. Here's the online help introduction: The Xilinx MCode block is a container for executing a user-supplied MATLAB function within Simulink. A parameter on the block specifies the m-code function name. The block executes the m-code to calculate block outputs during a Simulink simulation. The same code is translated in a straightforward way into equivalent behavioral VHDL when hardware is generated. The block's Simulink interface is derived from the MATLAB function signature, and from block mask parameters. There is one input port for each parameter to the function, and one output port for each value the function returns. Port names and ordering correspond to the names and ordering of parameters and return values. The MCode block does not support the entire MATLAB language; the supported subset is described below. There are several noteworthy additional restrictions on the block and its usage: - All block inputs and outputs must be of Xilinx fixed point type. - The block must have at least one output port. - The code for the block must exist on the MATLAB path or in the same directory as the model that uses the block. This block provides a convenient and flexible way to implement arithmetic functions and to build finite state machines and control logic. The MCode block tutorial shows three examples of functions for the MCode block. The first example (also described below) consists of a function xlmax which returns the maximum of its inputs. The second illustrates how to do simple arithmetic. The third shows how to build finite state machines. The example models are linked from the Examples section of this guide. <snip> #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 now i need to find the plug-in for the XESS XAS-3S1000 board which the XESS website does not provide .. duhz ... any other way of installing my board for the hardware simulation? #### samcheetah Joined May 25, 2004 Messages 645 Helped 39 Reputation 78 Reaction score 10 Trophy points 1,298 Location Pakistan Activity points 6,916 chinsin83 said: now i need to find the plug-in for the XESS XAS-3S1000 board which the XESS website does not provide .. duhz ... any other way of installing my board for the hardware simulation? are you talking about downloading the code to the FPGA??? well, i think the XSA board requires a tool that is available on their site. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 i think i read the manual wrongly .. now not really sure wat i wan .. going to read the manual again .. =( Added after 1 hours 29 minutes: i've used the System Generator Board Description Builder to create the plugin for my XESS XSA-3S1000 board .. when i used system generator to generate the block for simulink .. there's a error .. i think i specified the wrong system clock .. any1 have experienced with this board n can tell me which is the pin for the sytem clock? #### samcheetah Joined May 25, 2004 Messages 645 Helped 39 Reputation 78 Reaction score 10 Trophy points 1,298 Location Pakistan Activity points 6,916 i think you can easily find information about the clock from the manual. the XTAL should be connected to one of the GCKx pins. its always good to read the manual carefully. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 hmm .. did a simple arithmetic function constant input ----> HW co-sim block: input * 10 , output/ 2 ----> output displayed hmm .. i run the simulation n it took like 10 secs for the display in my simulation to show me the value .. is it coz my board is lousy? -,- .. btw .. how do i display an image if my board is connect to a monitor? .. do i need to add in the pins for the vga port in my plug-in? #### samcheetah Joined May 25, 2004 Messages 645 Helped 39 Reputation 78 Reaction score 10 Trophy points 1,298 Location Pakistan Activity points 6,916 the simulation might be slow due to two reasons; one is that your computer is slow and the other is that your approach is not correct. your board being lousy is not an issue because you dont simulate on the board. as for the vga. search on google with the keywords vga and fpga and you will get alot of useful info. and you should check the vga core at opencores.org #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 how do i calculate how much time the algorithms run in the fpga? .. #### FPGAs ##### Member level 1 Joined Dec 22, 2004 Messages 41 Helped 3 Reputation 6 Reaction score 3 Trophy points 1,288 Location Shanghai, China Activity points 306 i think the xilinx FPGA boards aren't worth so much #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 er .. huh? .. system generator for dsp 7.1i can support only some features of matlab functions, =( .. it cant even support 'for loop' .. does any1 know how to include simulnk's blockset inside the configurable subsystem for system generator to import as configurable subsystem for the hw-co-simulation? .. the xilinx block sets only accepts xilinx input .. so u have to use the gateway_in; but system generator cannot compile the gateway_in to be included in the hw-co-simulation. #### chinsin83 ##### Junior Member level 2 Joined Dec 22, 2005 Messages 20 Helped 0 Reputation 0 Reaction score 0 Trophy points 1,281 Activity points 1,539 hihi .. im back with new question .. system generator's gateway cannot take in frame-based? .. then how do i simulate my data in hardware?.. if i din use xilinx's blockset .. can i still use the hardware co-simulation? .. coz when i click generate, theres an error specifying netlist is empty after trimming useless hardware. wat does that mean? Status Not open for further replies.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.39730992913246155, "perplexity": 5328.510116053623}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573197.34/warc/CC-MAIN-20220818124424-20220818154424-00071.warc.gz"}
https://arbital.greaterwrong.com/p/alternating_group_five_conjugacy_classes?l=4kr
# Conjugacy classes of the alternating group on five elements This page lists the con­ju­gacy classes of the al­ter­nat­ing group $$A_5$$ on five el­e­ments. See a differ­ent lens for a deriva­tion of this re­sult us­ing less the­ory. $$A_5$$ has size $$5!/2 = 60$$, where the ex­cla­ma­tion mark de­notes the fac­to­rial func­tion. We will as­sume ac­cess to the con­ju­gacy class table of $$S_5$$ the sym­met­ric group on five el­e­ments; $$A_5$$ is a quo­tient of $$S_5$$ by the sign ho­mo­mor­phism. We have that a con­ju­gacy class splits if and only if its cy­cle type is all odd, all dis­tinct. (Proof.) This makes the clas­sifi­ca­tion of con­ju­gacy classes very easy. # The table We must re­move all the lines of $$S_5$$’s table which cor­re­spond to odd per­mu­ta­tions (that is, those which are the product of odd-many trans­po­si­tions). In­deed, those lines are classes which are not even in $$A_5$$. We are left with cy­cle types $$(5)$$, $$(3, 1, 1)$$, $$(2, 2, 1)$$, $$(1,1,1,1,1)$$. Only the $$(5)$$ cy­cle type can split into two, by the split­ting con­di­tion. It splits into the class con­tain­ing $$(12345)$$ and the class which is $$(12345)$$ con­ju­gated by odd per­mu­ta­tions in $$S_5$$. A rep­re­sen­ta­tive for that lat­ter class is $$(12)(12345)(12)^{-1} = (21345)$$. $$\begin{array}{|c|c|c|c|} \hline \text{Representative}& \text{Size of class} & \text{Cycle type} & \text{Order of element} \\ \hline (12345) & 12 & 5 & 5 \\ \hline (21345) & 12 & 5 & 5 \\ \hline (123) & 20 & 3,1,1 & 3 \\ \hline (12)(34) & 15 & 2,2,1 & 2 \\ \hline e & 1 & 1,1,1,1,1 & 1 \\ \hline \end{array}$$ Children: Parents: • Alternating group The al­ter­nat­ing group is the only nor­mal sub­group of the sym­met­ric group (on five or more gen­er­a­tors).
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 2, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5833767652511597, "perplexity": 6199.643758368009}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875145941.55/warc/CC-MAIN-20200224102135-20200224132135-00059.warc.gz"}
https://socratic.org/questions/58ac3a117c01493875b11e6e
Chemistry Topics # Question 11e6e Feb 22, 2017 ${43}^{\circ} \text{C}$ #### Explanation: The idea here is that you can assume that the heat lost by the iron bar as it cools will be equal to the heat gained by the water as it warms. $\textcolor{b l u e}{\underline{\textcolor{b l a c k}{- {q}_{\text{iron" = q_"water}}}}}$ The minus sign is used here because heat lost is, by definition, negative. Your tool of choice here will be the equation $\textcolor{b l u e}{\underline{\textcolor{b l a c k}{q = m \cdot c \cdot \Delta T}}}$ Here • $q$ is the heat lost or gained by the substance • $m$ is the mass of the sample • $c$ is the specific heat of the substance • $\Delta T$ is the change in temperature, defined as the difference between the final temperature and the initial temperature of the sample You should know that the specific heat of water is ${c}_{\text{water" = "1 cal g"^(-1)""^@"C}}^{- 1}$ If you take ${T}_{f}$ to be the final temperature reached by the iron bar + water system, you can say that you have • For the iron bar q_"iron" = 45.0 color(red)(cancel(color(black)("g"))) * "0.11 cal" color(red)(cancel(color(black)("g"^(-1)))) color(red)(cancel(color(black)(""^@"C"^(-1)))) * (T_f - 100)color(red)(cancel(color(black)(""^@"C"))) ${q}_{\text{iron" = 4.95 * (T_f - 100)" cal}}$ • For the water q_"water" = 90.0 color(red)(cancel(color(black)("g"))) * "1 cal" color(red)(cancel(color(black)("g"^(-1)))) color(red)(cancel(color(black)(""^@"C"^(-1)))) * (T_f - 40)color(red)(cancel(color(black)(""^@"C")))# ${q}_{\text{water" = 90.0 * (T_f - 40)" cal}}$ Since the heat lost by the iron bar must be equal to heat gained by the water, you will have -- do not forget about the minus sign used for the heat lost by the iron bar! $- 4.95 \cdot \left({T}_{f} - 100\right) \textcolor{red}{\cancel{\textcolor{b l a c k}{\text{cal"))) = 90.0 * (T_f - 40)color(red)(cancel(color(black)("cal}}}}$ This will get you $- 4.95 \cdot {T}_{f} + 495 = 90.0 \cdot {T}_{f} - 3600$ $94.95 \cdot {T}_{f} = 4095 \implies {T}_{f} = \frac{4095}{94.95} = 43.13$ Therefore, you can say that the iron bar will cool to a final temperature of $\textcolor{\mathrm{da} r k g r e e n}{\underline{\textcolor{b l a c k}{{T}_{f} = {43}^{\circ} \text{C}}}}$ I'll leave the answer rounded to two sig figs, but keep in mind that you only have one significant figure for the starting temperatures of the iron bar and of the water. ##### Impact of this question 279 views around the world
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 17, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7671588063240051, "perplexity": 556.3855150903266}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590348513321.91/warc/CC-MAIN-20200606124655-20200606154655-00477.warc.gz"}
http://efc.2015.uob.bram.xyz/Exercises/week05p.html
# Week 5 programming: lists¶ Subsite home page: http://bit.ly/focs-aux ## A: Head and tail, first and second¶ You can get the head and the tail of a list with functions hd and tl. In OCaml/utop, you first have to type open List;; to get access to these functions. You can get the first and the second part of a pair with the functions fst and snd. This only works for 2-tuples (e.g. (3, "abc")), not smaller tuples (like (), (2, 3, 4), ...) . • Can you find the head and the tail of the terms from last week, if they exist? • Can you find the first and the second component of any pairs in those terms? Check your answers with ocaml/utop. ## B: Cons exercises¶ Which of these terms is valid? If it is valid, what is its value, and what is its type? Check your answers with ocaml/utop. 1. 3 :: [] 2. 3 :: 4 :: [] 3. 3 :: [4] 4. [3] :: [] 5. [3] :: 4 6. [] :: [] 7. [] :: [[]] 8. [[]] :: [] 9. [[]] :: [[]] 10. [] :: [], [] 11. [] :: []; [] 12. [[] :: []; []] (Note: this is very different!) ## C: List exercises¶ Do the exercises from https://ocaml.org/learn/tutorials/99problems.html; solutions are provided. Additional hints given on Canvas: Some of the question use the OCaml option type. This type is used to return a special value (None) when a function cannot be computed, instead of throwing an exception. For example, the standard hd function is: let hd = function | [] -> failwith "hd" | x :: _ -> x and has type val hd : 'a list -> 'a = <fun> If you want to avoid throwing a failure the alternative is this: let hd = function | [] -> None | x :: _ -> Some x and has type val hd : 'a list -> 'a option = <fun> For the purpose of this assignment you can ignore the option type and use failwith instead of None. For a longer discussion of option types read this article: https://blogs.janestreet.com/making-something-out-of-nothing-or-why-none-is-better-than-nan-and-null/ See also: the assignment on Canvas We did some exercises without option. I have here filled in our solutions. 1. Make a function last that finds the last element of a list. We want that: last [3; 5] = 5 last [3] = 3 last [] = an error (Now you can click on the button below; there is an answer.) (answer goes here) We are making a function on lists, so probably we’ll need recursion. (Use let rec.) We can make a first attempt by just writing down these cases. Last of a two-element list should return the second element, and so forth. In the “otherwise” case (_), we wish to give an error. let rec last l = match l with | [x; y] -> y | [x] -> x | [] -> failwith "last" | _ -> failwith "dunno" We can make the last case more specific, because the list will not be empty: let rec last l = match l with | [x; y] -> y | [x] -> x | [] -> failwith "last" | x::xs -> failwith "dunno" We can not list all the possible list length, because a list can be arbitrarily long. So we need to use recursion. Usually, structural recursion is enough. What’s the relation between last [2; 3; 4] and last of its tail, last [3; 4]? They are equal. So we write: let rec last l = match l with | [x; y] -> y | [x] -> x | [] -> failwith "last" | x::xs -> last (tl l) We simplify this in two ways. Firstly, in the last case, we already have tl l in a variable, namely xs. let rec last l = match l with | [x; y] -> y | [x] -> x | [] -> failwith "last" | x::xs -> last xs We also don’t need to do the special case for two elements. let rec last l = match l with | [x; y] -> y | [x] -> x | [] -> failwith "last" | x::xs -> last xs 2. Make a function pen that finds the penultimate element of a list. Write down some example inputs and outputs like above, then write the solution. (answer goes here) We want pen [] = error pen [2; 3; 4] = 3 pen [2] = error pen [2; 3] = 2 Solution: let rec pen l = match l with | [] -> failwith "pen empty" | [a] -> failwith "pen of one element" | [a; b] -> a | x::xs -> pen xs ;; 3. Make a function pens that finds the last two elements of a list and puts them in a pair. Write down some example inputs and outputs like above, then write the solution. (answer goes here) We want pens [3; 4] = (3, 4) pens [4] = error pens [] = error pens [2; 3; 4] = (3, 4) Solution: let rec pens l = match l with | [] -> failwith "pens empty" | [a] -> failwith "pens of one element" | [a; b] -> (a, b) | x::xs -> pens xs ;; 4. Make a function at that finds the n’th element of a list. We want: at 3 [4; 5; 6; 7; 8] = 8 at 3 [] = error at 0 [4; 5; 6] = error at (-2) [4; 5; 6] = error at 1 [4; 5; 6] = 4 Note that usually when programming, we count from 0 because it tends to make things easier. For some mysterious reason, the exercise designers have chosen to count from 1 here. (answer goes here) let rec at k l = match l with | [] -> failwith "at" | x::xs -> if k < 1 then failwith "at" else if k = 1 then x else at (k-1) xs;; (Note that I forgot to write the if k = 1 then x else in the tutorial, which some students helpfully pointed out.)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4189477562904358, "perplexity": 3756.2871606680237}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943747.51/warc/CC-MAIN-20230321225117-20230322015117-00105.warc.gz"}
https://tug.org/pipermail/macostex-archives/2005-February/013225.html
# [OS X TeX] .dvi files Martin Buchmann Martin.Buchmann at uni-jena.de Fri Feb 4 13:55:57 CET 2005 Lawrence, AFAIK there's only a pdfetex binary on your system and latex is just a wrapper which calls pdfetex with dvi output. pdfetex can also produce dvi output is the \pdfoutput is set to 0. On my system calling latex leads to a dvi file. Maybe there's something wrong with your system. Until you figured out what's wrong a constrcus like this should help you out: \documentclass{article} \usepackage{ifpdf} \ifpdf \pdfoutput=0 \fi \begin{document} Test 1234 \end{document} HTH Martin -- If all the world's a stage, I want to operate the trap door. -- Paul Beatty --------------------- Info --------------------- Mac-TeX Website: http://www.esm.psu.edu/mac-tex/ & FAQ: http://latex.yauh.de/faq/ TeX FAQ: http://www.tex.ac.uk/faq List Post: <mailto:MacOSX-TeX at email.esm.psu.edu>
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9143847227096558, "perplexity": 29510.861156418378}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202450.86/warc/CC-MAIN-20190320190324-20190320212324-00199.warc.gz"}
http://davidlowryduda.com/tag/reflection/
# Tag Archives: reflection ## Math 100 Fall 2016: Concluding Remarks It is that time of year. Classes are over. Campus is emptying. Soon it will be mostly emptiness, snow, and grad students (who of course never leave). I like to take some time to reflect on the course. How did it go? What went well and what didn’t work out? And now that all the numbers are in, we can examine course trends and data. Since numbers are direct and graphs are pretty, let’s look at the numbers first. ## Math 100 grades at a glance Let’s get an understanding of the distribution of grades in the course, all at once. These are classic box plots. The center line of each box denotes the median. The left and right ends of the box indicate the 1st and 3rd quartiles. As a quick reminder, the 1st quartile is the point where 25% of students received that grade or lower. The 3rd quartile is the point where 75% of students received that grade or lower. So within each box lies 50% of the course. Each box has two arms (or “whiskers”) extending out, indicating the other grades of students. Points that are plotted separately are statistical outliers, which means that they are $1.5 \cdot (Q_3 – Q_1)$ higher than $Q_3$ or lower than $Q_1$ (where $Q_1$ denotes the first quartile and $Q_3$ indicates the third quartile).
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8039034008979797, "perplexity": 957.8379507832609}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463607704.68/warc/CC-MAIN-20170523221821-20170524001821-00208.warc.gz"}
https://docs.sympy.org/0.7.6/outreach.html
# SymPy Papers¶ A list of papers which either use or mention SymPy, as well as presentations given about SymPy at conferences can be seen at SymPy Papers. # Planet SymPy¶ We have a blog aggregator at http://planet.sympy.org. # SymPy logos¶ SymPy has a collection of official logos, which can be generated from $$sympy.svg <https://github.com/sympy/sympy/tree/master/doc/src/sympy.svg>$$ in your local copy of SymPy by: $cd doc$ make logo # will be stored in the _build/logo subdirectory # Projects using SymPy¶ This is an (incomplete) list of projects that use SymPy. If you use SymPy in your project, please let us know on our mailinglist, so that we can add your project here as well. • SfePy (simple finite elements in Python) • Quameon (Quantum Monte Carlo in Python) # Blogs, News, Magazines¶ You can see a list of blogs/news/magazines mentioning SymPy (that we know of) on our wiki page.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3597177565097809, "perplexity": 5716.290178952586}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039742968.18/warc/CC-MAIN-20181116004432-20181116030432-00090.warc.gz"}
https://socratic.org/questions/a-solid-disk-with-a-radius-of-7-m-and-mass-of-12-kg-is-rotating-on-a-frictionles-2
Physics Topics # A solid disk with a radius of 7 m and mass of 12 kg is rotating on a frictionless surface. If 15 W of power is used to increase the disk's rate of rotation, what torque is applied when the disk is rotating at 12 Hz? Sep 6, 2017 The torque is $= 0.20 N m$ #### Explanation: The mass of the disc is $m = 12 k g$ The radius of the disc is $r = 7 m$ The power $= P$ and the torque $= \tau$ are related by the following equation $P = \tau \omega$ where $\omega =$ the angular velocity Here, The power is $P = 15 W$ The frequency is $f = 12 H z$ The angular velocity is $\omega = 2 \pi f = 2 \cdot \pi \cdot 12 = 24 \pi r a {\mathrm{ds}}^{-} 1$ Therefore, the torque is $\tau = \frac{P}{\omega} = \frac{15}{24 \pi} = 0.20 N m$ ##### Impact of this question 391 views around the world
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 11, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9693424701690674, "perplexity": 584.928804110238}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780058373.45/warc/CC-MAIN-20210927060117-20210927090117-00026.warc.gz"}
https://mathhelpboards.com/threads/series-convergence-with-a-floor.3031/
# series convergence with a floor #### Lisa91 ##### New member Jan 15, 2013 29 I have one series $$\sum_{n=13}^{\infty}(-1)^{\left\lfloor\frac{n}{13}\right\rfloor} \frac{ \ln(n) }{n \ln(\ln(n)) }$$. How to investigate its convergence? I wanted to group the terms of this series but I don't know whether it's a good idea as we have 13 terms with minus and then 13 with plus and so on. What do you think? #### Klaas van Aarsen ##### MHB Seeker Staff member Mar 5, 2012 8,866 Hi Lisa91! Actually, that does seem like a good idea. If you can proof that the sequence without the (-1) is declining towards zero, then summing 13 consecutive terms and repeating that will also be declining towards zero. #### Fernando Revilla ##### Well-known member MHB Math Helper Jan 29, 2012 661 I have one series $$\sum_{n=13}^{\infty}(-1)^{\left\lfloor\frac{n}{13}\right\rfloor} \frac{ \ln(n) }{n \ln(\ln(n)) }$$. How to investigate its convergence? Hint: Prove that (i) $T_k=\left |\sum_{i=13k}^{13(k+1)}a_i\right|$ is strictly decreasing and has limit $0$. (ii) The sequence $(S_n)_{n\geq 13}$ of the partial sums of the given series is a Cauchy sequence. Last edited: #### chisigma ##### Well-known member Feb 13, 2012 1,704 I have one series $$\sum_{n=13}^{\infty}(-1)^{\left\lfloor\frac{n}{13}\right\rfloor} \frac{ \ln(n) }{n \ln(\ln(n)) }$$. How to investigate its convergence? I wanted to group the terms of this series but I don't know whether it's a good idea as we have 13 terms with minus and then 13 with plus and so on. What do you think? The series is the sum of 13 series and can be written as... $\displaystyle \sum_{k=0}^{12} \sum_{j=0}^{\infty} (-1)^{j}\ \frac{\ln (13\ j + k)}{(13\ j + k)\ \ln \{\ln (13\ j + k)\}}$ (1) Each of the inner series is 'alternating' and converges for the Leibnitz's criterion, so that the whole series converges... Kind regards $\chi$ $\sigma$ #### Lisa91 ##### New member Jan 15, 2013 29 I tried to do it this way but I don't know how to prove that it decreases and that the limit is zero. I tried estimating it RHS and LHS using $$\frac{t}{t+1}< \ln(t+1) < t$$ but in one case I've got -1... $$(-1)^{j}\ \frac{\ln (13\ j + 1)}{(13\ j + 1)\ \ln \{\ln (13\ j + 1)\}}$$. #### Fernando Revilla ##### Well-known member MHB Math Helper Jan 29, 2012 661 The series is the sum of 13 series and can be written as... $\displaystyle \sum_{k=0}^{12} \sum_{j=0}^{\infty} (-1)^{j}\ \frac{\ln (13\ j + k)}{(13\ j + k)\ \ln \{\ln (13\ j + k)\}}$ (1) Each of the inner series is 'alternating' and converges for the Leibnitz's criterion, so that the whole series converges... A priori, you don't know if the series is convergent, so ¿what associativity and/or conmutativity properties are you using? #### chisigma ##### Well-known member Feb 13, 2012 1,704 I tried to do it this way but I don't know how to prove that it decreases and that the limit is zero. I tried estimating it RHS and LHS using $$\frac{t}{t+1}< \ln(t+1) < t$$ but in one case I've got -1... $$(-1)^{j}\ \frac{\ln (13\ j + 1)}{(13\ j + 1)\ \ln \{\ln (13\ j + 1)\}}$$. We pratically can consider the term $\ln \{\ln (13\ j + k)\}$ as a constant term [it is very slowly changing with j...] and can analyse the term... $\displaystyle \frac{\ln (13\ j + k)}{13\ j + k}$ (1) Any doubt about the fact that, at least for j 'large enough', it is decreasing with j and that its limit is 0?... Kind regards $\chi$ $\sigma$ #### Lisa91 ##### New member Jan 15, 2013 29 We pratically can consider the term $\ln \{\ln (13\ j + k)\}$ as a constant term [it is very slowly changing with j...] and can analyse the term... $\displaystyle \frac{\ln (13\ j + k)}{13\ j + k}$ (1) Any doubt about the fact that, at least for j 'large enough', it is decreasing with j and that its limit is 0?... Kind regards $\chi$ $\sigma$ Thank you! I don't have any doubts about the fact that the limit of this guy is zero $$\frac{\ln (13\ j + k)}{13\ j + k}$$. Do you think the explanation 'we may consider it as a constant term' is good for the exam? I think I feel what you mean. Indeed, it increases very slowly but still I am looking for something more formal. #### chisigma ##### Well-known member Feb 13, 2012 1,704 Thank you! I don't have any doubts about the fact that the limit of this guy is zero $$\frac{\ln (13\ j + k)}{13\ j + k}$$. Do you think the explanation 'we may consider it as a constant term' is good for the exam? I think I feel what you mean. Indeed, it increases very slowly but still I am looking for something more formal. All right!... all what You have to do is to consider that, for j 'large enough', is $\displaystyle |\ln \{\ln (13\ j + k)\}| >1$ so that is... $\displaystyle |\frac{\ln (13\ j + k)}{(13\ j + k)\ \ln \{\ln (13\ j + k)\}}|< |\frac {\ln (13\ j + k)}{(13\ j + k)}|$ (1) Kind regards $\chi$ $\sigma$ #### Lisa91 ##### New member Jan 15, 2013 29 $$\ln \{\ln (13j + k) >1$$ $$\ln (13j + k)>e$$ $$13j + k >e^{e}$$ so if we take j=1 and k=0,1,... it's true and we can also take k=9 and j=0,1... Is it ok? #### Klaas van Aarsen ##### MHB Seeker Staff member Mar 5, 2012 8,866 Sorry, I have lost you in your argument. But did you consider that for [TEX]n \ge 3[/TEX]: [TEX]0 < {1 \over n \ln(\ln n)} < {\ln n \over n \ln(\ln n)} < {n \over n \ln(\ln n)} = {1 \over \ln(\ln n)} \to 0[/TEX] The LHS and RHS both approach zero when [TEX]n \to \infty[/TEX]. Moreover they do so monotonously. #### Lisa91 ##### New member Jan 15, 2013 29 Sorry, I have lost you in your argument. But did you consider that for [TEX]n \ge 3[/TEX]: [TEX]0 < {1 \over n \ln(\ln n)} < {\ln n \over n \ln(\ln n)} < {n \over n \ln(\ln n)} = {1 \over \ln(\ln n)} \to 0[/TEX] The LHS and RHS both approach zero when [TEX]n \to \infty[/TEX]. Moreover they do so monotonously. Thank you so much! It's so beautiful!!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9843013286590576, "perplexity": 1342.4476842552037}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141191511.46/warc/CC-MAIN-20201127073750-20201127103750-00373.warc.gz"}
http://math.stackexchange.com/questions/123710/normal-distribution-properties
# Normal Distribution Properties Hi all i have this question in which i don't really understand the line of reasoning. Given rvs X ∼ N(0, 1), Y ∼ N(0, 4), is P(X > 3) < P(Y < –6)? The reasoning is : P(Y < –6) = P(Y > 6) < P(Y / 2 > 3) = P(X > 3). I don't understand this part: P(Y > 6) < P(Y / 2 > 3) Isn't P(Y / 2 > 3) exactly equal to P(Y > 6)? Also i don't really understand why P(Y / 2 > 3) = P(X > 3) Could anyone enlighten me? Thanks :) - The random variable $Y$ has a mean of $0$ and a standard deviation of $\sqrt4=2$, so $Y/2$ has a mean of $0$ and a standard deviation of $2/2=1$ and is therefore $\mathcal{N}(0,1)$, just like $X$. Thus, $$\Bbb P\left(\frac{Y}2>3\right)=\Bbb P(X>3)\;.$$ For the rest, I agree with you that $Y/2>3$ exactly when $Y>6$, so that $$\Bbb P\left(\frac{Y}2>3\right)=\Bbb P(Y>6)\;.$$ It follows that $\Bbb P(Y<-6)=\Bbb P(X>3)$. It appears to me that the $<$ that’s bothering you is probably just a misprint.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9078055620193481, "perplexity": 54.90755590562361}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394023865238/warc/CC-MAIN-20140305125105-00093-ip-10-183-142-35.ec2.internal.warc.gz"}
http://ptsymmetry.net/?cat=129
## Magnetoelectric Effects in Local Light-Matter Interactions Konstantin Y. Bliokh, Yuri S. Kivshar, Franco Nori We study the generic interaction of a monochromatic electromagnetic field with bi-isotropic nanoparticles. Such an interaction is described by dipole-coupling terms associated with the breaking of dual, P- and T-symmetries, including the chirality and the nonreciprocal magnetoelectric effect. We calculate absorption rates, radiation forces, and radiation torques for the nanoparticles and introduce novel characteristics of the field quantifying the transfer of energy, momentum, and angular-momentum in these interactions. In particular, we put forward the concept of ‘magnetoelectric energy density’, quantifying the local PT-symmetry of the field. Akin to the ‘super-chiral’ light suggested recently for sensitive local probing of molecular chirality [Phys. Rev. Lett. 104, 163901 (2010); Science 332, 333 (2011)], here we describe a complex field for sensitive probing of the nonreciprocal magnetoelectric effect in nanoparticles or molecules. http://arxiv.org/abs/1312.4325 Optics (physics.optics); Other Condensed Matter (cond-mat.other); Quantum Physics (quant-ph) ## Pseudo Parity-Time Symmetry in Optical Systems Xiaobing Luo, Honghua Zhong, Jiahao Huang, Xizhou Qin, Qiongtao Xie, Yuri S. Kivshar, Chaohong Lee We introduce a novel concept of the {\em pseudo} parity-time ($$\mathcal{PT}$$) symmetry in periodically modulated optical systems with balanced gain and loss. We demonstrate that whether the original system is $$\mathcal{PT}$$-symmetric or not, we can manipulate the property of the $$\mathcal{PT}$$ symmetry by applying a periodic modulation in such a way that the effective system derived by the high-frequency Floquet method is $$\mathcal{PT}$$ symmetric. If the original system is non-$$\mathcal{PT}$$ symmetric, the $$\mathcal{PT}$$ symmetry in the effective system will lead to quasi-stationary propagation that can be associated with the \emph{pseudo $$\mathcal{PT}$$ symmetry}. Our results provide a promising approach for manipulating the $$\mathcal{PT}$$ symmetry of realistic systems. http://arxiv.org/abs/1302.1091 Optics (physics.optics); Quantum Physics (quant-ph) ## Discrete vortex solitons and PT symmetry Daniel Leykam, Vladimir V. Konotop, Anton S. Desyatnikov We study the effect of lifting the degeneracy of vortex modes with a PT symmetric defect, using discrete vortices in a circular array of nonlinear waveguides as an example. When the defect is introduced, the degenerate linear vortex modes spontaneously break PT symmetry and acquire complex eigenvalues, but nonlinear propagating modes with real propagation constants can still exist. The stability of nonlinear modes depends on both the magnitude and the sign of the vortex charge, thus PT symmetric systems offer new mechanisms to control discrete vortices. http://arxiv.org/abs/1301.1052 Optics (physics.optics); Pattern Formation and Solitons (nlin.PS) ## Breathers in PT-symmetric optical couplers I. V. Barashenkov, Sergey V. Suchkov, Andrey A. Sukhorukov, Sergey V. Dmitriev, Yuri S. Kivshar We show that the parity-time (PT) symmetric coupled optical waveguides with gain and loss support localised oscillatory structures similar to the breathers of the classical $$\phi^4$$ model. The power carried by the PT-breather oscillates periodically, switching back and forth between the waveguides, so that the gain and loss are compensated on the average. The breathers are found to coexist with solitons and be prevalent in the products of the soliton collisions. We demonstrate that the evolution of the small-amplitude breather’s envelope is governed by a system of two coupled nonlinear Schrodinger equations, and employ this Hamiltonian system to show that the small-amplitude PT-breathers are stable. http://arxiv.org/abs/1211.1835 Pattern Formation and Solitons (nlin.PS); Optics (physics.optics) ## Solitons in a chain of PT-invariant dimers Sergey V. Suchkov, Boris A. Malomed, Sergey V. Dmitriev, Yuri S. Kivshar Dynamics of a chain of interacting parity-time invariant nonlinear dimers is investigated. A dimer is built as a pair of coupled elements with equal gain and loss. A relation between stationary soliton solutions of the model and solitons of the discrete nonlinear Schrodinger (DNLS) equation is demonstrated. Approximate solutions for solitons whose width is large in comparison to the lattice spacing are derived, using a continuum counterpart of the discrete equations. These solitons are mobile, featuring nearly elastic collisions. Stationary solutions for narrow solitons, which are immobile due to the pinning by the effective Peierls-Nabarro potential, are constructed numerically, starting from the anti-continuum limit. The solitons with the amplitude exceeding a certain critical value suffer an instability leading to blowup, which is a specific feature of the nonlinear PT-symmetric chain, making it dynamically different from DNLS lattices. A qualitative explanation of this feature is proposed. The instability threshold drops with the increase of the gain-loss coefficient, but it does not depend on the lattice coupling constant, nor on the soliton’s velocity. http://arxiv.org/abs/1110.1501 Optics (physics.optics) ## Nonlinearly-PT-symmetric systems: spontaneous symmetry breaking and transmission resonances Andrey E. Miroshnichenko, Boris A. Malomed, Yuri S. Kivshar We introduce a class of PT-symmetric systems which include mutually matched nonlinear loss and gain (inother words, a class of PT-invariant Hamiltonians in which both the harmonic and anharmonic parts are non-Hermitian). For a basic system in the form of a dimer, symmetric and asymmetric eigenstates, including multistable ones, are found analytically. We demonstrate that, if coupled to a linear chain, such a nonlinear PT-symmetric dimer generates new types of nonlinear resonances, with the completely suppressed or greatly amplified transmission, as well as a regime similar to the electromagnetically-induced transparency (EIT). The implementation of the systems is possible in various media admitting controllable linear and nonlinear amplification of waves. http://arxiv.org/abs/1104.0849 Mathematical Physics (math-ph) ## Nonlinear suppression of time-reversals in PT-symmetric optical couplers Andrey A. Sukhorukov, Zhiyong Xu, Yuri S. Kivshar We reveal a generic connection between the effect of time-reversals and nonlinear wave dynamics in systems with parity-time (PT) symmetry, considering a symmetric optical coupler with balanced gain and loss where these effects can be readily observed experimentally. We show that for intensities below a threshold level, the amplitudes oscillate between the waveguides, and the effects of gain and loss are exactly compensated after each period due to {periodic time-reversals}. For intensities above a threshold level, nonlinearity suppresses periodic time-reversals leading to the symmetry breaking and a sharp beam switching to the waveguide with gain. Another nontrivial consequence of linear PT-symmetry is that the threshold intensity remains the same when the input intensities at waveguides with loss and gain are exchanged. http://arxiv.org/abs/1009.5428 Optics (physics.optics)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7929911613464355, "perplexity": 2478.5809458706703}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027330786.8/warc/CC-MAIN-20190825173827-20190825195827-00033.warc.gz"}
https://couryes.com/game-theory-dai-xie-econ3301-4/
## 经济代写|博弈论代写Game Theory代考|ECON3301 2022年12月29日 couryes-lab™ 为您的留学生涯保驾护航 在代写博弈论Game Theory方面已经树立了自己的口碑, 保证靠谱, 高质且原创的统计Statistics代写服务。我们的专家在代写博弈论Game Theory代写方面经验极为丰富,各种代写博弈论Game Theory相关的作业也就用不着说。 • Statistical Inference 统计推断 • Statistical Computing 统计计算 • (Generalized) Linear Models 广义线性模型 • Statistical Machine Learning 统计机器学习 • Longitudinal Data Analysis 纵向数据分析 • Foundations of Data Science 数据科学基础 couryes™为您提供可以保分的包课服务 ## 经济代写|博弈论代写Game Theory代考|IMPERFECT INFORMATION In this section, we will focus on games with imperfect information. Recall, these occur when players do not know one or more actions taken hy the other players and so find themselves in an information set with more than one history. Assuming that Suzette has already made her decisions about how to carry out the bidding process, Mark and Ben are the players in this game. The game is sequential in nature, and Ben has incomplete information about Mark’s action. To create a simple model, we assume Mark can choose one of three actions: not bid (No), bid low (Lo), or bid high (Hi). Ben only knows whether Mark bids or does not bid; if Mark does bid, Ben does not know whether he bid high or low. Since Ben is not interested in doing the work for a low price, we assume he chooses between two actions: not bid (No) or bid high (Hi). Assigning vNM utilities consistent with this discussion, we obtain the game tree in Figure $6.1$ and Table 6.1. We can see from the game tree the non-terminal histories are partitioned into three information sets, listed in Table 6.2. Only Ben faces an information set, Ben2, with more than one node (non-terminal history). Here Ben does not know which sequence of actions has occurred and is making his decision whether to choose No or Hi without knowing whether Mark has bid Hi or Lo. It is important to consider players’ overall strategies, rather than just individual actions. In FoodPro, one pure strategy for Ben is to always choose No. Another is to choose No when Mark Bids (at information set Ben2) and Hi if Mark does not bid (at information set Ben1). Mark has three pure strategies, corresponding to the actions at his only nonterminal history: No, Lo, or Hi. For strategic games, we allowed players to adopt mixed strategies. While we could allow players in extensive games to adopt such mixed strategies, it is typically sufficient and more natural to allow players to randomize action choices at each information set based on probability distributions. We call such strategies behavior strategies and formally define them below. Definition 6.1.1. A pure strategy for player $i$ is a function $s_i$ which assigns to each of the player’s information sets a possible action. A behavior strategy for player $i$ is a function $s_i$ which assigns to each of the player’s information sets a probability distribution over possible actions. If $s$ is used to denote a pure or behavior strategy, we will use $s_i(I)$ or simply $s(I)$ to denote the action or probability distribution over actions chosen by player $i$ at the assigned information set $I$, and $s_i(a \mid I)$ or simply $s(a \mid I)$ will denote the probability that player $i$ will choose action $a$ at information set $I$. ## 经济代写|博弈论代写Game Theory代考|ROMANS AND GERMANS In this section, we will construct a model for an unfortunately common scenario in human history, that of war. We’ll use a historical scenario, but our model and analysis can apply to many modern-day situations. Barron [8] describes a method for modeling a fictitious battle in the Greco-Persian wars. We present an adaptation of his model for the Greek and Roman battle described above. Using RE and GT to name the players representing the Roman Empire and the Germanic tribe, the extensive game tree is given in Figure 6.8. To more easily discern the reasoning behind the payoffs, a short description of each outcome and the corresponding payoffs are rank ordered in Tables $6.3$ and 6.4. Observe that payoffs were assigned so that positive numbers correspond to an overall “win” for that player and negative numbers correspond to an overall “loss” for that player, and the numbers are to be interpreted as vNM utilities. We encourage the reader to consider whether different numbers might be more realistic. This model is different from the three games analyzed in the first section because both players lack information. This lack of information is again indicated by the dashed line boxes around three pairs of the decision nodes. Within each of these boxes, the player who is making the decision does not know which history has occurred. We have also labeled each decision point so that they can be easily referenced. For example, within the GT1 box, when the German tribe is making its decision whether to defend the forest or lake, it does not know the Roman Empire legion’s decision whether to advance through the forest or over the lake, however, at the GT2 node (which may also be considered as an information set containing just this node), the German tribe knows that they have defended an advance across the lake but the Roman empire legion has attacked the village after advancing through the forest. The Warfare game has three subgames: First, the entire game has the empty history as its root. Second, the subgame with root (Forest, Lake) which corresponds in Figure $6.8$ to the node labeled GT2 and everything to the right of GT2. Third, the subgame with root (Lake, Forest) which corresponds in Figure $6.8$ to the node labeled GT3 and everything to the right of GT3. The effect of requiring the consistency condition to hold on every subgame rather than just the entire game is to ensure players choose an equilibrium even on subgames that have zero probability of being reached. We now find the weak sequential equilibria. Figure $6.8$ includes variable names below edges and nodes corresponding to the probabilities associated with an arbitrary assessment. Assume that the assessment is a weak sequential equilibrium. # 博弈论代考 ## 经济代写|博弈论代写Game Theory代考|ROMANS AND GERMANS Barron [8] 描述了一种模拟希波战争中虚构战斗的方法。我们展示了他对上述希腊和罗马战争模型的改编。 Warfare 游戏分为三个子游戏:首先,整个游戏以空历史为根。二、图中对应的根为(Forest, Lake)的子博弈6.8到标记为 GT2 的节点以及 GT2 右侧的所有内容。三、图中对应的根为(Lake, Forest)的子博弈6.8到标记为 GT3 的节点以及 GT3 右侧的所有内容。要求一致性条件适用于每个子博弈而不仅仅是整个博弈的效果是确保玩家即使在达到零概率的子博弈中也选择均衡。 ## 有限元方法代写 tatistics-lab作为专业的留学生服务机构,多年来已为美国、英国、加拿大、澳洲等留学热门地的学生提供专业的学术服务,包括但不限于Essay代写,Assignment代写,Dissertation代写,Report代写,小组作业代写,Proposal代写,Paper代写,Presentation代写,计算机作业代写,论文修改和润色,网课代做,exam代考等等。写作范围涵盖高中,本科,研究生等海外留学全阶段,辐射金融,经济学,会计学,审计学,管理学等全球99%专业科目。写作团队既有专业英语母语作者,也有海外名校硕博留学生,每位写作老师都拥有过硬的语言能力,专业的学科背景和学术写作经验。我们承诺100%原创,100%专业,100%准时,100%满意。 ## MATLAB代写 MATLAB 是一种用于技术计算的高性能语言。它将计算、可视化和编程集成在一个易于使用的环境中,其中问题和解决方案以熟悉的数学符号表示。典型用途包括:数学和计算算法开发建模、仿真和原型制作数据分析、探索和可视化科学和工程图形应用程序开发,包括图形用户界面构建MATLAB 是一个交互式系统,其基本数据元素是一个不需要维度的数组。这使您可以解决许多技术计算问题,尤其是那些具有矩阵和向量公式的问题,而只需用 C 或 Fortran 等标量非交互式语言编写程序所需的时间的一小部分。MATLAB 名称代表矩阵实验室。MATLAB 最初的编写目的是提供对由 LINPACK 和 EISPACK 项目开发的矩阵软件的轻松访问,这两个项目共同代表了矩阵计算软件的最新技术。MATLAB 经过多年的发展,得到了许多用户的投入。在大学环境中,它是数学、工程和科学入门和高级课程的标准教学工具。在工业领域,MATLAB 是高效研究、开发和分析的首选工具。MATLAB 具有一系列称为工具箱的特定于应用程序的解决方案。对于大多数 MATLAB 用户来说非常重要,工具箱允许您学习应用专业技术。工具箱是 MATLAB 函数(M 文件)的综合集合,可扩展 MATLAB 环境以解决特定类别的问题。可用工具箱的领域包括信号处理、控制系统、神经网络、模糊逻辑、小波、仿真等。
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.691368043422699, "perplexity": 1648.6988202808311}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945433.92/warc/CC-MAIN-20230326044821-20230326074821-00085.warc.gz"}
http://mathematica.stackexchange.com/users/5891/trissse?tab=summary
TriSSSe Reputation 228 Next privilege 250 Rep. 1 5 Impact ~2k people reached • 0 posts edited • 0 helpful flags • 4 votes cast Questions (2) 19 Usage of Assuming for Integration 6 Reproducing the Integral Definition of the Modified Bessel function Reputation (228) This user has no recent positive reputation changes This user has not answered any questions Tags (3) 0 special-functions 0 assumptions 0 calculus-and-analysis Accounts (10) Mathematics 308 rep 110 Mathematica 228 rep 15 TeX - LaTeX 224 rep 129 Stack Overflow 139 rep 4 Ask Different 123 rep 3
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9835293889045715, "perplexity": 21814.919551891955}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701987329.64/warc/CC-MAIN-20160205195307-00131-ip-10-236-182-209.ec2.internal.warc.gz"}
http://physicsandphysicists.blogspot.com/2011/11/kilogram-to-be-tied-to-planck-constant.html
## Monday, November 07, 2011 ### The Kilogram To Be Tied to Planck Constant I've mentioned before of the issue related to the last of the 7 SI units that still is not tied to a fundamental constant - the kilogram. This is compounded by the fact that one of the standard mass that is being kept outside of Paris is changing mass. They think it is losing weight, but they're not sure now if it is a weight loss or a weight gain (really?!). Still, at the end of the article in that last link, there was something that I didn't know about. Mr Picard said the kilogram had gained (or lost) the equivalent of a small grain of sand in weight, but that was enough to throw out calculations in everything from precision engineering to trade. It will be replaced by the Planck Constant, named after Max Planck, which is the smallest packet of energy (or quanta) that two particles can exchange. Last month the General Conference on Weights and Measures (CGPM) agreed to use the constant to calculate the value of the kilo - but not before 2014. Whoa! I didn't know they decided on that already! So I went and did a search, and found it immediately. This is a press release from BIPM. A redefinition of the kilogram first requires highly accurate measurements of a fundamental constant of nature in terms of the mass of the international prototype of the kilogram, currently exactly equal to 1 kilogram. The numerical value of the fundamental constant will then be fixed and the same experiment will later be used to measure the mass of objects including the international prototype. Several facilities throughout the world capable of carrying out such measurements will be needed after the redefinition in order to make practical use of the new kilogram definition. The target uncertainty for the most accurate of such measurements is 20 microgram per kilogram, which is the same as 20 parts in one thousand million. It is remarkable that at least two experimental approaches are very close to achieving this goal. One approach uses a special electronic balance – a “watt balance” – in order to measure the kilogram in terms of the Planck constant, which is the fundamental constant of quantum mechanics. A second technique compares one kilogram to the mass of a single atom of the chemical element silicon. Physics tells us that the results of these two seemingly different approaches can be accurately compared with each other and, of course, they should agree. The present situation has been examined by the CODATA Task Group on Fundamental Constants based on work published through the end of 2010. They conclude that the present uncertainty of the Planck constant from all relevant experimental approaches is the equivalent of 44 microgram per kilogram. The CGPM will not adopt the proposed new definitions until present difficulties are resolved. However, on Friday 21 October 2011, the General Conference took a historic step towards the revision by adopting Resolution 1 and thereby outlining the proposed New SI as well as the steps required for the final completion of this project. The text of Resolution 1 is that of Draft Resolution A, which had been publicly available for some months on the BIPM “New SI” website, with only minor changes made during the Conference. One of these asks the International Committee for Weights and Measures (CIPM) to continue its work to render the language of the New SI as far as possible understandable for users in general, while maintaining scientific rigour and clarity and without altering the basic content and structure of the New SI as set forth in Resolution 1. In the resolution of the most recent meeting, there is a bit more description on how the kilogram will be tied to the Planck constant. the kilogram will continue to be the unit of mass, but its magnitude will be set by fixing the numerical value of the Planck constant to be equal to exactly 6.626 06X ×10–34 when it is expressed in the SI unit m2 kg s–1, which is equal to J s . . . the mass of the international prototype of the kilogram m(K) will be 1 kg but with a relative uncertainty equal to that of the recommended value of h just before redefinition and that subsequently its value will be determined experimentally, Well, there ya go. Looks like some time in 2014, that standard mass will be nothing more than a historical, museum piece, just like the 1 meter rod. Zz.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8716508746147156, "perplexity": 478.21778933983813}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257830064.24/warc/CC-MAIN-20160723071030-00306-ip-10-185-27-174.ec2.internal.warc.gz"}
http://www.r-statistics.com/2013/01/stargazer-package-for-beautiful-latex-tables-from-r-statistical-models-output/
# {stargazer} package for beautiful LaTeX tables from R statistical models output stargazer is a new R package that creates LaTeX code for well-formatted regression tables, with multiple models side-by-side, as well as for summary statistics tables. It can also output the content of data frames directly into LaTeX. Compared to available alternatives, stargazer excels in three regards:  its ease of use, the large number of models it supports, and its beautiful aesthetics. ### Ease of use stargazer was designed with the user’s comfort in mind. The learning curve is very mild and all arguments are very intuitive, so that even a beginning user of R or LaTeX can quickly become familiar with the package’s many capabilities. The package is intelligent, and tries to minimize the amount of effort the user has to put into adjusting argument values. If stargazer is given a set of regression model objects, for instance, the package will create a side-by-side regression table. By contrast, if the user feeds it a data frame, stargazer will know that the user is most likely looking for a summary statistics table or – if the summary argument is set to false – wants to output the content of the data frame. A quick reproducible example shows just how easy stargazer is to use. You can install stargazer from CRAN in the usual way: install.packages("stargazer") library(stargazer) To create a summary statistics table from the ‘attitude’ data frame (which should be available with your default installation of R), simply run the following: stargazer(attitude) To output the contents of the first four rows of same data frame, specify the part of the data frame you would like to see, and set the summary option to FALSE: stargazer(attitude[1:4,], summary=FALSE) Now, let us try to create a simple regression table with three side-by-side models – two Ordinary Least Squares (OLS) and one probit regression model – using the lm() and glm() functions. We can set the align argument to TRUE, so that coefficients in each column are aligned along the decimal point: ## 2 OLS models linear.1 <- lm(rating ~ complaints + privileges + learning + raises + critical, data=attitude) linear.2 <- lm(rating ~ complaints + privileges + learning, data=attitude)   ## create an indicator dependent variable, and run a probit model   attitude$high.rating <- (attitude$rating > 70) probit.model <- glm(high.rating ~ learning + critical + advance, data=attitude, family = binomial(link = "probit"))   stargazer(linear.1, linear.2, probit.model, title="Regression Results", align=TRUE) ### Many supported models stargazer supports objects from the most widely used statistical functions and packages. These include objects from betareg (betareg), coxph (survival), clm (ordinal), clogit (survival), ergm (ergm),gam (mgcv), gee (gee), glm (stats), glmer (lme4), gls (nlme), hurdle (pscl), ivreg (AER), lm (stats), lmer (lme4), lmrob (robustbase), multinom (nnet), nlmer (lme4), plm (plm), pmg (plm), polr (MASS), rlm (MASS), svyglm (survey), survreg (survival), tobit (AER), zeroinfl (pscl), as well as from the implementation of these in Zelig. In addition, stargazer also supports several Zelig models for social network analysis: cloglog.net, gamma.net, probit.net, and logit.net. The number of models and objects can stargazer can accommodate puts it ahead of most of the alternative R-to-LaTeX options. As the development of the package continues, this list will continue expanding to include linear mixed effects models, matching models, as well as new, user-made, or customized statistical models. ### Beautiful aesthetics stargazer is very pleasing to the eye, and allows the user to customize the formatting of the resulting table, including all variable labels. Below is an example of a good-looking, complex regression table created by stargazer: If you’d like to create tables that look like those from your discipline’s leading journal, stargazer can help you with that as well. You can use the style argument to choose a template of your choice. Economics and management scholars can thus create tables that resemble those published in the American Economic Review, in the Quarterly Journal of Economics, or in Administrative Science Quarterly. Political scientists can avail themselves of templates based on the American Political Science Review, the American Journal of Political Science, and on International Organization. For sociologists and demographers, the American Sociological Review, the American Sociological Reviews and Demography are available. stargazer, of course, is not the only R package that creates LaTeX code from R statistical output. Other packages with similar capabilities include apsrtable, xtable, memisc, texreg and outreg. Each of these has its own strengths and weaknesses, and users should explore all of them to find the best fit for their needs. ### Some extra Q&A with Marek Hlavac (the package author) What was your motivation for starting the package? (self use, for students, for other people etc.) As a doctoral student in Political Economy and Government at Harvard University, I saw an urgent need for an easy-to-use tool to create well-formatted stargazer tables. Although other packages were available for this task, none of them combined a large number of supported models, good aesthetics, and simplicity of use in a way that I thought would be ideal. Why is the “summary” parameter turned on as default for data.frames? In published papers in the social sciences, summary statistics table are found more commonly than direct print-outs of chunks of data sets. For this reason, I thought the stargazer package should default to the user’s most likely need. This has been a guest post by Marek Hlavac, the author of the {stargazer} R package for beautiful LaTeX tables from R’s statistical models’ outputs. ## 27 thoughts on “{stargazer} package for beautiful LaTeX tables from R statistical models output” 1. Jeremy says: Nice, but no changelog posted on CRAN? And no changelog in the .tar file? 2. roth says: thanks a lot! 3. Brian says: It seems that this package and many similar R routines combine what could be two separate steps. First, they take data (or other R-objects) along with simple formatting parameters and construct a generalized table (including such things as merged cells, and per-cell formatting). Second, they output that table to a back-end (LaTeX in this case). If these steps were formally separated, users could tweak formatting (making changes to the generalized table) and could provide new back-ends (e.g. HTML, Excel, etc.). Do any of these packages do this? 1. Tormod Bøe says: Look into the package “xtable”, this allows you to output the table as hmtl, perfect for use with markdown in Rstudio for instance! 4. Josh says: Awesome package! I’m sure I’ll be using it soon 🙂 5. Roberto says: These tables really look nice. Moreover it is really convenient tool for those who are no experts in TeX. Is there any chance for the user to modify the command such that standard errors are given (instead of t-statistics) and to report a Mc Fadden R² for Probit Models? 6. I am not sure if this is a bug, but when I try to output a dataframe with decimal numbers, I get an error: > myframe V1 V2 1 0.457 41 2 23.000 57 3 28.500 39 4 45.000 58 5 49.000 31 > stargazer(myframe,summary=FALSE) % Table created by StarGazer v.3.0 by Marek Hlavac, Harvard University. E-mail: hlavac at fas.harvard.edu % Date and time: ma, jan 28, 2013 – 14:33:06 begin{table}[htb] centering caption{} label{} footnotesize begin{tabular}{@{extracolsep{5pt}} c c } [-1.8ex]hline hline [-1.8ex] V1 & V2 hline [-1.8ex] When I use the same dataframe with integers only, it works fine. 7. ignacio martinez says: is it possible to create a .tex file with the table instead of writing the table to the screen? Thanks! Fantastic work. Quick question – is it possible to display number of NAs in a summary table? 9. rasmus says: When i do stargazaer(dataframe) it omits the rows. How do I keep the rows? 10. Charlie says: Hi, thanks for this package! I was wondering if stargazer had an option for long tables? 11. Michael says: I have trouble reading the table. When I type: library(stargazer) stargazer(attitude) The output is like: % Table created by stargazer v.5.1 by Marek Hlavac, Harvard University. E-mail: hlavac at fas.harvard.edu % Date and time: Tue, May 05, 2015 – 14:07:16 \begin{table}[!htbp] \centering \caption{} \label{} \begin{tabular}{@{\extracolsep{5pt}}lccccc} \\[-1.8ex]\hline \hline \\[-1.8ex] Statistic & \multicolumn{1}{c}{N} & \multicolumn{1}{c}{Mean} & \multicolumn{1}{c}{St. Dev.} & \multicolumn{1}{c}{Min} & \multicolumn{1}{c}{Max} \\ \hline \\[-1.8ex] rating & 30 & 64.633 & 12.173 & 40 & 85 \\ complaints & 30 & 66.600 & 13.315 & 37 & 90 \\ privileges & 30 & 53.133 & 12.235 & 30 & 83 \\ learning & 30 & 56.367 & 11.737 & 34 & 75 \\ raises & 30 & 64.633 & 10.397 & 43 & 88 \\ critical & 30 & 74.767 & 9.895 & 49 & 92 \\ advance & 30 & 42.933 & 10.289 & 25 & 72 \\ \hline \\[-1.8ex] \end{tabular} \end{table} Could anybody help me fix it ? I am using a Mac and I just installed MacTax. Thanks! 1. Gabi says: What are you trying to fix though? Can you be more specific? This is what you expect from the output, it gives you LaTeX code that will create a table when compiled (you do have to compile it). Or are you getting some errors when compiling? 2. Gabi says: What are you trying to fix though? Can you be more specific? This is what you expect from the output, it gives you LaTeX code that will create a table when compiled (you do have to compile it within a LaTeX document). Or are you getting some errors when compiling? 12. amos says: Is there an implementation of stargazer or similar function for an object of type ‘geeglm’ for modelling repeated measures using a generalized linear model? I tried but got the error ‘Error: Unrecognized object type.’. 13. Bernardo Paz says: Thankyou very much Marek Hlavac you make a great contribution ! 14. thanks a bunch, Marek. This looks wonderful. I wonder if there is anything similar available for JAGS output? 15. eszter gagyi says: Hy! I get the same problem like Maxim Kovalenko. Can you let me knwo how to fix it? thanx in advance 🙂
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4031262993812561, "perplexity": 3469.4362641718094}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049276567.28/warc/CC-MAIN-20160524002116-00099-ip-10-185-217-139.ec2.internal.warc.gz"}
http://mathoverflow.net/questions/56762/representations-and-support
# Representations and support. I am interested in the question: Does there are exist concept of support in representation theory? When I say support I mean number of non-zero values of $f \in C[G]$. Do you know theorems which talks about the action of elements of $C[G]$ with small support in different representations? The only example I know about is uncertainty principle which says that for abelian group $supp(f)supp(\hat{f})\geq |A|$. - I think the question would benefit from being slightly more precise. It seems better to me if one asks for "theorems that show something like X happens" rather than "theorems which use the concept X". –  Yemon Choi Feb 28 '11 at 7:20 Although perhaps the question was directed more at finite groups: for reductive or semi-simple real Lie groups, (serious) results of Harish-Chandra (starting in the early 1950s) show that among regular semi-simple elements of the group, the supports of characters of principal series are the smallest, being confined to (maximally) split conjugacy classes, while at the opposite end the characters of discrete series have non-trivial support on all semi-simple conjugacy classes. The in-between repns, i.e., induced from discrete series on Levis, have in-between supports, in terms of split-ness. I have the impression that suitable analogues hold for p-adic reductive groups, tho' the corresponding results are relatively much-newer (by a displacement of 15-20 years?), not to mention the complication in description of discrete series (supercuspidal repns) as induced from compact-open subgroups (Kutzko-Bushnell-et-alia). Early computations for the p-adic case go back to MacDonald (SL2) in the 1950s, and Shalika, Sally in the 1960s (SL2), Jacquet in the late 1960s, so far as I know. -
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.645357072353363, "perplexity": 350.6309473183228}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-32/segments/1438042988250.59/warc/CC-MAIN-20150728002308-00020-ip-10-236-191-2.ec2.internal.warc.gz"}
http://gmatclub.com/forum/if-k-is-an-integer-and-2-k-7-for-how-many-different-135543.html?oldest=1
If k is an integer and 2 < k < 7, for how many different : GMAT Problem Solving (PS) Check GMAT Club Decision Tracker for the Latest School Decision Releases http://gmatclub.com/AppTrack It is currently 22 Jan 2017, 23:26 ### GMAT Club Daily Prep #### Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email. Customized for You we will pick new questions that match your level based on your Timer History Track every week, we’ll send you an estimated GMAT score based on your performance Practice Pays we will pick new questions that match your level based on your Timer History # Events & Promotions ###### Events & Promotions in June Open Detailed Calendar # If k is an integer and 2 < k < 7, for how many different Author Message TAGS: ### Hide Tags Math Expert Joined: 02 Sep 2009 Posts: 36601 Followers: 7098 Kudos [?]: 93489 [0], given: 10563 If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 09 Jul 2012, 02:34 Expert's post 16 This post was BOOKMARKED 00:00 Difficulty: 35% (medium) Question Stats: 58% (01:39) correct 42% (00:37) wrong based on 920 sessions ### HideShow timer Statistics If k is an integer and 2 < k < 7, for how many different values of k is there a triangle with sides of lengths 2, 7, and k? (A) one (B) two (C) three (D) four (E) five Diagnostic Test Question: 19 Page: 22 Difficulty: 600 [Reveal] Spoiler: OA _________________ Manager Affiliations: Project Management Professional (PMP) Joined: 30 Jun 2011 Posts: 213 Location: New Delhi, India Followers: 5 Kudos [?]: 67 [0], given: 12 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 09 Jul 2012, 03:37 Bunuel wrote: The Official Guide for GMAT® Review, 13th Edition - Quantitative Questions Project If k is an integer and 2 < k < 7, for how many different values of k is there a triangle with sides of lengths 2, 7, and k? (A) one (B) two (C) three (D) four (E) five Diagnostic Test Question: 19 Page: 22 Difficulty: 650 GMAT Club is introducing a new project: The Official Guide for GMAT® Review, 13th Edition - Quantitative Questions Project Each week we'll be posting several questions from The Official Guide for GMAT® Review, 13th Edition and then after couple of days we'll provide Official Answer (OA) to them along with a slution. We'll be glad if you participate in development of this project: 2. Please vote for the best solutions by pressing Kudos button; 3. Please vote for the questions themselves by pressing Kudos button; 4. Please share your views on difficulty level of the questions, so that we have most precise evaluation. Thank you! Believe ans is A as k can be 3,4,5 or 6. for all cases except k=6, sum of teo sides can be less than equal to third side which should not be true As as per triangle property sum of two sides is greater than third side. _________________ Best Vaibhav If you found my contribution helpful, please click the +1 Kudos button on the left, Thanks Joined: 28 Mar 2012 Posts: 314 Location: India GMAT 1: 640 Q50 V26 GMAT 2: 660 Q50 V28 GMAT 3: 730 Q50 V38 Followers: 29 Kudos [?]: 399 [0], given: 23 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 09 Jul 2012, 05:20 1 This post was BOOKMARKED Bunuel wrote: If k is an integer and 2 < k < 7, for how many different values of k is there a triangle with sides of lengths 2, 7, and k? (A) one (B) two (C) three (D) four (E) five Hi, Difficulty level: 600 |7-2| < k < |7+2| or 5 < k < 9 thus k = 6, 7, 8, but 2 < k < 7 therefore, k = 6 Regards, Math Expert Joined: 02 Sep 2009 Posts: 36601 Followers: 7098 Kudos [?]: 93489 [1] , given: 10563 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 13 Jul 2012, 01:59 1 KUDOS Expert's post 5 This post was BOOKMARKED SOLUTION If k is an integer and 2 < k < 7, for how many different values of k is there a triangle with sides of lengths 2, 7, and k? (A) one (B) two (C) three (D) four (E) five Relationship of the Sides of a Triangle: The length of any side of a triangle must be larger than the positive difference of the other two sides, but smaller than the sum of the other two sides. According to the above the following must be true: $$(7-2)<k<(7+2)$$ --> $$5<k<9$$. So, $$k$$ could be 6, 7 or 8. Since also given that $$2 < k < 7$$, then $$k=6$$. Hence $$k$$ can take only one value: 6. _________________ Intern Joined: 14 Nov 2012 Posts: 3 Followers: 0 Kudos [?]: 0 [0], given: 0 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 25 Nov 2012, 13:50 Why is it only one value? What if 2 is the smallest side and 7 is the largest? Then, k can be 6, 7 or 8. since: 2 + 7 = 9 7 - 2 = 5 So 5 < k < 9 The problem does not specify that 2 and 7are the smaller two sides. Last edited by morfin on 26 Nov 2012, 09:05, edited 1 time in total. Math Expert Joined: 02 Sep 2009 Posts: 36601 Followers: 7098 Kudos [?]: 93489 [0], given: 10563 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 26 Nov 2012, 01:05 morfin wrote: What if 2 is the smallest side and 7 is the largest? Then, the answer is B (two) since: 2 + 7 = 9 7 - 2 = 6 So 6 < k < 9 The problem does not specify that 2 and 7are the smaller two sides. Not sure I understand your question. First of all: 7-2=5, not 6. Next, obviously since the lengths of the sides are 2, 7 and k, where 2<k<7, then the length of the smallest side is 2 and the length of the largest side is 7. Check here for complete solution: if-k-is-an-integer-and-2-k-7-for-how-many-different-135543.html#p1104032 _________________ Senior Manager Joined: 29 Oct 2013 Posts: 297 Concentration: Finance GMAT 1: 750 Q V46 GPA: 3.7 WE: Corporate Finance (Retail Banking) Followers: 14 Kudos [?]: 377 [1] , given: 197 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 19 May 2014, 22:39 1 KUDOS Hi Bunuel, 39% got it wrong. Do u wanna tag it 600-700 instead of sub600? thanks _________________ My journey V46 and 750 -> http://gmatclub.com/forum/my-journey-to-46-on-verbal-750overall-171722.html#p1367876 Math Expert Joined: 02 Sep 2009 Posts: 36601 Followers: 7098 Kudos [?]: 93489 [0], given: 10563 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 19 May 2014, 22:43 MensaNumber wrote: Hi Bunuel, 39% got it wrong. Do u wanna tag it 600-700 instead of sub600? thanks ____________ Done. Thank you. _________________ Senior Manager Joined: 29 Oct 2013 Posts: 297 Concentration: Finance GMAT 1: 750 Q V46 GPA: 3.7 WE: Corporate Finance (Retail Banking) Followers: 14 Kudos [?]: 377 [0], given: 197 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 19 May 2014, 22:49 Bunuel wrote: MensaNumber wrote: Hi Bunuel, 39% got it wrong. Do u wanna tag it 600-700 instead of sub600? thanks ____________ Done. Thank you. Great ! Thanks for a quick reply. _________________ My journey V46 and 750 -> http://gmatclub.com/forum/my-journey-to-46-on-verbal-750overall-171722.html#p1367876 Intern Joined: 14 May 2014 Posts: 45 Followers: 0 Kudos [?]: 40 [1] , given: 1 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 20 May 2014, 03:17 1 KUDOS Relationship of the Sides of a Triangle: The length of any side of a triangle must be larger than the positive difference of the other two sides, but smaller than the sum of the other two sides. Above relationship is in the core of this solution. Many times I have found students having difficulty is assimilating this concept. Best way to get this concept is trying to actually draw triangles which contradict this. For example, try to draw a triangle with following sides (actual scale) 4, 3, 8 Once you failed, you will realize that The length of any side of a triangle must be larger than the positive difference of the other two sides, but smaller than the sum of the other two sides. _________________ Help me with Kudos if it helped you " Mathematics is a thought process. GMAT Club Legend Joined: 09 Sep 2013 Posts: 13508 Followers: 577 Kudos [?]: 163 [0], given: 0 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 09 Aug 2015, 13:35 Hello from the GMAT Club BumpBot! Thanks to another GMAT Club member, I have just discovered this valuable topic, yet it had no discussion for over a year. I am now bumping it up - doing my job. I think you may find it valuable (esp those replies with Kudos). Want to see all other topics I dig out? Follow me (click follow button on profile). You will receive a summary of all topics I bump in your profile area as well as via email. _________________ GMAT Club Legend Joined: 09 Sep 2013 Posts: 13508 Followers: 577 Kudos [?]: 163 [0], given: 0 Re: If k is an integer and 2 < k < 7, for how many different [#permalink] ### Show Tags 13 Aug 2016, 12:47 Hello from the GMAT Club BumpBot! Thanks to another GMAT Club member, I have just discovered this valuable topic, yet it had no discussion for over a year. I am now bumping it up - doing my job. I think you may find it valuable (esp those replies with Kudos). Want to see all other topics I dig out? Follow me (click follow button on profile). You will receive a summary of all topics I bump in your profile area as well as via email. _________________ Re: If k is an integer and 2 < k < 7, for how many different   [#permalink] 13 Aug 2016, 12:47 Similar topics Replies Last post Similar Topics: 1 How many possible positive integer values are there for k if -24 < 9k 2 13 Nov 2016, 10:12 35 How many different pairs of integers (a, b) exist such that 2 <= a <= 6 04 May 2015, 05:05 8 If k is an integer and 121 < k^2 < 225, then k can have at most how 12 27 Mar 2015, 04:28 8 How many even integers n ; 13<=n<=313 are of the form 3k+4, 10 13 Sep 2013, 23:01 8 For how many integers k is k^2 = 2^k ? 11 03 Jan 2012, 05:35 Display posts from previous: Sort by
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4890553951263428, "perplexity": 2565.5927890424014}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560282140.72/warc/CC-MAIN-20170116095122-00070-ip-10-171-10-70.ec2.internal.warc.gz"}
http://mathhelpforum.com/calculus/77059-showing-function-sobolev-space-print.html
# Showing function is in sobolev space I am trying to show that the function $u(r,\theta)=r^{2/3} sin((2\theta+\pi)/3)$ defined on the domain $0 \le r \le 1$ and $- \pi /2 \le \theta \le \pi$ is in $H^1(\Omega)$ but not in $H^2(\Omega)$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 5, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7785015106201172, "perplexity": 39.51517175106556}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368704218408/warc/CC-MAIN-20130516113658-00024-ip-10-60-113-184.ec2.internal.warc.gz"}
http://math.stackexchange.com/questions/66010/what-is-this-variation-of-the-traveling-salesman-problem-called
# What is this variation of the traveling salesman problem called? I've encountered a description of a well-known, hard to compute problem. Let us consider $n$ sites that need to be connected in the shortest possible way. If I am only allowed to connect the sites I have a traveling salesman problem. Now, I am allowed to add nodes to the map and connect the sites to those nodes (or even connect two nodes). What it the name of this problem? - It's hard to understand what you're asking here. You said $n$ sites and that they have to be connected in the "shortest" way; are you given pairwise distances for all sites? When you add a new site, what are the distances from it to all other sites? – Fixee Sep 20 '11 at 6:48 ## 1 Answer Assuming the sites are given as points in the plane, this is the Steiner tree problem. By the way, if you are trying to do the same thing without introducing additional nodes, as in the initial part of your question, then this is the problem of computing a minimum spanning tree, not the travelling salesman problem, and can be solved in polynomial time. -
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.575501561164856, "perplexity": 199.28842417795312}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257827791.21/warc/CC-MAIN-20160723071027-00241-ip-10-185-27-174.ec2.internal.warc.gz"}
http://www.pcmag.com/encyclopedia/term/38364/backslash
Definition of: backslash backslash The symbol (\) used as a separator between folder and file names in DOS and Windows. For example, the path to the Windows version of this encyclopedia is c:\"program files"\CDEweb\CDEweb.exe, which points to the CDEweb.exe file in the CDEweb folder within the Program Files folder on the C: drive. An Unfortunate, Confusing Symbol While the backslash (\) is used in Windows addresses, the forward slash (/) is used in Internet addresses, which are Unix based. In addition, the backslash key is in a non-standard keyboard location. Windows followed DOS, which was modeled after CP/M, and CP/M used the forward slash for command line parameters without regard to Unix path compatibility (Unix was big iron at the time, and CP/M was for personal computers). Thus we are stuck with two different symbols to separate file and folder names. Double Backslashes (\\) Two backslashes are used as a prefix to a server name (hostname). For example, \\a5\c\expenses is the path to the EXPENSES folder on the C: drive on server A5. See UNC, \\, path and forward slash. Platform     Path to Ovens Folder Windows:     \products\kitchen\ovens Unix/Linux, Mac OS X:    /products/kitchen/ovens Path to Ovens Page Internet:    greatproducts.com/kitchen/ovens Three Symbols Separate Names The Unix-based Mac uses the forward slash, while Windows uses the backslash. Starting with Windows Vista, the Explorer path uses right arrows; however, backslashes must still be used at the command line (see cmd abc's).
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9816159605979919, "perplexity": 4500.6612183904645}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-18/segments/1461860118321.95/warc/CC-MAIN-20160428161518-00143-ip-10-239-7-51.ec2.internal.warc.gz"}
https://hal-insu.archives-ouvertes.fr/insu-03388642
# Identifying nonlinear wave interactions in plasmas using two-point measurements: A case study of Short Large Amplitude Magnetic Structures (SLAMS) Abstract : Two fundamental quantities for characterizing nonlinear wave phenomena in plasmas are the spectral energy transfer associated with the energy redistribution between Fourier modes, and the linear growth rate. It is shown how these quantities can be estimated simultaneously from dual-spacecraft data using Volterra series models. We consider magnetic field data gathered upstream the Earth's quasiparallel bow shock, in which Short Large Amplitude Magnetic Structures (SLAMS) supposedly play a leading role. The analysis attests the dynamic evolution of the SLAMS and reveals an energy cascade toward high-frequency waves. These results put constraints on possible mechanisms for the shock front formation. Document type : Journal articles Domain : https://hal-insu.archives-ouvertes.fr/insu-03388642 Contributor : Thierry Dudok de Wit Connect in order to contact the contributor Submitted on : Tuesday, November 9, 2021 - 3:36:18 PM Last modification on : Tuesday, May 10, 2022 - 3:24:57 PM Long-term archiving on: : Thursday, February 10, 2022 - 7:10:41 PM ### File 1999JA900134.pdf Publisher files allowed on an open archive ### Citation Thierry Dudok de Wit, V. Krasnosel'Skikh, M. Dunlop, H. Lühr. Identifying nonlinear wave interactions in plasmas using two-point measurements: A case study of Short Large Amplitude Magnetic Structures (SLAMS). Journal of Geophysical Research Space Physics, American Geophysical Union/Wiley, 1999, 104 (A8), pp.17079-17090. ⟨10.1029/1999JA900134⟩. ⟨insu-03388642⟩ Record views
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9084222912788391, "perplexity": 6843.742767332774}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573908.30/warc/CC-MAIN-20220820043108-20220820073108-00753.warc.gz"}
https://www.zora.uzh.ch/id/eprint/62569/
# Varied but not necessarily random: Human performance under variability contingencies is affected by instructions Souza, Alessandra S; Pontes, Thaissa Neves R; Abreu-Rodrigues, Josele (2012). Varied but not necessarily random: Human performance under variability contingencies is affected by instructions. Learning & Behavior:e1-e13. ## Abstract The goal of the present study was to evaluate the role of verbal stimuli in the production of response variability in humans. College students were distributed into three groups and asked to type three-digit sequences. Participants in the systematic group were instructed to produce sequences according to a rule of their choice; those in the random group were instructed to produce sequences according to chance; and those in the control group were not instructed about how to produce sequences. The experiment employed an ABA design. During the A phases, low-frequent sequences were reinforced (variability contingency), whereas during the B phase, reinforcement was withdrawn (extinction). The results indicated the following: (1) The instructions were efficient at producing systematic and random-like patterns for the systematic and random groups, respectively; in the absence of instructions, a mix of both patterns was observed. (2) Behavior was sensitive to extinction independently of the instructions provided. (3) Systematic patterns favored a more equiprobable distribution of sequences across trials. (4) Reac- tion times were longer for responding in a systematic than in a random-like fashion. The present findings suggest that individual differences in meeting variability contingencies may be due, at least partially, to instructional control. ## Abstract The goal of the present study was to evaluate the role of verbal stimuli in the production of response variability in humans. College students were distributed into three groups and asked to type three-digit sequences. Participants in the systematic group were instructed to produce sequences according to a rule of their choice; those in the random group were instructed to produce sequences according to chance; and those in the control group were not instructed about how to produce sequences. The experiment employed an ABA design. During the A phases, low-frequent sequences were reinforced (variability contingency), whereas during the B phase, reinforcement was withdrawn (extinction). The results indicated the following: (1) The instructions were efficient at producing systematic and random-like patterns for the systematic and random groups, respectively; in the absence of instructions, a mix of both patterns was observed. (2) Behavior was sensitive to extinction independently of the instructions provided. (3) Systematic patterns favored a more equiprobable distribution of sequences across trials. (4) Reac- tion times were longer for responding in a systematic than in a random-like fashion. The present findings suggest that individual differences in meeting variability contingencies may be due, at least partially, to instructional control. ## Statistics ### Citations Dimensions.ai Metrics 2 citations in Web of Science® 2 citations in Scopus® ### Altmetrics Detailed statistics
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8082296848297119, "perplexity": 4104.077972957898}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247506094.64/warc/CC-MAIN-20190221172909-20190221194909-00290.warc.gz"}
https://www.physicsforums.com/threads/amperes-law-application-problem.764759/
# Homework Help: Ampere's Law Application Problem 1. Aug 5, 2014 ### BlackMelon 1. The problem statement, all variables and given/known data I want to know why the book says that when ρ>c, there is no current and the magnetic will be zero 2. Relevant equations Hø=I/2∏ρ 3. The attempt at a solution I was trying to find the magnetic between the gap by using the formula... I've found that the magnetic magnetic field between the outer conductor and the inner one will try to cancel each other... I think the book shouldn't just say that when ρ>c, there's no current and the field will be zero. It should say that the magnetic field from the inner and the outer conductor will completely cancel each other. Here is the book: https://www.mediafire.com/?f0cpqvavy1ddl29 PS: The problem is on pages 190-192 of the attached book #### Attached Files: File size: 20.4 KB Views: 189 • ###### A2.PNG File size: 9.7 KB Views: 145 Last edited: Aug 5, 2014 2. Aug 6, 2014 ### rude man I don't see the distinction between saying 'the fields cancel' and 'the currents cancel'. Both are correct. 3. Aug 7, 2014 ### BlackMelon Rude man, if the outer layer of the conductor doesn't completely enclose the inner one but it carries the same magnitude of current as the inner one in an opposite direction, will the fields cancel each other completely? (In the case that my chosen magnetic field (H) path encloses both inner and outer conductors and vice versa) File size: 12.4 KB Views: 125 4. Aug 7, 2014 ### rude man You'd have to draw me a picture of how the outer conductor somehow doesn't fully enclose the inner one. That certainly isn't the case with your illustration. I have to warn you that the only reason you get a nice homogeneous B field around a circular path enclosing a wire carrying a current is if (1) the wire is of infinite length, and (2) the symmetry of the configuration is maintained. In other words, the configuration must be symmetrical as you go around the circular path. 5. Aug 9, 2014 ### BlackMelon I've started to confuse about words from the book that I've attached to the very first post. Could you explain me the sentence at the page 191 "If the radius ρ is larger than the outer radius of the outer conductor, no current is enclosed" and how is it different from my second post? Thanks 6. Aug 9, 2014 ### rude man The image in your second post shows an asymmetry in the outer conductor. So the H field is not constant in magnitude as you go around your indicated path. But the net current through the path is still zero: I + (-I) = 0 and the contour integral ∫H*ds = 0 also. H*ds is the dot product of H and element of path length ds. H and ds are vectors. 7. Aug 29, 2014 ### BlackMelon So even a broken shield in my second post can still be a perfect magnetic shield? (Just let the shield carry the same amount of current as the conductor inside?) 8. Aug 29, 2014 ### rude man No. Only the integral ∫H*ds = 0. H will not be zero everywhere around a contour outside the cable. But the integral will. So, as you go around the cable there will be places where H > 0 and other places where H < 0 (opposite direction).
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9009071588516235, "perplexity": 911.9983675259472}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676589573.27/warc/CC-MAIN-20180717051134-20180717071134-00244.warc.gz"}
https://www.tutorialspoint.com/Tower-Of-Hanoi-Problem
Tower Of Hanoi Problem Misc AlgorithmsData StructureAlgorithms Tower of Hanoi is a puzzle problem. Where we have three stands and n discs. Initially, Discs are placed in the first stand. We have to place discs into the third or destination stand, the second or auxiliary stand can be used as a helping stand. • But there are some rules to follow. •  We can transfer only one disc for each movement. • Only the topmost disc can be picked up from a stand. • No bigger disc will be placed at the top of the smaller disc. This problem can be solved easily by recursion. At first, using recursion the top (n-1) discs are placed from source to auxiliary stand, then place the last disc from source to destination, then again place (n-1) disc from auxiliary stand to destination stand by recursion. Input and Output Input: Number of discs: 3 Output: 1. Move disk 1 from A to C 2. Move disk 2 from A to B 3. Move disk 1 from C to B 4. Move disk 3 from A to C 5. Move disk 1 from B to A 6. Move disk 2 from B to C 7. Move disk 1 from A to C Algorithm toh(n, s, a, d) Input: Number of discs, source, auxiliary, destination. Output: Steps to move discs from source to destination maintaining proper rules. Begin if n = 1, then display move disc from s to d toh(n-1, s, d, a) display move disc from s to d toh(n-1, a, s, d) End Example #include<iostream> using namespace std; void TOH(int n, char s, char a, char d) { static int count = 0;          //store number of counts if(n == 1) { count++; cout << count<< ". Move disk " << n << " from "<<s <<" to "<<d<<endl; return;      //base case, when only one disk } TOH(n-1, s, d, a);               //recursive call the function count++; cout << count<< ". Move disk " << n << " from "<<s <<" to"<<d<<endl; TOH(n-1, a, s, d); } int main() { int n; cout << "Enter the number of disks: "; cin >> n; TOH(n, 'A','B','C'); } Output Enter the number of disks: 3 1. Move disk 1 from A to C 2. Move disk 2 from A to B 3. Move disk 1 from C to B 4. Move disk 3 from A to C 5. Move disk 1 from B to A 6. Move disk 2 from B to C 7. Move disk 1 from A to C Published on 12-Jul-2018 11:40:37
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3004481792449951, "perplexity": 5237.193620175807}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662515466.5/warc/CC-MAIN-20220516235937-20220517025937-00745.warc.gz"}
http://www.mathworks.com/help/ident/ref/lsim.html?nocookie=true
Accelerating the pace of engineering and science # lsim Simulate time response of dynamic system to arbitrary inputs ## Syntax lsim lsim(sys,u,t) lsim(sys,u,t,x0) lsim(sys,u,t,x0,'zoh') lsim(sys,u,t,x0,'foh') lsim(sys) ## Description lsim simulates the (time) response of continuous or discrete linear systems to arbitrary inputs. When invoked without left-hand arguments, lsim plots the response on the screen. lsim(sys,u,t) produces a plot of the time response of the dynamic system model sys to the input time history t,u. The vector t specifies the time samples for the simulation (in system time units, specified in the TimeUnit property of sys), and consists of regularly spaced time samples. t = 0:dt:Tfinal The matrix u must have as many rows as time samples (length(t)) and as many columns as system inputs. Each row u(i,:) specifies the input value(s) at the time sample t(i). The LTI model sys can be continuous or discrete, SISO or MIMO. In discrete time, u must be sampled at the same rate as the system (t is then redundant and can be omitted or set to the empty matrix). In continuous time, the time sampling dt=t(2)-t(1) is used to discretize the continuous model. If dt is too large (undersampling), lsim issues a warning suggesting that you use a more appropriate sample time, but will use the specified sample time. See Algorithms for a discussion of sample times. lsim(sys,u,t,x0) further specifies an initial condition x0 for the system states. This syntax applies only to state-space models. lsim(sys,u,t,x0,'zoh') or lsim(sys,u,t,x0,'foh') explicitly specifies how the input values should be interpolated between samples (zero-order hold or linear interpolation). By default, lsim selects the interpolation method automatically based on the smoothness of the signal U. Finally, lsim(sys1,sys2,...,sysN,u,t) simulates the responses of several LTI models to the same input history t,u and plots these responses on a single figure. As with bode or plot, you can specify a particular color, linestyle, and/or marker for each system, for example, lsim(sys1,'y:',sys2,'g--',u,t,x0) The multisystem behavior is similar to that of bode or step. When invoked with left-hand arguments, [y,t] = lsim(sys,u,t) [y,t,x] = lsim(sys,u,t) % for state-space models only [y,t,x] = lsim(sys,u,t,x0) % with initial state return the output response y, the time vector t used for simulation, and the state trajectories x (for state-space models only). No plot is drawn on the screen. The matrix y has as many rows as time samples (length(t)) and as many columns as system outputs. The same holds for x with "outputs" replaced by states. lsim(sys) opens the Linear Simulation Tool GUI. For more information about working with this GUI, see Working with the Linear Simulation Tool. ## Examples ### Simulate Response to Square Wave Simulate and plot the response of the following system to a square wave with period of four seconds: Create the transfer function, and generate the square wave with gensig. Sample every 0.1 second during 10 seconds. H = [tf([2 5 1],[1 2 3]);tf([1 -1],[1 1 5])]; [u,t] = gensig('square',4,10,0.1); Then simulate with lsim. lsim(H,u,t) The plot displays both the applied signal and the response. ### Simulate Response of Identified Model Simulate the response of an identified linear model using the same input signal as the one used for estimation and the initial states returned by the estimation command. load(fullfile(matlabroot,'toolbox','ident','iddemos','data','dcmotordata')); z = iddata(y,u,0.1,'Name','DC-motor'); [sys,x0] = n4sid(z,4); [y,t,x] = lsim(sys, z.InputData, [], x0); Compare the simulated response y to measured response z.OutputData. plot(t,z.OutputData,'k',t,y,'r') legend('Measured','Simulated') expand all ### Algorithms Discrete-time systems are simulated with ltitr (state space) or filter (transfer function and zero-pole-gain). Continuous-time systems are discretized with c2d using either the 'zoh' or 'foh' method ('foh' is used for smooth input signals and 'zoh' for discontinuous signals such as pulses or square waves). The sampling period is set to the spacing dt between the user-supplied time samples t. The choice of sampling period can drastically affect simulation results. To illustrate why, consider the second-order model $\begin{array}{cc}H\left(s\right)=\frac{{\omega }^{2}}{{s}^{2}+2s+{\omega }^{2}},& \omega =62.83\end{array}$ To simulate its response to a square wave with period 1 second, you can proceed as follows: w2 = 62.83^2; h = tf(w2,[1 2 w2]); t = 0:0.1:5; % vector of time samples u = (rem(t,1) >= 0.5); % square wave values lsim(h,u,t) lsim evaluates the specified sample time, and issues a warning: Warning: Input signal is undersampled. Sample every 0.016 sec or faster. To improve on this response, discretize H(s) using the recommended sampling period: dt = 0.016; ts = 0:dt:5; us = (rem(ts,1) >= 0.5); hd = c2d(h,dt); lsim(hd,us,ts) This response exhibits strong oscillatory behavior that is hidden in the undersampled version.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 1, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8606640100479126, "perplexity": 2869.3231957938215}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-42/segments/1414119645329.44/warc/CC-MAIN-20141024030045-00277-ip-10-16-133-185.ec2.internal.warc.gz"}
http://ch.mathworks.com/help/optim/ug/lsqnonlin.html?nocookie=true
# lsqnonlin Solve nonlinear least-squares (nonlinear data-fitting) problems ## Equation Solves nonlinear least-squares curve fitting problems of the form $\underset{x}{\mathrm{min}}{‖f\left(x\right)‖}_{2}^{2}=\underset{x}{\mathrm{min}}\left({f}_{1}{\left(x\right)}^{2}+{f}_{2}{\left(x\right)}^{2}+...+{f}_{n}{\left(x\right)}^{2}\right)$ with optional lower and upper bounds lb and ub on the components of x. x, lb, and ub can be vectors or matrices; see Matrix Arguments. ## Syntax ```x = lsqnonlin(fun,x0)x = lsqnonlin(fun,x0,lb,ub)x = lsqnonlin(fun,x0,lb,ub,options)x = lsqnonlin(problem)[x,resnorm] = lsqnonlin(...)[x,resnorm,residual] = lsqnonlin(...)[x,resnorm,residual,exitflag] = lsqnonlin(...)[x,resnorm,residual,exitflag,output] = lsqnonlin(...)[x,resnorm,residual,exitflag,output,lambda] = lsqnonlin(...)[x,resnorm,residual,exitflag,output,lambda,jacobian] = lsqnonlin(...) ``` ## Description `lsqnonlin` solves nonlinear least-squares problems, including nonlinear data-fitting problems. Rather than compute the value ${‖f\left(x\right)‖}_{2}^{2}$ (the sum of squares), `lsqnonlin` requires the user-defined function to compute the vector-valued function $f\left(x\right)=\left[\begin{array}{c}{f}_{1}\left(x\right)\\ {f}_{2}\left(x\right)\\ ⋮\\ {f}_{n}\left(x\right)\end{array}\right]$ Then, in vector terms, you can restate this optimization problem as $\underset{x}{\mathrm{min}}{‖f\left(x\right)‖}_{2}^{2}=\underset{x}{\mathrm{min}}\left({f}_{1}{\left(x\right)}^{2}+{f}_{2}{\left(x\right)}^{2}+...+{f}_{n}{\left(x\right)}^{2}\right)$ where x is a vector or matrix and f(x) is a function that returns a vector or matrix value. For details of matrix values, see Matrix Arguments. `x = lsqnonlin(fun,x0)` starts at the point `x0` and finds a minimum of the sum of squares of the functions described in `fun`. `fun` should return a vector of values and not the sum of squares of the values. (The algorithm implicitly computes the sum of squares of the components of `fun(x)`.) Note:   Passing Extra Parameters explains how to pass extra parameters to the vector function f, if necessary. `x = lsqnonlin(fun,x0,lb,ub)` defines a set of lower and upper bounds on the design variables in `x`, so that the solution is always in the range `lb `` x `` ub`. You can fix the solution component `x(i)` by specifying `lb(i) = ub(i)`. `x = lsqnonlin(fun,x0,lb,ub,options)` minimizes with the optimization options specified in `options`. Use `optimoptions` to set these options. Pass empty matrices for `lb` and `ub` if no bounds exist. `x = lsqnonlin(problem)` finds the minimum for `problem`, where `problem` is a structure described in Input Arguments. Create the `problem` structure by exporting a problem from Optimization app, as described in Exporting Your Work. `[x,resnorm] = lsqnonlin(...)` returns the value of the squared 2-norm of the residual at `x`: `sum(fun(x).^2)`. `[x,resnorm,residual] = lsqnonlin(...)` returns the value of the residual `fun(x)` at the solution `x`. `[x,resnorm,residual,exitflag] = lsqnonlin(...)` returns a value `exitflag` that describes the exit condition. ```[x,resnorm,residual,exitflag,output] = lsqnonlin(...)``` returns a structure `output` that contains information about the optimization. ```[x,resnorm,residual,exitflag,output,lambda] = lsqnonlin(...)``` returns a structure `lambda` whose fields contain the Lagrange multipliers at the solution `x`. ```[x,resnorm,residual,exitflag,output,lambda,jacobian] = lsqnonlin(...) ``` returns the Jacobian of `fun` at the solution `x`. Note:   If the specified input bounds for a problem are inconsistent, the output `x` is `x0` and the outputs `resnorm` and `residual` are `[]`.Components of `x0` that violate the bounds `lb ≤ x ≤ ub` are reset to the interior of the box defined by the bounds. Components that respect the bounds are not changed. ## Input Arguments Function Arguments contains general descriptions of arguments passed into `lsqnonlin`. This section provides function-specific details for `fun`, `options`, and `problem`: `fun` The function whose sum of squares is minimized. `fun` is a function that accepts a vector `x` and returns a vector `F`, the objective functions evaluated at `x`. The function `fun` can be specified as a function handle to a file: `x = lsqnonlin(@myfun,x0)` where `myfun` is a MATLAB® function such as ```function F = myfun(x) F = ... % Compute function values at x``` `fun` can also be a function handle for an anonymous function. `x = lsqnonlin(@(x)sin(x.*x),x0);` If the user-defined values for `x` and `F` are matrices, they are converted to a vector using linear indexing. Note   The sum of squares should not be formed explicitly. Instead, your function should return a vector of function values. See Examples. If the Jacobian can also be computed and the Jacobian option is `'on'`, set by `options = optimoptions('lsqnonlin','Jacobian','on')` the function `fun` must return, in a second output argument, the Jacobian value `J`, a matrix, at `x`. By checking the value of `nargout`, the function can avoid computing `J` when `fun` is called with only one output argument (in the case where the optimization algorithm only needs the value of `F` but not `J`). ```function [F,J] = myfun(x) F = ... % Objective function values at x if nargout > 1 % Two output arguments J = ... % Jacobian of the function evaluated at x end``` If `fun` returns a vector (matrix) of `m` components and `x` has length `n`, where `n` is the length of `x0`, the Jacobian `J` is an `m`-by-`n` matrix where `J(i,j)` is the partial derivative of `F(i)` with respect to `x(j)`. (The Jacobian `J` is the transpose of the gradient of `F`.) `options` Options provides the function-specific details for the `options` values. `problem` `objective` Objective function `x0` Initial point for `x` `lb`Vector of lower bounds `ub`Vector of upper bounds `solver` `'lsqnonlin'` `options` Options created with `optimoptions` ## Output Arguments Function Arguments contains general descriptions of arguments returned by `lsqnonlin`. This section provides function-specific details for `exitflag`, `lambda`, and `output`: `exitflag` Integer identifying the reason the algorithm terminated. The following lists the values of `exitflag` and the corresponding reasons the algorithm terminated: `1` Function converged to a solution `x`. `2` Change in `x` was less than the specified tolerance. `3` Change in the residual was less than the specified tolerance. `4` Magnitude of search direction was smaller than the specified tolerance. `0` Number of iterations exceeded `options.MaxIter` or number of function evaluations exceeded `options.MaxFunEvals`. `-1` Output function terminated the algorithm. `-2` Problem is infeasible: the bounds `lb` and `ub` are inconsistent. `-4` Line search could not sufficiently decrease the residual along the current search direction. `lambda` Structure containing the Lagrange multipliers at the solution `x` (separated by constraint type). The fields are `lower` Lower bounds `lb` `upper` Upper bounds `ub` `output` Structure containing information about the optimization. The fields of the structure are `firstorderopt` Measure of first-order optimality `iterations` Number of iterations taken `funcCount` The number of function evaluations `cgiterations` Total number of PCG iterations (trust-region-reflective algorithm only) `stepsize` Final displacement in `x` (Levenberg-Marquardt algorithm) `algorithm` Optimization algorithm used `message` Exit message ## Options Optimization options. Set or change options using the `optimoptions` function. Some options apply to all algorithms, some are only relevant when you are using the trust-region-reflective algorithm, and others are only relevant when you are using the Levenberg-Marquardt algorithm. See Optimization Options Reference for detailed information. ### Algorithm Options Both algorithms use the following options: `Algorithm` Choose between `'trust-region-reflective'` (default) and `'levenberg-marquardt'`. Set the initial Levenberg-Marquardt parameter λ by setting `Algorithm` to a cell array such as `{'levenberg-marquardt',.005}`. The default λ = `0.01`.The `Algorithm` option specifies a preference for which algorithm to use. It is only a preference, because certain conditions must be met to use each algorithm. For the trust-region-reflective algorithm, the nonlinear system of equations cannot be underdetermined; that is, the number of equations (the number of elements of `F` returned by `fun`) must be at least as many as the length of `x`. The Levenberg-Marquardt algorithm does not handle bound constraints. For more information on choosing the algorithm, see Choosing the Algorithm. `DerivativeCheck` Compare user-supplied derivatives (gradients of objective or constraints) to finite-differencing derivatives. The choices are `'on'` or the default `'off'`. `Diagnostics` Display diagnostic information about the function to be minimized or solved. The choices are `'on'` or the default `'off'`. `DiffMaxChange` Maximum change in variables for finite-difference gradients (a positive scalar). The default is `Inf`. `DiffMinChange` Minimum change in variables for finite-difference gradients (a positive scalar). The default is `0`. `Display` Level of display:`'off'` or `'none'` displays no output.`'iter'` displays output at each iteration, and gives the default exit message.`'iter-detailed'` displays output at each iteration, and gives the technical exit message.`'final'` (default) displays just the final output, and gives the default exit message.`'final-detailed'` displays just the final output, and gives the technical exit message. `FinDiffRelStep` Scalar or vector step size factor. When you set `FinDiffRelStep` to a vector `v`, forward finite differences `delta` are```delta = v.*sign(x).*max(abs(x),TypicalX);```and central finite differences are`delta = v.*max(abs(x),TypicalX);`Scalar `FinDiffRelStep` expands to a vector. The default is `sqrt(eps)` for forward finite differences, and `eps^(1/3)` for central finite differences. `FinDiffType` Finite differences, used to estimate gradients, are either `'forward'` (default), or `'central'` (centered). `'central'` takes twice as many function evaluations, but should be more accurate.The algorithm is careful to obey bounds when estimating both types of finite differences. So, for example, it could take a backward, rather than a forward, difference to avoid evaluating at a point outside bounds. `FunValCheck` Check whether function values are valid. `'on'` displays an error when the function returns a value that is `complex`, `Inf`, or `NaN`. The default `'off'` displays no error. `Jacobian` If `'on'`, `lsqnonlin` uses a user-defined Jacobian (defined in `fun`), or Jacobian information (when using `JacobMult`), for the objective function. If `'off'` (default), `lsqnonlin` approximates the Jacobian using finite differences. `MaxFunEvals` Maximum number of function evaluations allowed, a positive integer. The default is `100*numberOfVariables`. `MaxIter` Maximum number of iterations allowed, a positive integer. The default is `400`. `OutputFcn` Specify one or more user-defined functions that an optimization function calls at each iteration, either as a function handle or as a cell array of function handles. The default is none (`[]`). See Output Function. `PlotFcns` Plots various measures of progress while the algorithm executes, select from predefined plots or write your own. Pass a function handle or a cell array of function handles. The default is none (`[]`):`@optimplotx` plots the current point.`@optimplotfunccount` plots the function count.`@optimplotfval` plots the function value.`@optimplotresnorm` plots the norm of the residuals.`@optimplotstepsize` plots the step size.`@optimplotfirstorderopt` plots the first-order optimality measure.For information on writing a custom plot function, see Plot Functions. `TolFun` Termination tolerance on the function value, a positive scalar. The default is `1e-6`. `TolX` Termination tolerance on `x`, a positive scalar. The default is `1e-6`. `TypicalX` Typical `x` values. The number of elements in `TypicalX` is equal to the number of elements in `x0`, the starting point. The default value is `ones(numberofvariables,1)`. `lsqnonlin` uses `TypicalX` for scaling finite differences for gradient estimation. ### Trust-Region-Reflective Algorithm Only The trust-region-reflective algorithm uses the following options: `JacobMult` Function handle for Jacobian multiply function. For large-scale structured problems, this function computes the Jacobian matrix product `J*Y`, `J'*Y`, or `J'*(J*Y)` without actually forming `J`. The function is of the form `W = jmfun(Jinfo,Y,flag) ` where `Jinfo` contains the matrix used to compute `J*Y `(or `J'*Y`, or `J'*(J*Y)`). The first argument `Jinfo` must be the same as the second argument returned by the objective function `fun`, for example, by `[F,Jinfo] = fun(x)` `Y` is a matrix that has the same number of rows as there are dimensions in the problem. `flag` determines which product to compute: • If `flag == 0` then `W = J'*(J*Y)`. • If `flag > 0` then ```W = J*Y```. • If `flag < 0` then ```W = J'*Y```. In each case, `J` is not formed explicitly. `lsqnonlin` uses `Jinfo` to compute the preconditioner. See Passing Extra Parameters for information on how to supply values for any additional parameters `jmfun` needs. Note   `'Jacobian'` must be set to `'on'` for `lsqnonlin` to pass `Jinfo` from `fun` to `jmfun`. `JacobPattern` Sparsity pattern of the Jacobian for finite differencing. Set `JacobPattern(i,j) = 1` when `fun(i)` depends on `x(j)`. Otherwise, set ```JacobPattern(i,j) = 0```. In other words, `JacobPattern(i,j) = 1` when you can have ∂`fun(i)`/∂`x(j)` ≠ 0. Use `JacobPattern` when it is inconvenient to compute the Jacobian matrix `J` in `fun`, though you can determine (say, by inspection) when `fun(i)` depends on `x(j)`. `lsqnonlin` can approximate `J` via sparse finite differences when you give `JacobPattern`. In the worst case, if the structure is unknown, do not set `JacobPattern`. The default behavior is as if `JacobPattern` is a dense matrix of ones. Then `lsqnonlin` computes a full finite-difference approximation in each iteration. This can be very expensive for large problems, so it is usually better to determine the sparsity structure. `MaxPCGIter` Maximum number of PCG (preconditioned conjugate gradient) iterations, a positive scalar. The default is ```max(1, numberOfVariables/2)```. For more information, see Algorithms. `PrecondBandWidth` Upper bandwidth of preconditioner for PCG, a nonnegative integer. The default `PrecondBandWidth` is `Inf`, which means a direct factorization (Cholesky) is used rather than the conjugate gradients (CG). The direct factorization is computationally more expensive than CG, but produces a better quality step towards the solution. Set `PrecondBandWidth` to `0` for diagonal preconditioning (upper bandwidth of 0). For some problems, an intermediate bandwidth reduces the number of PCG iterations. `TolPCG` Termination tolerance on the PCG iteration, a positive scalar. The default is `0.1`. ### Levenberg-Marquardt Algorithm Only The Levenberg-Marquardt algorithm uses the following options: `InitDamping` Initial value of the Levenberg-Marquardt parameter, a positive scalar. Default is `1e-2`. For details, see Levenberg-Marquardt Method. `ScaleProblem` `'Jacobian'` can sometimes improve the convergence of a poorly scaled problem; the default is `'none'`. ## Examples Find x that minimizes $\sum _{k=1}^{10}{\left(2+2k-{e}^{k{x}_{1}}-{e}^{k{x}_{2}}\right)}^{2},$ starting at the point `x = [0.3, 0.4]`. Because `lsqnonlin` assumes that the sum of squares is not explicitly formed in the user-defined function, the function passed to `lsqnonlin` should instead compute the vector-valued function ${F}_{k}\left(x\right)=2+2k-{e}^{k{x}_{1}}-{e}^{k{x}_{2}},$ for k = 1 to 10 (that is, `F` should have `10` components). First, write a file to compute the `10`-component vector `F`. ```function F = myfun(x) k = 1:10; F = 2 + 2*k-exp(k*x(1))-exp(k*x(2));``` Next, invoke an optimization routine. ```x0 = [0.3 0.4] % Starting guess [x,resnorm] = lsqnonlin(@myfun,x0); % Invoke optimizer``` After about 24 function evaluations, this example gives the solution ```x,resnorm x = 0.2578 0.2578 resnorm = 124.3622``` ## Diagnostics ### Memory and Jacobians You can use the trust-region reflective algorithm in `lsqnonlin`, `lsqcurvefit`, and `fsolve` with small- to medium-scale problems without computing the Jacobian in `fun` or providing the Jacobian sparsity pattern. (This also applies to using `fmincon` or `fminunc` without computing the Hessian or supplying the Hessian sparsity pattern.) How small is small- to medium-scale? No absolute answer is available, as it depends on the amount of virtual memory in your computer system configuration. Suppose your problem has `m` equations and `n` unknowns. If the command `J = sparse(ones(m,n))` causes an `Out of memory` error on your machine, then this is certainly too large a problem. If it does not result in an error, the problem might still be too large. You can only find out by running it and seeing if MATLAB runs within the amount of virtual memory available on your system. ### Trust-Region-Reflective Optimization The trust-region-reflective method does not allow equal upper and lower bounds. For example, if `lb(2)==ub(2)`, `lsqlin` gives the error `Equal upper and lower bounds not permitted.` `lsqnonlin` does not handle equality constraints, which is another way to formulate equal bounds. If equality constraints are present, use `fmincon`, `fminimax`, or `fgoalattain` for alternative formulations where equality constraints can be included.) ## Limitations The function to be minimized must be continuous. `lsqnonlin` might only give local solutions. `lsqnonlin` can solve complex-valued problems directly with the `levenberg-marquardt` algorithm. However, this algorithm does not accept bound constraints. For a complex problem with bound constraints, split the variables into real and imaginary parts, and use the `trust-region-reflective` algorithm. See Fit a Model to Complex-Valued Data. ### Trust-Region-Reflective Optimization The trust-region-reflective algorithm for `lsqnonlin` does not solve underdetermined systems; it requires that the number of equations, i.e., the row dimension of F, be at least as great as the number of variables. In the underdetermined case, the Levenberg-Marquardt algorithm is used instead. The preconditioner computation used in the preconditioned conjugate gradient part of the trust-region-reflective method forms JTJ (where J is the Jacobian matrix) before computing the preconditioner; therefore, a row of J with many nonzeros, which results in a nearly dense product JTJ, can lead to a costly solution process for large problems. If components of x have no upper (or lower) bounds, `lsqnonlin` prefers that the corresponding components of `ub` (or `lb`) be set to `inf` (or `-inf` for lower bounds) as opposed to an arbitrary but very large positive (or negative for lower bounds) number. Trust-Region-Reflective Problem Coverage and Requirements For Large Problems • Provide sparsity structure of the Jacobian or compute the Jacobian in `fun`. • The Jacobian should be sparse. ### Levenberg-Marquardt Optimization The Levenberg-Marquardt algorithm does not handle bound constraints. Since the trust-region-reflective algorithm does not handle underdetermined systems and the Levenberg-Marquardt does not handle bound constraints, problems with both these characteristics cannot be solved by `lsqnonlin`. collapse all ### Trust-Region-Reflective Optimization By default, `lsqnonlin` chooses the trust-region-reflective algorithm. This algorithm is a subspace trust-region method and is based on the interior-reflective Newton method described in [1] and [2]. Each iteration involves the approximate solution of a large linear system using the method of preconditioned conjugate gradients (PCG). See Trust-Region-Reflective Least Squares, and in particular Large Scale Nonlinear Least Squares. ### Levenberg-Marquardt Optimization If you set the `Algorithm` option to `'levenberg-marquardt'` using `optimoptions`, `lsqnonlin` uses the Levenberg-Marquardt method [4], [5], and [6]. See Levenberg-Marquardt Method. ## References [1] Coleman, T.F. and Y. Li, "An Interior, Trust Region Approach for Nonlinear Minimization Subject to Bounds," SIAM Journal on Optimization, Vol. 6, pp. 418–445, 1996. [2] Coleman, T.F. and Y. Li, "On the Convergence of Reflective Newton Methods for Large-Scale Nonlinear Minimization Subject to Bounds," Mathematical Programming, Vol. 67, Number 2, pp. 189-224, 1994. [3] Dennis, J.E., Jr., "Nonlinear Least-Squares," State of the Art in Numerical Analysis, ed. D. Jacobs, Academic Press, pp. 269–312, 1977. [4] Levenberg, K., "A Method for the Solution of Certain Problems in Least-Squares," Quarterly Applied Math. 2, pp. 164–168, 1944. [5] Marquardt, D., "An Algorithm for Least-Squares Estimation of Nonlinear Parameters," SIAM Journal Applied Math., Vol. 11, pp. 431–441, 1963. [6] Moré, J.J., "The Levenberg-Marquardt Algorithm: Implementation and Theory," Numerical Analysis, ed. G. A. Watson, Lecture Notes in Mathematics 630, Springer Verlag, pp. 105–116, 1977.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 6, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9084458351135254, "perplexity": 1217.2834246752138}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1430453976406.4/warc/CC-MAIN-20150501041936-00083-ip-10-235-10-82.ec2.internal.warc.gz"}
http://math.stackexchange.com/questions/902017/how-far-can-one-see-over-the-ocean
# How far can one see over the ocean? Since Earth is a sphere, one has only a limited visibility radius. How far is that, actually? This Q&A was inspired by this question, about whether or not Legolas can see the 24km distant Riders of Rohan. - I have up-voted both the question and the answer by M.Herzkamp, but I also think that answer is somewhat more complicated than it needs to be. Accordingly I've added my own answer. – Michael Hardy Aug 18 '14 at 15:23 Both answers need an additional factor of about two. You don't have to be able to see the orcs' toes, just the tops of their heads. – Pete Becker Aug 18 '14 at 16:05 The factor of two is only right, if the Orcs are of the same height as you (or elevation). As I remember, Legolas stood on a hill, while the Orcs rode over the plains of Rohan. – M.Herzkamp Aug 19 '14 at 6:34 @M.Herzkamp - that's why I said about two. – Pete Becker Aug 19 '14 at 14:08 Maybe Middle Earth has a planar geometry. :-) – peterh Aug 20 '14 at 11:45 I have up-voted the answer by M.Herzkamp, but I also think he makes it somewhat more complicated than it needs to be. The distance from the center of the earth to your eye is $r+h$, where $r$ is the radius of the earth and $h$ is the height of your eye above the ground. The distance from the center of the earth to a point on the horizon is $r$. The distance from your eye to the point on the horizon let us call $d$. The three sides of a right triangle are then the legs, $r$ and $d$, and the hypotenuse $r+h$. Applying the Pythagorean theorem, we have $$r^2 + d^2 = (r+h)^2.$$ It follows that $$d^2 = (r+h)^2 - r^2$$ so $$d=\sqrt{(r+h)^2-r^2}.$$ This admits simplifiction: $$d=\sqrt{(r+h)^2-r^2} = \sqrt{(r^2+2rh+h^2) - r^2} = \sqrt{2rh + h^2}.$$ When $h$ is tiny compared to $r$, we can say $$d \approx \sqrt{2rh\,{}}.$$ - I might be worth clarifying that any objects beyond the horizon can of course be visible if they are high or tall enough - an object at height $H$ is visible to an oberver at height $h$ at distance $d \approx \sqrt{2rh} + \sqrt{2rH}$. – JohannesD Aug 18 '14 at 16:08 @JohannesD I guess an object on the precise polar opposite of you, could be as high as you want, and you still couldn't see it. – Cruncher Aug 18 '14 at 16:34 @Red_Shadow you also need to allow your vantage point to be arbitarily high. At any finite height the planet casts a shadow covering some angle. – JHance Aug 18 '14 at 19:11 @Red_Shadow Actually to see anything (except something perfectly opposite you) both you and the object must be arbitrarily tall/high. If your height is bounded, then you can only see things that are not "below" the infinite conic surface defined by your position and the horizon circle. – JohannesD Aug 18 '14 at 19:20 A handy formula implementing the above result is $$d\,[{\rm km}]=3.56\cdot\sqrt{h\,[{\rm m}]}\ .$$ – Christian Blatter Aug 19 '14 at 18:03 Let us suppose, an observer of height $h$ stands on a perfectly spherical planet of radius $r$: Edit: here is an easier way, making use of the right angle between the line of sight and the radial ray. You can just use the definition of the cosine: $$\cos(\theta_T) = \frac{r}{r+h} \qquad \Rightarrow \qquad s = r\cdot\theta_T = r\cdot\cos^{-1}\!\!\left(\frac{r}{r+h}\right)$$ which is equivalent to the solution obtained by the complicated method. /Edit The distance $s$ to the farthest point he can then see is determined by the tangent to the semi circle through his head. If you describe the semi circle in a cartesian coordinate system by $$y^2+x^2 = r^2,$$ the observer's head is at $y=r+h,\ x=0$. To obtain the slope of the tangent, we plug the tangent equation $y=mx+r+h$ into the circle equation and solve for $x$: $$x_{1/2} = -(r+h)\frac{m}{1+m^2} \pm \sqrt{\frac{(r+h)^2m^2}{(1+m^2)^2}+\frac{r^2-b^2}{1+m^2}}$$ Those are two intersection points, and in order to have a tangent, they must be equal. That is the case, if the term under the square root is zero. The resulting equation can be solved for $m$: $$m_{\pm} = \pm \sqrt{\frac{(r+h)^2}{r^2}-1}$$ Let's take the negative solution for the tangent on the right (it does not matter), and calculate the tangent point: $$x_T = -(r+h)\frac{m_-}{1+m^2_-} = \frac{r}{r+h}\sqrt{(r+h)^2-r^2}$$ The viewing distance angle is $\theta_T = \text{asin}(x_T)$. To get the viewing distance, we observe that $$\frac{s}{2\pi r} = \frac{\theta_T}{\text{full angle}} = \frac{\theta_T}{2\pi}\text{, with angle in radian}$$ $$\Rightarrow s(h) = r\cdot\text{asin}\left(\sqrt{1-\frac{r^2}{(r+h)^2}}\right)$$ If you plot this for $h$ small compared to $r$, it resembles a square root function, and indeed, $$\lim_{h\rightarrow0^+}\frac{s(h)}{\sqrt{h}} = \sqrt{2r}$$ which means that for small heights, the viewing distance can be described as $$s(h) \approx \sqrt{2rh}$$ On Earth ($r\approx6371\text{km}$), a normal person ($h\approx1.8\text{m}$) can see the surface about 4.8km away. Not much further. If you climb a hill or tree ($h\approx 50\text{m}$), your range increases to 25km! - I'd have called it the Pythagorean theorem rather than the "circle equation". ${}\qquad{}$ – Michael Hardy Aug 18 '14 at 15:15 Now we only need to know the height of an average orc ;-) – Stefan Aug 18 '14 at 15:25 This approach, involving solving an equation for $x$, is unnecessarily complicated. I've shown a simpler way in my answer. – Michael Hardy Aug 18 '14 at 16:00 Note that this calculates the distance to the horizont where the completely smooth Earth begins to be in the way of itself. In order words, the distance where you can see the whole orc, including soles of his shoes. If you can live with seeing less of the orc, the distance is greater. – Thorbjørn Ravn Andersen Aug 19 '14 at 14:24 Holy cow this is freaking complicated. You should've thrown general relativity in there too! – Mehrdad Aug 19 '14 at 22:43 Atmospheric refraction cannot be neglected. As mentioned here the effect of this can be taken into account approximately by pretending as if the Earth's radius is larger by a factor of 7/6. This makes the distance $d$ to the horizon when the height $h$ is much less than the Earth's radius $R$ equal to $$d = \sqrt{\frac{7}{3} R h}$$ - Physicist invasion! Regardless of how true this is, it's a comment; not an answer. (Unless you provide an answer, and factor the effect due to refraction too) – Ollie Ford Aug 18 '14 at 19:19 I've updated the post. – Count Iblis Aug 18 '14 at 19:36 If you go sailing, you'll have about 5 nautical miles of visibility (1nm = 1852m). I personally find the nomogram a delightful invention: Simply draw a straight line between the height of the observer and the height of the object on the horizon (in this case = 0), then read off the geographical range. - Unfortunately the scales are very nearly too large for a 1.8m observer watching a 1.8m object. On the other hand, your chart appears to show about 5 nautical miles, so about 9.26km, which roughly matches M.Herzkamp's answer of 9.6km. – Mooing Duck Aug 19 '14 at 22:24
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8894584774971008, "perplexity": 437.7287074681372}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049282327.67/warc/CC-MAIN-20160524002122-00215-ip-10-185-217-139.ec2.internal.warc.gz"}
https://www.cfd-online.com/W/index.php?title=Isentropic_Mach_number&oldid=8154
# Isentropic Mach number The isentropic Mach number is the Mach number you would have without any losses in the flow. This property is often used to investigate the ideal surface Mach number you would have without losses and walls without any friction (slip conditions). The isentropic Mach number is for example often plotted on turbomachinery blades. The isentropic Mach number can be computed from the isentropic flow relations using the following formula: $\frac{p_0^{freestream}}{p}=(1+\frac{\gamma-1}{2}M^2)^{\frac{\gamma}{\gamma-1}}$ This gives the following formula for the isentropic Mach number: $M^{isentropic} = \sqrt{ \left( ( \frac{p_0^{freestream}}{p} ) ^{\frac{\gamma - 1}{\gamma}} - 1 \right) \frac{2}{\gamma - 1}}$ Where: $p_0^{freestream}$ is the Total pressure in the freestream outside of the boundary layers. Usually the inlet total pressure is used. $p$ is the local Static pressure. $\gamma$ is the ratio of specific heats, for air $\gamma = 1.4$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 6, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9649629592895508, "perplexity": 730.1847646016273}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560282631.80/warc/CC-MAIN-20170116095122-00237-ip-10-171-10-70.ec2.internal.warc.gz"}
https://infoscience.epfl.ch/record/205837
## Measurement of tokamak error fields using plasma response and its applicability to ITER The nonlinear response of a low-beta tokamak plasma to non-axisymmetric fields offers an alternative to direct measurement of the non-axisymmetric part of the vacuum magnetic fields, often termed ‘error fields’. Possible approaches are discussed for determination of error fields and the required current in non-axisymmetric correction coils, with an emphasis on two relatively new methods: measurement of the torque balance on a saturated magnetic island, and measurement of the braking of plasma rotation in the absence of an island. The former is well suited to ohmically heated discharges, while the latter is more appropriate for discharges with a modest amount of neutral beam heating to drive rotation. Both can potentially provide continuous measurements during a discharge, subject to the limitation of a minimum averaging time. The applicability of these methods to ITER is discussed, and an estimate is made of their uncertainties in light of the specifications of ITER’s diagnostic systems. The use of plasma response-based techniques in normal ITER operational scenarios may allow identification of the error field contributions by individual central solenoid coils, but identification of the individual contributions by the outer poloidal field coils or other sources is less likely to be feasible. Published in: Nuclear Fusion, 54, 073004 Year: 2014 Keywords: Laboratories: SPC CRPP
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.816987156867981, "perplexity": 1476.6387871575016}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676589237.16/warc/CC-MAIN-20180716080356-20180716100356-00430.warc.gz"}
https://www.physicsforums.com/threads/proofs-proof-by-contraposition.645981/
# Homework Help: [Proofs] Proof by contraposition 1. Oct 21, 2012 ### twoski 1. The problem statement, all variables and given/known data For n ∈ Z+ prove by contrapositive that if 2n3 + 3n2 + 4n + 5 is odd then n is even. 2. Relevant equations 3. The attempt at a solution If n is odd then ( 2n³ + 3n² + 4n + 5 ) is even. By definition, a number n is odd if n = 2k + 1 for some integer k. A number n is even if n = 2k for some integer k. If x and y are two integers for which x + y is even, then by definition x and y have the same parity. It follows that two integers will have the same parity if they are both odd or both even. 2(2k + 1)³ + 3(2k + 1)² + 4(2k + 1) + 5 We see here that 2(2k + 1)³ can be written as 2n where n = (2k+1)³. Thus we conclude 2(2k + 1)³ is an even number. By the same logic, 4(2k + 1) can be written as 2n where n = 2(2k+1). I get stuck after this... How can i prove 3(2k + 1)² is odd so that i can add it to 5 and get even parity? 2. Oct 22, 2012 ### Staff: Mentor I would expand 2(2k + 1)³ + 3(2k + 1)² + 4(2k + 1) + 5 and see what I got. 3. Oct 22, 2012 ### LCKurtz Multiply it out, add 5 to it, and look at what you get.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9261255264282227, "perplexity": 693.0691091039369}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267156613.38/warc/CC-MAIN-20180920195131-20180920215531-00495.warc.gz"}
https://de.zxc.wiki/wiki/Spektrallinie
# Spectral line Lines of a spectrum of emitted ( emission lines ) or absorbed ( absorption lines ) of electromagnetic waves , in the narrower sense within the wavelength range of visible light ( light spectrum ) , are defined as spectral lines or resonance lines . Spectral lines are characterized by wavelength , line intensity and line width . The cause of the spectral lines are the light-excited electronic transitions inAtoms or molecules . Historically, the name spectral line goes back to the fact that there is an entrance slit in conventional spectrometers , the shape of which is reflected on the screen or in the eye of the beholder. The name was later also transferred to the peaks (i.e. maxima) in a spectrum recorded as an intensity curve. Spectrum of a low-pressure cadmium vapor lamp, upper image with a 256-pixel line sensor , lower image with a camera Spectral lines appear in instrumental atomic spectroscopy (such as nuclear magnetic resonance spectroscopy ) or in flame coloration . They are used, among other things, in astronomy to analyze the molecular structure of stars , planets and interstellar matter , which would otherwise be impossible. They were in the refraction of light from the sun through a prism discovered in the 19th century, from which then spectroscopes developed that allow a wide field of applications for the spectral analysis emerged. ## Basics Continuous light spectrum without spectral lines A spectral line is the light of a precisely defined frequency that is emitted (emitted) or absorbed (absorbed) by an atom or molecule due to a transition from one energy level to another. The frequency is determined by the energy of the emitted or absorbed photon ; this is equal to the difference between the energies of the two quantum mechanical states . The frequency is characteristic of this particular transition in the given type of atom. Therefore, one can distinguish types of atoms by observing spectral lines. ### Emission line Emission lines An emission line shows up as a light line in the spectrum. It arises during the transition from a higher to a lower energy level , for example when an electron passes from an excited state to the ground state. A photon is emitted here. This can either happen spontaneously ( spontaneous emission ) or, e.g. B. with laser , are excited by light of a suitable frequency ( stimulated emission ). ### Absorption line Absorption lines Resonance absorption of H 2 O gas at 1519 nm When light is irradiated with a continuous spectrum (i.e. a frequency mixture), resonance absorption of photons of a suitable frequency results in an absorption line in which a transition from a lower to a higher energy level is induced - for example, when an electron passes through the photon from the valence band into the Conduction band is "lifted" (see photoelectric effect ). When falling back to the lower energy level, photons become isotropic , i.e. H. in any direction, emitted. Both of these lead to the light being diffusely scattered by the material being irradiated at this frequency . As long as there are enough absorbing atoms, it comes about • a dark line in the continuous spectrum of light shining through ( Fraunhofer line ); this is mostly meant by the term absorption line • a bright line against a dark background when analyzing the scattered light exiting the side of the gas; this type of light line is not called an emission line due to historical terminology ; Lines are only designated as such if the excitation was not made by light of the same frequency . ## Emission profiles The light of a spectral line does not contain a single, sharply defined frequency, but comprises a (narrow) frequency range. The half- width of this area is called the line width . The line width of an emission line is made up of several contributions: Lorentz profile The natural line width results from the lifetime of the initial state through Heisenberg's uncertainty principle . This has the shape of a Lorentz curve . It is not possible to reduce this. Gaussian profile The thermal movement of the atoms creates a Doppler effect that shifts the light from an individual atom or molecule to red or blue, depending on the direction of movement. The statistical movement results in a broader frequency distribution overall. This effect is called Doppler broadening . It has the form of a Gaussian curve and is dependent on the temperature . Usually the Doppler width clearly dominates over the natural line width. The mechanism is also known as inhomogeneous line broadening . Voigt profile During a measurement, a Lorentz curve of finite width appears to have changed from its known form if the apparatus function of the measuring arrangement has a half- width in the order of magnitude of the Lorentz curve under consideration. The line shape can then be described by the convolution of the Lorentz curve and the apparatus function. If the apparatus function is a Gaussian curve, the result of the convolution is called a Voigt profile . ## history Absorption lines were first discovered in 1802 by William Hyde Wollaston and in 1814, independently of him, by Joseph von Fraunhofer in the spectrum of the sun . These dark lines in the solar spectrum are also called Fraunhofer lines . The spectral lines, among other effects, contributed to the development of quantum mechanics . According to classical electrodynamics , an electron bound in an atom could emit electromagnetic waves of any frequency; the existence of discrete lines could not be explained classically. The discovery that the frequencies of the spectral lines of the hydrogen atom are proportional to an expression of the shape with whole numbers and led to the concept of the quantum number and finally brought Niels Bohr to his Bohr model of the atom , the first - now obsolete - quantum mechanical atom model. Modern quantum mechanics can predict the spectral lines of atoms with a very high degree of accuracy. ${\ displaystyle (1 / n ^ {2} -1 / m ^ {2})}$${\ displaystyle m}$${\ displaystyle n}$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 3, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.890322208404541, "perplexity": 479.65170739705724}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446706285.92/warc/CC-MAIN-20221126080725-20221126110725-00251.warc.gz"}
https://blog.computationalcomplexity.org/2009/08/what-is-least-boring-number-not-usual.html
## Tuesday, August 18, 2009 ### What is the least boring number (NOT the usual paradox) What is the first boring natural number? I am NOT going to present that the first boring natural number is interesting crap, which may qualify as the most boring paradox. The question is, of course, ill defined. I will define it a little better by only considering mathematical properties of numbers. (e.g., 7 is interesting because there are 7 days of the week will not work.) Here are my opinions, and my opinions of my opinions. I will write WEAK if I think the justification for calling that number interesting is weak. In those cases if you know a better one, then comment on it. 1. 1 is interesting as it is the multplicative identity. 2. 2 is interesting because it is the only even prime. Also the first prime. 3. 3 is interesting because it is the first odd prime. Also the first Mersenne prime. 4. 4 is interesting because it is the first non-trivial square. Also it is the first number that is the sum of two primes. 5. 5 is interesting because it is the first number that is the sum of two distinct squares and the first number that is the sum of two distinct primes. (WEAK) 6. 6 is the first perfect number (though there are so few perfect numbers that ALL of them are interesting.) 7. 7 is the first number such that the number of squares needed to add up to it is 4 (All numbers are the sum of 4 or less squares. There are an infinite number of numbers that require 4 squares: all of the numbers congruent to 7 mod 8.) 8. 8 is the first non-trivial cube. 9. 9 is the first non-trivial odd square. (weak) 10. 10 is the first number that is the sum of two distinct odd squares. First triangular number that is the sum of 3 squares. (weak) have not been able to come up with anything interesting about 11. I could say that 11 is the first number that is the sum of 2 distinct numbers in 5 different ways. But that seems very weak: every number of the form 2n+1 is the first number that is the sum of 2 distinct numbers in n ways. Also, every number of the form 2n is the first number that is the sum of 2 numbers in n different ways. If we allowed that definition of interesting then all numbers would be interesting. 1. http://www2.stetson.edu/~efriedma/numbers.html 2. 5 is interesting also because Z[\sqrt{-5}] isn't a UFD; 5 is the smallest integer with this property. A_5 is the smallest nonabelian simple group. In group theory, 11 is related to the smallest Mathieu group (the smallest sporadic simple group) -- it's one of only four special cases of 4-transitive permutation groups. 8 is extremely interesting (more or less because it divides 24, which is probably the MOST interesting number) -- the E_8 lattice is one example. My vote for the first boring number is 9, which doesn't have nearly as many interesting algebraic, geometric, etc. properties as the numbers around it. 3. How many times do these numbers appear in Sloane's Encyclopedia of Integer Sequences? I searched for 1, 2, ..., 31. As you might expect there's a decreasing trend. But 9 is the first number to appear less often than its successor (9 appears 53266 times, 10 appears 57972 times). This doesn't necessarily mean 9 is uninteresting, though; it might just mean 10 is interesting. And of course there are flaws in this method. For example there are 10 hits for 196884 -- which occurs in the study of the Monster group -- but most of them are really "the same" as there are a lot of trivial variations on the same sequence, or sequences that don't include 196884 at all but include it in the explanatory text. 4. Yes, that's an unusual paradox: "The least boring number is boring" 5. In finite geometry, 9 is the order of the smallest Non-Desarguesian projective plane, and also the first order for which the projective plane is non-unique. 6. The integers -1,0,1 are extremely interesting because of their role in fields. The integer 2 is extremely interesting, especially in computer science, because it is the least prime and because (-1)^2 = 1. No other integers are anywhere near as interesting as {-1,0,1,2}, so I claim 3 is the least boring natural number. By the way it's silly to say that 2 is interesting because it is the only even prime. That is equivalent to saying that 2 is the only prime divisible by 2, which would be true if "2" were replaced by any other prime! The number 2 is so important that there are special words for numbers that are divisible by it, but that is a consequence of its being interesting, not the reason. 7. 4 is the first composite number 8. There's also this list... 9. It's worth mentioning that numbers that require four squares aren't just those congruent to 7 mod 8, but all those of the form 4^a (8b + 7). I happen to collect lists of uninteresting numbers on my website, so I found this post particularly interesting.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8445063233375549, "perplexity": 289.56523097088825}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794865411.56/warc/CC-MAIN-20180523024534-20180523044534-00455.warc.gz"}
http://starlink.eao.hawaii.edu/devdocs/sun211.htx/sun211ss497.html
### Interval A region representing an interval on one or more axes of a Frame #### Description: The Interval class implements a Region which represents upper and/or lower limits on one or more axes of a Frame. For a point to be within the region represented by the Interval, the point must satisfy all the restrictions placed on all the axes. The point is outside the region if it fails to satisfy any one of the restrictions. Each axis may have either an upper limit, a lower limit, both or neither. If both limits are supplied but are in reverse order (so that the lower limit is greater than the upper limit), then the interval is an excluded interval, rather than an included interval. Note, The Interval class makes no allowances for cyclic nature of some coordinate systems (such as SkyFrame coordinates). A Box should usually be used in these cases since this requires the user to think about suitable upper and lower limits, #### Inheritance The Interval class inherits from the Region class. #### Attributes The Interval class does not define any new attributes beyond those which are applicable to all Regions. #### Functions The Interval class does not define any new functions beyond those which are applicable to all Regions.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8188888430595398, "perplexity": 505.05353411196535}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141711306.69/warc/CC-MAIN-20201202144450-20201202174450-00241.warc.gz"}
https://www.biostars.org/p/416799/#416862
error reading bed files for bedtools intersect 2 0 Entering edit mode 2.0 years ago rthapa ▴ 50 Hi, I am trying to use bedtools intersect to find the overlapping gene regions with corresponding SNPs. I have two files; SNP file and gene annotation file. I am getting the following error while using bedtools intersect. I do not see extra tab at the end of the lines, I am not able to find the problem. Appreciate any suggestion. Thanks! Error: Type checker found wrong number of fields while tokenizing data line. Perhaps you have extra TAB at the end of your line? Check with "cat -t" chr^Istart^Iend^Igenes chr1^I1951^I2616^IS.001G000100 chr1^I11180^I14899^IS.001G000200 chr1^I23399^I24152^IS.001G000300 chr1^I22391^I42443^IS.001G000400 chr1^I52891^I53594^IS.001G000501 chr1^I53781^I63305^IS.001G000700 chr1^I62892^I69306^IS.001G000800 chr1^I79159^I81636^IS.001G000900 chr1^I81932^I83350^IS.001G001000 chr^Ipos^Ieffect^ISNP chr9^I57068854^I0.355187213^IS1 chr9^I57068854^I21.59969981^IS2 chr9^I57068854^I0.326924349^IS3 chr6^I13897772^I^I3.351266271^IS4 chr9^I57068854^I18.61550849^IS5 chr2^I2244737^I1.158934285^IS6 chr9^I57068854^I26.81277167^IS7 chr2^I2244737^I5.017257342^IS8 chr9^I57054157^I26.5431411^IS9 bedtools intersect • 1.6k views 0 Entering edit mode what is the output of cat genesgff1.bed SNPs.bed| awk -F '\t' '{print NF;}' | uniq | sort | uniq 0 Entering edit mode Thank you, I did remove the headers from both bed files. The output of cat genesgff1.bed SNPs.bed| awk -F '\t' '{print NF;}' | uniq | sort | uniq is 4 5 0 Entering edit mode 2.0 years ago The output of is 4 5 some lines have more than 4 fields. you can display those lines using: awk -F '\t' '(NF!=4) {print FILENAME,NF,NR, $0;}' genesgff1.bed SNPs.bed ADD COMMENT 0 Entering edit mode There were many lines with more than 4 fields. So, to simplify the file, I removed the last two columns and retained only the first two columns. cat -t SNPs.bed Chr09^I57068854 Chr09^I57068854 Chr09^I57068854 Chr06^I13897772 Chr09^I57068854 Chr02^I2244737 Chr09^I57068854 Chr02^I2244737 Chr09^I57054157 Chr01^I60532342 awk -F '\t' '(NF!=2) {print FILENAME,NF,NR,$0;}' SNPs.bed awk -F '\t' '(NF!=4) {print FILENAME,NF,NR, $0;}' genesgff1.bed didn't display anything whichI think is expected. But the bedtools is still not working, bedtools intersect -a genesgff1.bed -b SNPs.bed -wa > output.txt Error: unable to open file or unable to determine types for file SNPs.bed - Please ensure that your file is TAB delimited (e.g., cat -t FILE). - Also ensure that your file has integer chromosome coordinates in the expected columns (e.g., cols 2 and 3 for BED). ADD REPLY 0 Entering edit mode 2.0 years ago same idea, check the lines where there is no integer in column 2 and 3 awk -F '\t' '(int($2)==0 || int($3)==0) {print FILENAME,NF,NR,$0;}' SNPs.bed 0 Entering edit mode It printed out number of lines. Is the error is due to this? How can I deal with this? SNPs.bed 2 1 Chr09 57068854 SNPs.bed 2 2 Chr09 57068854 SNPs.bed 2 3 Chr09 57068854 SNPs.bed 2 4 Chr06 13897772 SNPs.bed 2 5 Chr09 57068854 SNPs.bed 2 6 Chr02 2244737 SNPs.bed 2 7 Chr09 57068854 SNPs.bed 2 8 Chr02 2244737 0 Entering edit mode It looks like it printed out every line of the file. 0 Entering edit mode what is the output of file SNPs.bed genesgff1.bed ? 0 Entering edit mode SNPs.bed: ASCII text genesgff1.bed: ASCII text 0 Entering edit mode there are only two fields in line 1 of file SNPs.bed (see also the other lines). 0 Entering edit mode awk -F '\t' '(int($2)==0 || int($3)==0) {print FILENAME,NF,NR, \$0;}' SNPs.bed It didn't display anything, which I think is correct. But still, bedtools intersect is not working. 0 Entering edit mode It looks like the problem is from SNPs.bed file. In this file I can only have two columns - 1) Chromosome number and 2) SNP position but looks like bed file needs atleast 3 columns. I want to find the corresponding genes from gene annotation file by comparing gene location and SNP position. Do you have any suggestions to improve the file format and make it acceptable for bedtools?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.19358858466148376, "perplexity": 14972.678405267878}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320304134.13/warc/CC-MAIN-20220123045449-20220123075449-00501.warc.gz"}
https://portalrecerca.uab.cat/en/publications/changes-in-malaria-epidemiology-in-a-rural-area-of-cubal-angola
# Changes in malaria epidemiology in a rural area of Cubal, Angola Fernando Salvador, Yolima Cossio, Marta Riera, Adrián Sánchez-Montalvá, Cristina Bocanegra, Jacobo Mendioroz, Arlette N. Eugenio, Elena Sulleiro, Warren Meredith, Teresa López, Milagros Moreno, Israel Molina Research output: Contribution to journalArticleResearchpeer-review 5 Citations (Scopus) ## Abstract © 2015 Salvador et al.; licensee Biomed Central. Background: Scarce information about malaria epidemiology in Angola has been published. The objective of this study is to describe the epidemiology of malaria at the Hospital Nossa Senhora da Paz (Cubal, Angola) and the fatality rate due to malaria (total and in children under five years) in the last five years. Methods: A retrospective, observational study was performed at the Hospital Nossa Senhora da Paz, a 400-bed rural hospital located in Benguela Province of Angola. The study population included all patients who attended the hospital from January 2009 to December 2013. Outcome variables were calculated as follows: the percentage of malaria cases (number of positive thick blood films, divided by the total thick blood films performed); the percentage of in-patients for malaria (number of in-patients diagnosed with malaria, divided by the total number of in-patients); and, the fatality rate (number of deaths due to malaria divided by the number of positive thick blood films). Results: Overall, 23,106 thick blood films were performed, of which 3,279 (14.2%) were positive for Plasmodium falciparum infection. During this five-year period, a reduction of 40% (95% CI 37-43%, p < 0.001) in the malaria-positive slides was detected. Distribution of positive-malaria slides showed a seasonal distribution with a peak from December to March (rainy season). An average annual reduction of 52% (95% CI 50-54%, p <0.001) in the admissions due to malaria was observed. The overall fatality rate due to malaria was 8.3%, and no significant differences in the annual fatality rate were found (p =0.553). Conclusions: A reduction in the number of malaria cases and the number of admissions due to malaria has been observed at the Hospital Nossa Senhora da Paz, during the last five years, and incidence along the study period showed a seasonal distribution. All this information could be useful when deciding which malaria control strategies have to be implemented in this area. Original language English 21 Malaria Journal 14 1 https://doi.org/10.1186/s12936-014-0540-z Published - 21 Jan 2015 ## Keywords • Angola • Benguela • Epidemiology • Malaria • Plasmodium falciparum ## Fingerprint Dive into the research topics of 'Changes in malaria epidemiology in a rural area of Cubal, Angola'. Together they form a unique fingerprint.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8102946877479553, "perplexity": 10566.869105449836}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103033816.0/warc/CC-MAIN-20220624213908-20220625003908-00381.warc.gz"}
https://phys.libretexts.org/Core/Astronomy/Cosmology/Conceptual_Problems_in_Quantum_Gravity
Skip to main content $$\require{cancel}$$ # Conceptual Problems in Quantum Gravity Why is quantum gravity hard? There are a lot of particular answers, but most, if not all, of them, have the same root. According to general relativity, gravity is a characteristic of the structure of spacetime, so quantum gravity means quantizing spacetime itself. In a very basic sense, we have no idea what this means. For instance: 1. As a probabilistic theory, quantum mechanics gives time a special role: we would like to say, for example, that an electron has a total probability of one of being somewhere in the Universe at a given time. But if spacetime is quantized, we don't know what "at a given time" means. 2. We can try to define time as "the reading of a clock." But any quantum mechanical clock made of ordinary matter has a finite (though small) probability of occasionally running backwards. 3. We would like to require that causes precede effects; this is, in fact, a basic axiom in the construction of quantum field theories. But if spacetime is subject to quantum fluctuations, the notion of "preceding" gets smeared out, and we no longer know how to make sense of causality. We can't say "A comes before B" any more; the best we seem to be able to do is to say "A probably comes before B." 4. Typical observables in quantum theories are things like "the field at a given point x." But if spacetime is quantized, we no longer know what it means to talk about "a given point." It is probably true that all, or almost all, observables in quantum gravity are nonlocal, and we know very little about how to deal with such objects. 5. It's likely that the structure of spacetime itself at very small distances is quite different from what we're used to. For instance, there are variations on the uncertainty principle that seem to imply that there is a minimum observable length. (Observing small distances requires high momenta, which then curve spacetime and distort the distances.) Other hints suggest that spacetime loses dimensions at very short distances, perhaps becoming effectively two-dimensional. But we don't know what to replace our ordinary picture of spacetime with. 6. Typically, quantum field theories can't be solved exactly, and must be approached with systematic approximation methods, or "perturbation theory." For quantum gravity, ordinary perturbative methods start with flat spacetime and treat curvature as a small distortion. But there is no reason to believe that flat spacetime -- or any simple, smooth manifold at all -- is a good approximate solution to quantum gravity. I should stress that people who work on quantum gravity don't, for the most part, spend much time thinking about such problems in the abstract. Rather, they try various approaches that work elsewhere, run into some difficulty that's rooted in these conceptual problems, and try to solve it in a concrete instance; or they look for ways to reformulate general relativity or quantum mechanics to make these issues less important; or they look for simpler models in which some of these problems occur but may be solvable. There are lots of vague ideas about what quantum gravity might look like; the hard part is in getting any of them to actually work in detail. (For a nice review paper by Chris Isham on some of the conceptual issues in quantum gravity, go here.)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7434243559837341, "perplexity": 366.7277238855395}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218187144.60/warc/CC-MAIN-20170322212947-00547-ip-10-233-31-227.ec2.internal.warc.gz"}
https://edreports.org/reports/detail/ck-12-interactive-middle-school-math-for-ccss-2021/seventh-grade
## CK-12 Interactive Middle School Math for CCSS ##### v1.5 ###### Usability Our Review Process Showing: ### Overall Summary The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for Alignment to the CCSSM. In Gateway 1, the materials meet expectations for focus and coherence, and in Gateway 2, the materials meet expectations for rigor and practice-content connections. ###### Alignment Meets Expectations ###### Usability Partially Meets Expectations ### Focus & Coherence The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for focus and coherence. For focus, the materials assess grade-level content and provide all students extensive work with grade-level problems to meet the full intent of grade-level standards. For coherence, the materials are coherent and consistent with the CCSSM. ##### Gateway 1 Meets Expectations #### Criterion 1.1: Focus Materials assess grade-level content and give all students extensive work with grade-level problems to meet the full intent of grade-level standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for focus as they assess grade-level content and provide all students extensive work with grade-level problems to meet the full intent of grade-level standards. ##### Indicator {{'1a' | indicatorName}} Materials assess the grade-level content and, if applicable, content from earlier grades. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for assessing grade-level content. Overall, assessments are aligned to grade-level standards, and the instructional materials do not assess content from future grades. Each chapter has an End of Chapter Assessment in both Word and PDF formats. Examples of End of Chapter Assessment items aligned to grade-level standards include: • In Chapter 2, Item 1 states, “There is a wall in your living room that is 13 feet wide. Your TV is centered on the wall. You want to hang up two pictures on either side of the TV. Each picture should be 4.25 feet away from the center of the TV.  a. Label where the center of the TV is on the number line above.  b. Write two expressions to calculate where each picture will go. Then label their positions on the number line.” (7.NS.1) • In Chapter 3, Item 4 states, “Gary’s birthday is coming up! He wants to book the party room at his favorite restaurant. The restaurant requires him to spend a minimum of $200 in order to book the party room. Each person’s meal costs$12.50. He also plans to buy a cake from the restaurant for $45. a. Using the minimum requirement, write an inequality to represent the total amount he will spend at the restaurant for p people. b. Solve the inequality for p. Graph the solution on a number line. c. What is the least number of people you can have at your party? Show your Reasoning. Hint: Remember that you can’t have a fraction of a person.” (7.EE.4b) • In Chapter 4, Item 2 states, “At a restaurant, a burger is$10.99 and a drink is $2.00. a. The tax rate is 8.5%. What is the cost of a burger and drink including tax? Round your answer to the nearest cent. b. If you want to tip your waiter 20% of the total cost, including tax, how much should you tip? Round your answer to the nearest cent.” (7.RP.3) • In Chapter 7, Item 2 states, “Jessie is playing a game where a red ball, a blue ball and a gold ball are each placed under a cup. Then the three cups are shuffled around. If Jessie correctly guesses which cup has a gold ball underneath it she wins the game. a. If Jessie has no idea which cup the ball is under, what is the probability that she guesses correctly? Write your answer as a fraction in simplest form. If Jessie plays the game 12 times, how many times would you expect her to guess correctly?” (7.SP.6,7) • In Chapter 8, Item 1a states, “A news reporter wants to estimate how many of the 900,000 registered voters in Rhode Island plan to vote for the candidate Fiona Rodriguez in the state senatorial race. The reporter goes to a mall in a major city and surveys 50 people. He finds that 35% of people plan to vote for Fiona Rodriguez. Why might this sample not be representative of the population?” (7.SP.1) ##### Indicator {{'1b' | indicatorName}} Materials give all students extensive work with grade-level problems to meet the full intent of grade-level standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for giving all students extensive work with grade-level problems to meet the full intent of grade-level standards. All lessons contain a Warm-Up, two or more activities, Extension Activities, Inline Questions, and Review Questions that are at grade level. Inline Questions range in number, and lessons generally contain around 10, which are used throughout the lesson to check for understanding. Also, there are Supplemental Questions and Extension Activities. These questions and activities are only seen in the Teacher’s Edition. The Review Questions are mostly multiple choice, and there are approximately 10 per lesson. Examples include: 7.RP.A, Analyze proportional relationships and use them to solve real-world and mathematical problems. • In Lesson 1.6, Activity 1, Question 2 states, “If c = 0.79s is the equation that represents the proportional relationship in the interactive, what is k?” (7.RP.A) • In Lesson 4.2, Activity 2, Question 2 states, “Jake also has a monthly car note of$575. About what percent of his income is this?” (7.RP.3) 7.NS.1, Apply and extend previous understandings of addition and subtraction to add and subtract rational numbers. • In Lesson 2.3, the Warm Up states, “Think about the three cases of integer addition: when the addends (the two numbers that are being added) are both positive, both negative, or one of each. Discuss how you approach an addition problem differently in each of these cases and what sign the answer has.” (7.NS.1) • In Lesson 2.6, Question 7 states, “Solve -4 + 5 - 7” (7.NS.1) 7.EE.B, Solve real-life and mathematical problems using numerical and algebraic expressions and equations. • In Lesson 3.10, Activity 2 states, “Are there any values of n such that n + 1 < n?” (7.EE.4b) • In Lesson 6.7, Activity 1, Question 4 states, “If you are looking at a square pyramid that has a base length of x, a slant height of y, and the pyramid height of z, which of the following are true?  a)The area of the base is 4x.  b) The area of the base is x^2c) The area of a triangular face is 12xy.  d) The area of a triangular face is 12xz.” (7.EE.4a) The full intent of the standards can be found in the progressions of the chapters and lessons, for example: • In Lesson 4.2, students solve multiple word problems related to percentages. For example, in Activity 1, Question 1, students identify how percent is found, “Which of the following can be used to find 60% of 450 customers?  A. 450 - (45 + 45 + 45 + 45)  b. 225 + 45  c. 45 + 45 + 45 + 45 + 45 + 45  d. 225 - 45”. In Activity 3, Question 1, students use ratios and percentages to solve the problems, “Jake pays $1625 per month in rent. He makes$6735 per month. Which equations can be used to correctly calculate the percent of Jake's salary on rent each month?” (7.RP.3) • In Lesson 7.10, students solve multiple problems involving probability. In Activity 2, Question 4, students reason about the likelihood of probability by frequency. The question states, “The outcomes are not equally likely because the likelihood of getting a hit is less than 0.5. About one-third (0.333) of the time, the batter will hit the ball, and two-thirds of the time, he will not. Using a die roll to simulate an at-bat, what will the numbers 1 through 6 represent?” (7.SP.7) #### Criterion 1.2: Coherence Each grade’s materials are coherent and consistent with the Standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for coherence. The majority of the materials, when implemented as designed, address the major clusters of the grade, and the materials have supporting content that enhances focus and coherence simultaneously by engaging students in the major work of the grade. The materials also include problems and activities that serve to connect two or more clusters in a domain or two or more domains in a grade. The materials partially have content from future grades that is identified and related to grade-level work and relate grade-level concepts explicitly to prior knowledge from earlier grades. ##### Indicator {{'1c' | indicatorName}} When implemented as designed, the majority of the materials address the major clusters of each grade. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations that, when implemented as designed, the majority of the materials address the major clusters of each grade. • The approximate number of chapters devoted to major clusters of the grade is four out of eight, which is approximately 50%. • The number of lessons devoted to major clusters of the grade (including supporting clusters connected to the major clusters) is 53 out of 80, which is approximately 66%. • The number of days devoted to major clusters (including assessments and supporting clusters connected to the major clusters) is 58 out of 88, which is approximately 66%. A day-level analysis is most representative of the instructional materials, because this calculation includes assessment days that represent major clusters. As a result, approximately 66% of the instructional materials focus on major clusters of the grade. ##### Indicator {{'1d' | indicatorName}} Supporting content enhances focus and coherence simultaneously by engaging students in the major work of the grade. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations that supporting content enhances focus and coherence simultaneously by engaging students in the major work of the grade. Supporting standards/clusters are connected to the major standards/clusters of the grade. Lessons in Grade 7 incorporate supporting standards in ways that support and/or maintain the focus on major work standards. Examples of the connections between supporting and major work include the following: • Lesson 1.9 connects 7.G.1 and 7.RP.A. Students work with scaled drawings and solve problems with the use of proportions. For example, in Activity 1, the Teacher’s Edition directions are, “Discuss with students why scale drawings are proportional relationships. Why are they included in this chapter?... Also we can write the scale factor into an equation of proportionality.” Also, in the Activity 2 Interactive, students use proportions to create a scaled drawing. The problem states, “Tiana wants to rearrange the furniture in her bedroom. She knows that it is 12 feet by 10 feet, with a door and window on the opposite wall. In her bedroom, she needs her bed, a desk, bookshelf, dresser, and chair. She is drawing a scale mapping on graph paper and decides to make 1 foot = 2 squares (0.50 inch). Use the interactive to rearrange her furniture.” • Lesson 4.10 connects 7.G.1 and 7.RP.3. Students use proportional relationships to solve problems involving scale and percent. For example, in Activity 1, Inline Question 1 states, “Which methods correctly convert a scale factor of 45 to a percent?” In Activity 2, students scale an image, “A scale model of a car was built at 12.5% scale. If the width of the model car is 9.75 inches, what is the width of the real car?” • Lesson 7.6 connects 7.RP.2 and 7.SP.7. Students find the probability of possible outcomes with objects that have constant proportions. For example, in Activity 1 Interactive, students use the proportions of a dart board to predict probability. The problem states, “Below is a very basic dartboard, without a bullseye. There are 20 sections on the dartboard, and you have an equal likelihood of hitting any of them. The numbers around the circle represent the point values for each section if you are keeping score. Find probabilities for different sets of numbers (numbers less than 8, multiples of 3, etc) by clicking on the sections. The fraction will adjust accordingly.” • Lesson 8.10 connects 7.SP.2 and 7.RP. Students examine samples to make generalizations using ratio understanding. The Warm Up states, “You will be using sample proportions to estimate the population proportions. A sample proportion is the ratio of members in the sample with a certain attribute to the total number of members in the sample.” ##### Indicator {{'1e' | indicatorName}} Materials include problems and activities that serve to connect two or more clusters in a domain or two or more domains in a grade. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for including problems and activities that serve to connect two or more clusters in a domain or two or more domains in a grade. Examples include: • Lesson 2.3 connects 7.NS.A with 7.EE.B. For example, in Activity 3, Stock Market states, “A particular stock started at $11.15 at the beginning of the day. After 3 hours, it was up$0.67 and at the end of the day it was down $2.25. What was the value of the stock at the end of the day? Use the number line interactive below to help you find the answer.” Also, Inline question 5 states, “On Monday, a stock rose$.50. On Tuesday, it dropped $.15. On Wednesday, it dropped$.12. Thursday it rose $.42. Friday it fell$.07. What was the overall loss?” • Lesson 3.4 connects 7.EE.A with 7.EE.B. Students rewrite expressions in different forms while using variables to represent numbers. For example, Activity 3 Interactive states, “Your math teacher, Miss Nomer, gives you an extra credit problem to figure out her age. Half of her age three years ago is equal to one-third of her age nine years from now. How old is she currently? All the pieces to figure out Miss Nomer's age are in the box below. Your job is to make two equivalent expressions from the clues above. Set them equal to each other so you can determine her age and get extra credit.” • Lesson 4.3 connects 7.NS.A with 7.RP.A. Students use operations to solve real-world problems involving percentages. For example, Warm-up Inline Question 1 states, “The original value of a house is $300,000. The house is now worth$600,000. Which of the following statements are true?  a. The new price is 200% of the old price.  b. The new price is 2 times the old price.  c. The new price is 300% of the old price.  d The amount of increase was $300,000.” • Lesson 4.6 connects 7.RP.A with 7.NS.A. Students interpret and calculate percent error and absolute error. In Activity 1, Inline Question 3 states, “Which of the following statements are true about the distance between -4 and 7?” Review Question 5 states, “A student determines the volume of a cube to be 4.6cm^3. What is the percent error if the correct volume of the crystal is 4.3cm^3?” ##### Indicator {{'1f' | indicatorName}} Content from future grades is identified and related to grade-level work, and materials relate grade-level concepts explicitly to prior knowledge from earlier grades. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially meet expectations that content from future grades is identified and related to grade-level work, and materials relate grade-level concepts explicitly to prior knowledge from earlier grades. The materials do not clearly identify content from future grades, but the materials do relate grade-level concepts explicitly to prior knowledge from earlier grades. Examples where grade-level concepts are explicitly related to prior knowledge from earlier grades: • In Lesson 2.7, Multiplying Rational Numbers, the Teacher’s Edition, lists the Previous Learning Objectives: “Use positive and negative numbers to represent quantities in real-world contexts. (6.NS.C.5)’ to connect to the lesson’s focus standards 7.NS.A.2.a and 7.NS.A.2.c. In the Teacher Notes, it says “Students should know how to multiply fractions and decimals, having learned it in 6th grade. This text will review that a little, but the main focus is on multiplying when the numbers have different signs.” • In Lesson 6.7, Surface Area of Pyramids, the Teacher’s Edition, lists the Previous Learning Objectives: “Represent three-dimensional figures using nets made up of rectangles and triangles. (6.G.A.4)” and “Use nets to find surface area in the context of solving real-world problems. (6.G.A.4)”, the Teacher Notes states: “This lesson expands on the surface and addresses pyramids. The bases of the pyramids students will see will be either triangles or quadrilaterals. Students will use their previous work with nets throughout the lesson.” Relating the focus standard 7.G.B.6 to these 6th grade standards. • In Lesson 7.1, Understanding Likelihood, the Teacher’s Edition, list Previous Learning Objectives: “Recognize and write equivalent fractions. (4.NF.A.1)” and “Compare two fractions with different numerators and different denominators. (4.NF.A.2)”, the Teacher Notes states: “Students will use their understanding of fractions as they work with probabilities for example when they write or compare probabilities.” relating the focus standard 7.SP.C.5 to these 4th grade standards. There is one instance that alludes to content related to future grades, but the future grade-level content is not identified. In Lesson 3.7, Solving Two-Step Equations, the Teacher’s Edition, the Teacher Notes references future learning: “Solving equations is one of the most important lessons for 7th-grade students. It is the foundation of 8th-grade math and Algebra”, but there are no specifics given of how the lesson connects to future learning. ##### Indicator {{'1g' | indicatorName}} In order to foster coherence between grades, materials can be completed within a regular school year with little to no modification. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS can be completed within a regular school year with some modifications. As described below, the lessons and assessments provided within the materials can be completed in 121 days. An average lesson is 90 minutes with additional material available through Related Modalities and practice problems. In addition, lessons include a daily 15 minute review problem session which could easily be modified. Related Modalities content is included within each lesson, but there is no instruction for teachers as to how or when to utilize it. For homework assignments, the materials state, “It is the expectation that the Adaptive Practice will be used as homework. The students must correctly answer ten questions to receive full credit.” The material is viable for one school year with some modification. • Lessons typically follow this format: • Warm up: Ranging between 5-25 minutes • One to four Activities: Ranging between 10-25 minutes each • Review Questions: 15 minutes • Most lessons are 90 minutes long, but lessons range from 60 to 120 minutes. • There are eight chapters. Each chapter ends with an assessment, and the chapters include from six to eleven lessons. • No lessons are marked as supplementary or optional. • The total number of minutes (6235) was divided by an average class period of 55 minutes. This computation resulted in approximately 113 days of instruction. There are eight days for eight chapter assessments, for a total of 121 days. ###### Overview of Gateway 2 ### Rigor & the Mathematical Practices The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for rigor and balance and practice-content connections. The materials help students develop procedural skills, fluency, and application. The materials also make meaningful connections between the Standards for Mathematical Content and the Standards for Mathematical Practice (MPs). ##### Gateway 2 Meets Expectations #### Criterion 2.1: Rigor and Balance Materials reflect the balances in the Standards and help students meet the Standards’ rigorous expectations, by giving appropriate attention to: developing students’ conceptual understanding; procedural skill and fluency; and engaging applications. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for rigor. The materials develop conceptual understanding of key mathematical concepts, give attention throughout the year to procedural skill and fluency, and do not always treat the three aspects of rigor together or separately. The materials partially meet expectations for spending sufficient time working with engaging applications of mathematics, ##### Indicator {{'2a' | indicatorName}} Materials develop conceptual understanding of key mathematical concepts, especially where called for in specific content standards or cluster headings. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for developing conceptual understanding of key mathematical concepts, especially where called for in specific standards or cluster headings. The materials include problems and questions that develop conceptual understanding throughout the grade level. Chapter 1 has multiple opportunities for students to work independently to develop conceptual understanding of analyzing proportional relationships and using them to solve real-world and mathematical problems (7.RP.A) through the use of Interactives. Examples include: • In Lesson 1.6, Activity 1, students use the Interactive to develop understanding of proportional relationships by manipulating numbers to see a total cost based on the amount being bought. The student directions state, “A website offers music downloads for$0.79 per song. Use the slider to see how the cost changes as you increase the number of songs you buy. Use the record button to mark different price points on the table below, then use the data given.” (7.RP.2) • In Lesson 1.8, Warm-up, students manipulate the sliders in the Interactive to solve proportional relationships involving percents by finding out the discount on the amount being spent. The student directions state, “Use the Interactive and slide the tape diagram to adjust for each fraction. This will help you determine the discount. Then, subtract the discount from the original price to get the sale price.” (7.RP.3) Chapter 2 has multiple opportunities for students to work independently to build conceptual understanding of applying and extending previous understandings of operations with fractions (7.NS.A) through the use of interactives. Examples include: • In Lesson 2.3, Activity 3, students develop a conceptual understanding of the distance between numbers by manipulating the sliders on the Interactive activity and answering questions such as, “Which equation models the situation in the problem? A. 11.15 + (-.67) + (-2.25) B. 11.15 + .67 + 2.25 C. 11.15 + .67 + (-2.25) D. 11.15 + (-.67) + 2.25”. (7.NS.1) • In Lesson 2.7, students multiply rational numbers. In Activity 1, the context is owing friends money, and students answer, “Annie owes $6 to 3 friends. How much money does she owe? Remember owing money means you have a negative amount.” In Activity 2, the context is rewinding to the beginning of a TV show. Both of these contexts develop an understanding of multiplying signed rational numbers. (7.NS.2a) Practice questions at the end of the lesson in the student materials include problem 1, (-9) × (+8), and problem 2, (-5) ×(3), and practice questions from the teacher materials include problem 1, (2)(-8)(-3), and problem 4, 4 ×(-50). • In Lesson 2.10, Activity 2, students convert fractions to decimals in the Interactive to develop understanding of multiplying and dividing rational numbers. The student directions state, “Use the Interactive to match the fractions and decimals in the table. Then, select either T for terminating decimals or R for repeating decimals in the last column.” (7.NS.2) Chapter 3 has multiple opportunities for students to work independently to build conceptual understanding of using properties of operations to generate equivalent expressions and solving real-life and mathematical problems using numerical and algebraic expressions and equations (7.EE) through the use of Interactives. Examples include: • In Lesson 3.3, Activity 2, students manipulate the Interactive to sort expressions that are equivalent to the given expression, which develops their understanding of equivalent expressions. The Teacher Notes describe how the students will be independently working by stating, “For this Interactive, students practice matching equivalent expressions to the expression given at the top. Students can click and drag the expressions on the right into the yes or no column.” (7.EE.2) • In Lesson 3.7, Activity 1, students develop the conceptual understanding of solving multi-step problems with the Interactive by balancing the equations to solve for x. The student directions state, “The Interactive will tell you if it is not balanced and when the equation is solved correctly. Click on the buttons at the top of the Interactive to add and subtract ones and x's. At the end, division buttons will appear, so that you can isolate x”. (7.EE.3) ##### Indicator {{'2b' | indicatorName}} Materials give attention throughout the year to individual standards that set an expectation for procedural skill and fluency. The instructional materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for attending to those standards that set an expectation of procedural skill and fluency. The instructional materials develop procedural skill and fluency and provide opportunities for students to independently demonstrate procedural skill and fluency throughout the grade level, especially where called for by the standards (7.NS.1,2; 7.EE.1,4a). In Chapter 2, the materials develop and students independently demonstrate procedural skill in adding and subtracting (7.NS.1) and multiplying and dividing (7.NS.2) rational numbers. Examples include: • In Lesson 2.3, Activity 2 Interactive, students demonstrate procedural skill in adding rational numbers written as decimals. The directions state, “You may remember adding decimals and fractions from last year. Adding decimals is not that much different than adding whole numbers, just make sure you line up the decimal point. As with integers, whichever rational number has the greater absolute value, the answer will have that sign. You may use the Interactive below to brush up on adding decimals.” (7.NS.1) • In Lesson 2.4, the Warm-Up: Subtracting Integers states, “Subtraction is taking away a value from another. Adding -4 would mean moving 4 units to the left. With subtraction it is the opposite. Subtracting -4 would mean moving 4 units to the right. Therefore subtraction can also be defined as adding the opposite. 2 - (-4) + 2 + 4. When doing subtraction problems, change the problem to adding the opposite before starting.” Students complete practice problems, for example, Activity 1: Diving Depths, Inline Question 2 states, “If -5 - 12 models Fatima’s diving depth, what is another way to write this problem?” (7.NS.1) • In Lesson 2.7, students multiply rational numbers. In Activity 1: Annie’s Debt and Activity 2: TV Show Skip Back, students see the results of multiplying numbers with different signs. In Activity 3: Are you -8?, students determine which expressions are equal to -8. For example, Inline Question 1 states, “How would you multiply -\frac{2}{3}×2\frac{3}{4}?” The practice questions at the end of the lesson, such as “$$(-5) ×(3)$$ ,” give independent practice on multiplying integers. In Lesson 2.9, Review Questions, students demonstrate procedural skill in multiplying rational numbers, and some examples include, “6. Multiply the following rational numbers. \frac{1}{11} ×\frac{22}{21} × \frac{7}{10}” and “9. Multiply: \frac{1}{3} ×\frac{4}{12} × \frac{2}{9}.” (7.NS.2) In Chapter 3, the materials develop and students independently demonstrate procedural skill in applying properties of operations as strategies to add, subtract, factor, and expand linear expressions with rational coefficients (7.EE.1) and writing equations of the form px + q = r and p(x + q) = r to solve word problems (7.EE.4a). Examples include: • In Lesson 3.2, Review Questions, students demonstrate procedural skill in applying properties of operations to expressions with multiple examples. Some examples include, “1. Simplify the expression using the distributive property and combining like terms until there are two terms. -5(6t-8)-6(t+3)” and “4. Use the distributive property to write an equivalent expression. (-x+4).” (7.EE.1) • In Lesson 3.3, Activity 2: Are You Equivalent?, students develop procedural skill in applying properties of operations to determine equivalent expressions. The materials state, “Analyze the expression 4(x-3) - 2(5x+6)+10. In the box, there are several other expressions that may or may not be equivalent to it. Sort them depending on if they are equivalent or not to 4(x-3) - 2(5x+6) +10.” Also, students develop skill in the practice questions at the end of the lesson, for example, “8. Simplify the expression \frac{4x}{2} - 2(x+13)-5^2.” (7.EE.1) • In Lesson 3.7, students independently demonstrate procedural skill in solving two-step equations in the Review Questions, for example, “10. 13 - 8x = -3.” (7.EE.4a) ##### Indicator {{'2c' | indicatorName}} Materials are designed so that teachers and students spend sufficient time working with engaging applications of the mathematics. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially meet expectations for being designed so that teachers and students spend sufficient time working with engaging applications of mathematics. The materials include multiple opportunities for students to independently engage in routine application throughout the grade level, but the materials include limited opportunities for all students to engage, collectively or independently, in non-routine application problems. Examples of students engaging in routine application of grade-level skills and knowledge, within instruction and independently, include: • In Lesson 2.4, Subtracting Integers and Other Rational Numbers, students solve routine money problems. For example, Activity 2 states: “Gina has a savings account with$23.64 in it. She makes a withdrawal of $15.67 and then a deposit of$6.78. How much money is in her account?” (7.NS.3) • In Lesson 3.5, Solving Multi-Step Problems with Rational Numbers, students solve routine multi-step problems posed with positive and negative rational numbers. Activity 2: “A new skatepark, Sk8er L8er, is putting in two kickers, with a grind box in the middle, like the picture. To make the feature ready to skate on, the skatepark must also put a layer of Skatelite over the surface to make it smooth and perfect for tricks. The skatepark needs to figure out how many linear feet of Skatelite to buy for this feature. The ramp on each kicker is x feet long, and the grind box is \frac{3}{4}x feet long. The height is \frac{1}{2}x feet.” (7.EE.3) • In Lesson 4.2, Problem Solving with Percents, students use proportional relationships to solve multi-step percent problems . Activity 2 Interactive states, “Jake is renting an apartment for $1,800 a month, and his monthly income is$5,625. What percent of Jake’s monthly income is his rent? Begin by using a tape diagram to estimate the percent of Jake’s monthly income that his rent is.” (7.RP.3) • In Lesson 6.3, Solving Problems involving Circles, students use the formula for a circle to solve problems: Activity 3, Inline Question 3 states, “Sherry wants to put some decorative tile around the pool (circular pool with a diameter of 20 feet). If each tile is 6 inches long, how would she determine how many tiles she needs?”  (7.G.4) The materials provide limited opportunities for students to independently engage with non-routine application throughout the grade level. An example where a student would engage in a non-routine application is shown below. • In Lesson 7.11 Using Simulations to Estimate Probabilities of Compound Events,  Activity 3: Make a Simulation states,  "In the previous lesson you chose an event and created a simulation to model its probability. Appropriate tools were chosen to model the probability of the event. Either extend your scenario to model a compound event or choose a new scenario. Aim for a minimum of 3 simple events. Consider whether the events are dependent or independent. State the scenario, the probabilities, the research you did, and the simulation tools you chose." (7.SP.7) ##### Indicator {{'2d' | indicatorName}} The three aspects of rigor are not always treated together and are not always treated separately. There is a balance of the three aspects of rigor within the grade. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations in that the three aspects of rigor are not always treated together and are not always treated separately. There is a balance of the three aspects of rigor within the grade. Examples include: • In Lesson 3.1, Activity 2, students develop conceptual understanding of combining like terms. Students sort different parts of expressions in the Interactive Activity. Teacher directions state, “This Interactive is a visual example of combining like terms, given expressions with variables. The instructions mention the blue box on the graph represents 7x. Students will also see four yellow boxes each with their own values. Students can click and drag the red points at the corner of each of the boxes to move the boxes around. Students can add to the blue box by stacking the yellow boxes on top of the blue box ... Students can also visualize subtraction by placing the yellow boxes in the blue box.” (7.EE.2) • In Lesson 2.4, Interactive 3, students develop procedural skill by practicing subtraction with decimals. For example, Inline question 1 states, “Calculate: -56.902 - 12.45 - (-13.58) - (-27.9).  a) -41.567 b) -16.945 c) 33.124 d) -27.872.”  (7.NS.1) • In Lesson 1.6, Review Questions, students represent and solve proportional relationships presented through different real-world scenarios. For example, Question 5 states, “The amount of money Sebastian spends on shoes can be represented by the equation y = 50x , where x is the number of pairs of shoes he owns, and y is the total cost. How many pairs of shoes does Sebastian own if he's spent $650?” (7.RP.2) Multiple aspects of rigor are engaged simultaneously to develop students’ mathematical understanding of a single topic/unit of study throughout the materials. Examples include: • In Lesson 1.9, Activity 1, students develop understanding of the effects of scale factors in geometric shapes. The materials state, “Take Brainy’s photo and enlarge it and shrink it. See what sort of conclusions you can make about scale drawings. Determine what the scale factor would be if he started with an 8” by 8” photo and made a duplicate of 6” by 6”.” Later in the lesson, students build procedural skill in finding scale factors in the Review Questions. Review Question 4 states, “A map has a scale of 1 inch = 3 feet. What is the scale factor of the map?” • In Lesson 4.5, Activity 1: Mark Ups Interactive, students develop a conceptual understanding of using equations for percent problems. The materials state, “Use the Interactive below to explore how markup rates affect the sale price of a product. In this Interactive, students will get to experiment with markups and item prices, and see how that will affect the resulting purchase price.” Inline Question 3 states, “Change the markup rate to 160%. At this rate, what will you multiply each purchase price by to get the selling price?” In the Review Questions at the end of the lesson, students apply their knowledge of percents and equations to solving real-world problems. Review Question 2 states, “The marked price of a sweater at the clothing store was$24. During a sale a discount of 25% was given. A further 15% discount was given to the customers who have the store’s credit card. How much would a member customer need to pay for the sweater during the sale if the customer paid with the store's credit card? Round your answer to the nearest cent.” • In Lesson 6.3, students develop procedural skill in finding the circumference and area of circles. Activity 2 states, “Use 3.14 for \pi to determine the area and circumference of the circles in the interactive.” Inline Question 1 states, “The area of a circle is 81\pi. What are the steps to find the circumference?” In Activity 4: Room for pi?, students apply their understanding of circles to real-world situations. For example, Inline Question 3 states, “Sherry wants to put some decorative tile around the pool. If each tile is 6 inches long, how would she determine how many tiles she needs?” #### Criterion 2.2: Math Practices Materials meaningfully connect the Standards for Mathematical Content and Standards for Mathematical Practice (MPs). The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for practice-content connections. The materials meaningfully connect the Standards for Mathematical Content and the Standards for Mathematical Practice (MPs). ##### Indicator {{'2e' | indicatorName}} Materials support the intentional development of MP1: Make sense of problems and persevere in solving them; and MP2: Reason abstractly and quantitatively, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for supporting the intentional development of MP1: Make sense of problems and persevere in solving them; and MP2: Reason abstractly and quantitatively, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials intentionally identify and develop MP1 in connection with grade-level content by providing opportunities for the students to make sense of problems and persevere in solving them. Examples include: • In Lesson 4.4, Finding the Whole Given the Percent, MP1 is intentionally developed as they use an Interactive. During Activity 2: Increase and Decrease, “students are asked to find multiple methods for solving for the whole in an increase/decrease problem” as they use proportional relationships to solve multi-step percent problems. Students are asked to prove their method is equivalent using algebraic skills they have learned. (7.RP.3) • In Lesson 6.11, Volume of Composite 3D Solids, students intentionally develop MP 1 as they “look for an entry point (for finding volume of composite solids) and consider similar problems which may guide them.” In Activity 1: Step Stool Storage, the Teacher Notes explains, “In this Interactive, students will adjust the dimensions of a step stool to figure out its volume and how much storage it has. Students will start with a small step tool with three red points and dimensions, 8\times6\times5. Once students resize the stool to the correct dimensions, a slider will appear at the top of the screen so students can see how the step stool can be broken up into smaller shapes.” (7.G.6) The materials intentionally identify and develop MP2 in connection with grade-level content by providing opportunities for the students to reason abstractly and quantitatively. Examples include: • In Lesson 3.4, Equivalent Expressions Within a Context, students intentionally develop MP 2 throughout the lesson using the Interactives and Inline Questions. Activity 3: Mystery Age, states, “Your math teacher, Miss Nomer, gives you an extra credit problem to figure out her age. Half of her age three years ago is equal to one-third of her age nine years from now. How old is she currently? All the pieces to figure out Miss Nomer’s age are in the box below. Your job is to make two equivalent expressions from the clues above. Set them equal to each other so you can determine her age and get extra credit.” (7.EE.4) • In Lesson 7.3, Discerning Between Equally Likely and Non-Equally Likely Outcomes, students intentionally develop MP2. For example in Activity 1: Rolling the Dice, students are given this situation: “Sai and his friend Kai are rolling dice to see who gets the higher number. They each roll a six-sided die and whoever gets the higher number wins the round. If they roll the same number, then it is a tie and no one wins the round.” Students then are posed the following Discussion Questions, “Look at your results, is this what you expected? Are both boys equally likely to win? Why or why not?” as they use data to determine whether the simulation was fair or unfair. (7.SP.7) ##### Indicator {{'2f' | indicatorName}} Materials support the intentional development of MP3: Construct viable arguments and critique the reasoning of others, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for supporting the intentional development of MP3: Construct viable arguments and critique the reasoning of others, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials intentionally identify and develop MP3 in connection with grade-level content by providing opportunities for the students to construct viable arguments and critique the reasoning of others. Examples include: • In Lesson 2.9, Multi-Step Multiplication & Division Problems involving Rational Numbers, the materials help students to intentionally develop MP3 to its full intent. In the Discussion Question following Activity 1: Switch the Order, students critique the reasoning of a student and form an argument. “Jack claims that the expressions -8 ÷ 2 ÷(-5) ÷ 2 and -8 ⋅ \frac{1}{2} ⋅ (-\frac{1}{5}) ⋅ \frac{1}{2} are the same, do you agree or disagree? Provide evidence for your argument? Can you apply the Commutative or Associative Properties to one and not the other?” (7.NS.2 & 7.NS.3) • In Lesson 3.8, Using Two-Step Equations to Solve Problems, Activity 2, the Discussion Question asks, “Looking back at the answers for some of the inline questions, are any of the answers not possible? Could you automatically eliminate any?” The Teacher Notes state, “ The answer is, yes, some of the answers could definitely be eliminated. Ultimately, all students will have to take a standardized test at the end of the year, and it is always a good test-taking technique to learn how to eliminate answers that are not possible. In the case of 2, 3, and 5, seconds cannot be negative, so those "distractors" can automatically be eliminated, thus making the selection choice smaller and a greater likelihood of selecting the correct answer. For similar reasons, you could discuss with the class why some of the equations are incorrect in #1 and #5. For example, 160t cannot be positive in the equation because Alex is falling, meaning that 160 needs to be negative.” (7.EE.4) • In Lesson 4.4, Finding the Whole Given a Percent, Standards for Mathematical Practice: “MP3: In Activity 1, the students are asked to develop an argument around which approach they prefer to help understand a pie chart. The students are given the opportunity to analyze the arguments of their classmates.“ Activity 1: A Slice of Pie, Discussion Question: “When solving the problem above, Donna found it easier to start with smaller percentages, and Ellie found it easier to start with multiples of 10. Who do you side with and why?  Use evidence or an example to support your answer.” The Teacher Notes state, “Answers may vary. Students may have leaned toward small percentages or multiples of 10 for easy calculation. However, finding smaller percentages may be a more effective strategy if the smaller percentages are factors of larger percentages. Allow students to share their strategies.” (7.RP.3) • In Lesson 8.2, Visually Comparing Two Data Distributions, Activity 3, the Teacher Notes encourage the teacher to have students discuss mean and median in relation to visual data sets. The Discussion Questions asks, “Based on the data taken, which angle do you believe produced the most solar energy?” The Teacher Notes then state, “Allow students to discuss with a classmate and then share with the class. Encourage them to discuss how the mean, median, MAD and IQR influenced their conclusion. After several students have shared their results with the class, allow the groups to put all of their findings together to determine which angle produced the most solar energy.”(7.SP.3) ##### Indicator {{'2g' | indicatorName}} Materials support the intentional development of MP4: Model with mathematics; and MP5: Choose tools strategically, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for supporting the intentional development of MP4: Model with mathematics; and MP5: Choose tools strategically, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials intentionally identify and develop MP4 in connection with grade-level content by providing opportunities for the students to model with mathematics. Examples include: • In Lesson 2.5, Finding the Distance Between Two Numbers, students model the rise and run of slope with mathematics. In Activity 2: Getting around Washington DC, students use an online map of Washington, D.C. set in a grid to investigate distances between various landmarks. The Inline Questions state, “1. How many blocks is the Spy Museum for Union Station?  2. Select the two routes Cheryl could take to get to the Spy Museum from 3rd and C SW.  3. The DMV is located at Half St and K in SW. What would be the coordinates on our map?” The Teacher Notes state, this “Interactive is a real-world application of finding the distance between two points when they are not collinear.” (7.NS.1) • In Lesson 4.9, Solving Problems Involving Taxes, Commissions, and Fees , Activity 4: Big Tipper Continued, students are encouraged to use a variety of strategies to determine a 20% tip, such as multiplying by \frac{1}{5}, using 10% times 2, etc. Then they model with mathematics.  Students are given a sample meal check and are directed as follows: “Put a tip in the input box. Aim for a tip that is around 20%. Click the ‘Check’ button to check your answer. Click the ‘New Amount’ button to try a new bill.” (7.RP.3) • In Lesson 7.9, Solving Problems Involving Compound Probability, Activity 1: Deal or No Deal, the students model with mathematics the probability of getting a certain prize in a ‘Deal or No Deal’ type game. It states, “You start with four cases, and in each 'round,' choose one case to remove from the table. The goal is to try to leave the case with $10,000 for last. Play the "Deal or No Deal" game in the Interactive below several times. Review the tree diagram and the probability of each final outcome at the end of the game. Once you have a feel for the probabilities of each prize, answer the Inline Questions that follow. 1) How can you find the probability of winning$10,000? 2) Look at the tree diagram. Which of the following statements is true? 3) Look at the completed tree diagram. Select the true statements.” (7.SP.8) • In Lesson 8.8, Understanding Sampling Variability, Activity 3: Collecting Data, students collect data and analyze data on reaction time, using the Interactive to model with mathematics. The directions on the Interactive are as follows: “You will be recording the time it takes you to catch a bug on the screen 5 times. Then do so with three other classmates and observe the statistics for your results. Press the buttons to release the bug. Tap the bug to catch it, and stop the timer. Observe the results for each student in the given table.”  Students are directed to, “collect data from yourself and from 4 other classmates. Each classmate should provide 5 samples. Display each sample in a spreadsheet. Find the mean and MAD for each person’s data.” (7.SP.2) The materials intentionally identify and develop MP5 in connection with grade-level content by providing opportunities for the students to choose tools strategically. Examples include: • In Lesson 5.2, Triangle Construction,  students do the following:  “In Activity 1, students choose their own tools to create a conjecture about the restrictions on the side lengths of a triangle. The students can come up with a conjecture and provide examples and counterexamples to support their arguments. The students share the experience of how their choice of tool helped or hurt their conjecture.” Activity 1: Three Sides; Discussion Question states: “Can any three lengths create a triangle? Create a conjecture about the side lengths of a triangle.” Then the Teacher Notes state the following:  “For this question, the students will need physical tools to practice constructing triangles based on a chosen number of sides. Allow the students to choose their own tools to practice constructing triangles. The students can choose rulers, graph paper, string, etc. The students should choose three side lengths and then attempt to construct a triangle using those dimensions. If the sum of any two sides is not greater than the third side, the triangle sides will not connect. The students should work in small groups to maximize the data collection and to compare their tool choices with their peers. At the end of the activity, the students should share their conjectures and discuss how their choice of tools assists or hinders their efforts.” (7.G.2) • In Lesson 6.2, Area and Circumference of Circles: “ In Activity1, the students explore the challenges of using various tools to measure the circumference of a circle.” Activity 1 states, Parts of a Circle; Discussion Question: “What tool do you think would be best for measuring the circumference of a circle? What challenges would you have measuring the circumference with a circle?” After students discuss the questions with their classmates, the Teacher Notes instruct the teacher to print the circle below and allow the students to use real tools like rulers and measuring tape to make the activity more tangible.  (7.G.4) • In Lesson 7.10, Using Simulations to Estimate Probabilities of Simple Events states, “ In each Activity, the students must choose the appropriate tools to simulate the scenario presented in the Activity.” For example, in Activity 2: Batting Average; Discussion Questions state: “1. What other tools could you have used to simulate a 0.333 batting average, a 1-in-3 chance of getting a hit.  2. Would the tools you chose for the previous question still apply if the batting average was 0.3000 or 0.320? Which tools would work in these situations.” Activity 3: Make a Simulation states, “Choose an event and create a simulation to model its probability. Choose the appropriate tools to model the probability of the event. It may require research to find the probability of the possible outcomes. State the scenario, the probabilities, the research you did, and the simulation tools you chose.” (7.SP.6 & 7.SP.7) ##### Indicator {{'2h' | indicatorName}} Materials attend to the intentional development of MP6: Attend to precision; and attend to the specialized language of mathematics for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials reviews for CK-12 Interactive Middle School Math 7 meet expectations for supporting the intentional development of MP6: Attend to precision; and attend to the specialized language of mathematics, for students,  in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials intentionally develop MP6 through providing instruction on communicating mathematical thinking using words, diagrams, and symbols. Examples include: • In Lesson 3.2, Rewriting Expressions Using the Distributive Property, Activity 1, students are asked to expand expressions such as \frac{1}{2}(4a-5). The Teacher Notes state, “For questions #4 and #5, discuss what it means to be a factor. Students commonly get the words ‘factor’ and ‘multiple’ confused. It might be hard for students to see that a number with a + sign is a factor. If that’s the case, show them this example; if a = 2, b = 3, and c = 5, then 2 and 8 b+care factors of their product, 16. Notice that 3 nor 5 are factors, but their sum is.” (7.EE.1) • In Lesson 5.1, Special Angle Pairs, students learn about supplementary, complementary, adjacent, and vertical angles.  The Teacher Notes at the beginning of the lesson states, “Start by reviewing some important terminology: lines, line segments, types of angles, etc. Some of these terms may be new to students, like complementary and supplementary. If students are having difficulty with all the new vocabulary, you can provide them with a vocabulary toolkit or encourage them to make flash cards.” (7.G.5) • In Lesson 6.3, Solving Problems Involving Circles, Activity 2, students attend to precision as they are asked to, “Use 3.14 for \pi to determine the area and circumference of the circles in the Interactive. Repeat several times until you are comfortable with the two formulas. Remember the 3.14 is an estimation of \pi, but it does enable you to get a numerical answer, and not one in terms of \pi.“ (7.G.4) The materials use precise and accurate terminology and definitions when describing mathematics, and the materials also support students in using the terminology and definitions. There is no separate glossary in these materials, but definitions are found within the units in which the terms are used. The vocabulary words are in bold print. Examples include: • In Lesson 3.3, Identifying Equivalent Expressions,  Introduction, students read the definition of equivalent expressions. The definition is as follows: “Two expressions are equivalent if they can be simplified to the same third expression or if one of the expressions can be written like the other. In addition, you can also determine if two expressions are equivalent when values are substituted in for the variable and both arrive at the same answer.” In Activity 1 Interactive, students see an example of identifying equivalent expressions to help them understand the proper use. This is introduced to the students as, “In this Interactive, there is a column of expressions and a box with other expressions. Your job is to drag the expressions in the box to its equivalent expression in the column. The first one is done for you.” (7.EE.1) • In Lesson 5.1, Special Angle Pairs, Activity 1, students read and apply the exact definitions of terms relating to angles: “A line is composed of infinitely many points, but you only need two points to define a line. Three points are used to define an angle, where the middle point is always the vertex.”  Students are supported in using the terms to answer Inline Questions where they must identify angle terms from a diagram: “1. (Highlight) Based on the data in the image, select the points collinear with point A. 2. (Drag and Drop) Sort the terms below into the correct categories using the image for reference. Remember multiple items may use the same points. For example, points C and D could describe both a line segment and a ray. 3. Angles are labeled in the form ∠ABC, where the middle letter always describes the vertex. The other two letters may be in either order. Select all the correctly labeled angles below.”  (7.G.5) • In Lesson 8.5, Introduction to Sampling, Activity 1 states, “Sampling is the practice of using data obtained from a group to represent a population. A population is a group of objects with a common characteristic. The group selected from the population is called a sample. By studying small groups of a larger population, you can identify trends that might apply to the entire population. Throughout this chapter, you will explore what makes a good sample and how it can be used to make estimates about large populations.”(7.SP.1) ##### Indicator {{'2i' | indicatorName}} Materials support the intentional development of MP7: Look for and make use of structure; and MP8: Look for and express regularity in repeated reasoning, for students, in connection to the grade-level content standards, as expected by the mathematical practice standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for supporting the intentional development of MP7: Look for and make use of structure; and MP8: Look for and express regularity in repeated reasoning, for students, in connection to grade-level content standards, as expected by the mathematical practice standards. The materials intentionally identify and develop MP7 in connection with grade-level content by providing opportunities for the students to look for and make use of structure. Examples include: • In Lesson 2.4, Subtracting Integers and Other Rational Numbers, Activity 1, students develop MP7 as they use an Interactive to “explore the link between adding a negative and subtracting a positive number.”  The Teacher Notes state, “This interactive shows an application of how negative values are used; in this case, it is how deep a diver dives. Students will see a body of water with a diver, boat and some fish. There are also two number lines; the horizontal line ranges from -20 to 0 and the vertical line ranges 0 feet to 20 feet. The horizontal line has a red point that students can click and drag to make the diver dive. While the student moves the point in the negative direction the father down the diver will go. The arrow on the vertical line will travel down with the diver showing the student how many feet down the diver is.”  Discussion Question: “Why is subtracting a negative integer the same as adding a positive integer?” (7.NS.1) • In Lesson 2.11, Using Arithmetic Methods to Solve Multi-Step Problems, students intentionally develop MP7 as they “explore how parentheses placement can affect the simplification of an expression.” In Activity 1, students make use of structure as they investigate an equation. “See the equation with blue parentheses. Below is the expression from the left side of the equation.  Depending on which operation is performed first, the value of the expression changes.  For example, if you add 3 + 4 before performing any other operations in the first equation, then the expression equals 37. 1) Move the red points to change which operation is performed first. 2) Move the red point so the second expression equals 18. 3) Move the red point so the third expression equals 1.” (7.NS.2a & 7.NS.3) • In Lesson 3.9, Writing Two-Step Inequalities, Activity 1, MP7 is intentionally developed as students compare the structure of inequalities to the written expression that matches them.  In this interactive, you will practice matching inequalities to phrases.  Try to notice any keywords that help you match the phrase to the inequality.” In the Inline Questions, students continue to use the structure of inequalities and their expression in words. For example, Question 2: “How could the inequality -3>-2+\frac{1}{2}z be expressed differently than it was in the interactive? a) -3 is less than -2 and half a number. b) -3 is greater than -2 and half a number.  c)  -3 is less than half a number and 2 d) -3 is greater than -2 less than half a number. (7.EE.4) • In Lesson 6.2, Area and Circumference of Circles, Activity 2, Students use an interactive to determine the revolutions of a bowling ball as it rolls down an alley. Students apply their measurements as they use the ratio of circumference to diameter to help derive the circumference of a circle” guided by  Inline Questions: “1). Use the interactive. Approximately how many inches does the ball travel in on revolution?  2) How can we mathematically find the number of revolutions it takes the ball to reach the pins, 3. The number \pi is defined as the ratio of the circumference, C, to the diameter, d.  For the ball, this is \frac{27}{8.5}\approx 3.14 and 4) How can we rewrite the formula  as a formula to find the circumference? r represents the radius in the formulas.” (7.G.4) The materials intentionally identify and develop MP8 by providing opportunities for the students to look for and express regularity in repeated reasoning. Examples include: • In Lesson 1.4, Identifying Proportional Relationships, students intentionally develop MP8 as they “use repeated reasoning to develop a conjecture about the properties of graphs of proportional relationships.” Activity 1’s Discussion question asks, “What do you notice about this graph and the graph from the Driving Away interactive?” (7.RP.2) • In Lesson 6.10, Volume of Triangular Prisms, in the introduction, students compute the area of a triangle using a one square unit grid and then use that result to compute the volume of a one unit high triangular prism.  The Discussion Questions 3 and 4 lead them to use repeated reasoning to derive the formula for the volume of a triangular prism of any dimensions: 3) How many cubes would there be in 2 of these figures? 4) How many cubes would there be in n of these figures?” (7.G.6) ### Usability The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially meet expectations for Usability. The materials partially meet expectations for Teacher Supports (Criterion 1), meet expectations for Assessment (Criterion 2), and do not meet expectations for Student Supports (Criterion 3). ##### Gateway 3 Partially Meets Expectations #### Criterion 3.1: Teacher Supports The program includes opportunities for teachers to effectively plan and utilize materials with integrity and to further develop their own understanding of the content. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially meet expectations for Teacher Supports. The materials provide: teacher guidance with useful annotations and suggestions for enacting the student and ancillary materials; explanations of the instructional approaches of the program and identification of the research-based strategies; and a comprehensive list of supplies needed to support instructional activities. The materials contain adult-level explanations and examples of concepts beyond the current grade so that teachers can improve their own knowledge of the subject, but do not contain adult-level explanations and examples of the more complex grade-level concepts. The materials partially include standards and correlation information that explains the role of the standards in the context of the overall series. ##### Indicator {{'3a' | indicatorName}} Materials provide teacher guidance with useful annotations and suggestions for how to enact the student materials and ancillary materials, with specific attention to engaging students in order to guide their mathematical development. The materials reviewed  for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for providing teacher guidance with useful annotations and suggestions for how to enact the student materials and ancillary materials, with specific attention to engaging students in order to guide their mathematical development. Materials provide comprehensive guidance to assist teacher delivery of student materials. The Teacher Edition of the materials contains Teacher Notes throughout to assist the teacher in presenting the student lessons.  Examples include: • Important information about student learning at the beginning of lessons. For example, in Lesson 7.4, Theoretical Probability Models, the Introductory Teacher Notes states, “In this lesson, we shift from experimental probability to theoretical. Students will continue to learn how to calculate theoretical probabilities and use these probabilities to estimate and predict the likelihood of outcomes. Students should continue to use what they have learned about probabilities as they work through the examples to answer questions and make predictions. (7.SP.5 & 7.SP.6) • Answers to all Inline Questions • Instructions for help with the Interactives. For example, in Lesson 1.8, Solving Multi-Step Ratio Problems, Warm-Up, the Teacher Notes for the Interactive states, “This Interactive is intended to help with the discount problem in the text above. It has three sliders: A green slider to adjust price (of the item being discounted), a red slider to change the numerator of the fraction, and a blue slider to change the denominator. As the students change the fraction, an equivalent number of boxes will appear next to the fraction to help the student visualize the percentages. The denominator dictates how many boxes appear, and the numerator dictates how many boxes are colored.  Students can set these values to match the product price and discounts in the text above, and they will see how much was discounted (ex: set the Interactive fraction to \frac{1}{4} and price to $1,200 to find the price of the discount on the refrigerator -$300 off.) (7.RP.3) • Possible answers, further questions, and discussion ideas for the Discussion Questions are in the following examples. In Lesson 6.1, Area and Perimeter of Scale Drawings, Activity 2, the Discussion Question Teacher Notes state, “Students need to convert the scale so that both are in inches before determining the size of the model. Once the scale is all in inches, then also convert the measurements of the actual building into inches. Once everything is all in the same units, then it can be scaled down. As for the scale factor, it is reasonable. It might be a little big to bring a nearly 5-foot tall model to school, but it could work. As an extension, you could have students determine a more “reasonable” scale and then determine the dimensions of their model. Students need to explain why they chose their scale.” (7.G.1) • Specific learning standards and objectives for each lesson • Lesson-specific Teacher Notes • A scope-and-sequence at the end of the Teacher Edition Materials include sufficient and useful annotations and suggestions that are presented within the context of the specific learning objectives. In the Teacher Edition at the beginning of each lesson, there is an overview of the lesson to assist the teacher in planning lessons: • Common Core Standard—the focus and prerequisite standard(s) for each lesson is listed. • Standard for Mathematical Practice—the mathematical practice(s) for each lesson is listed as well as where in the lesson it is developed. • Previous Learning Objectives —a majority of the lessons list these objectives and the standard(s) or grade(s) the objective is connected. • Learning Objectives -goals for each lesson. • Agenda - there is an agenda listed for each lesson with the allotted times for the Warm-Up, the Activities, Review Questions, Related Modalities and Adaptive Practice. • Introductory Teacher Notes—located at the beginning of the lesson after the agenda, these notes describe what the students will be doing in the lesson. Some have helpful hints. • Interactives—Teacher Notes for the Interactive activities give the teacher direction on how the students are to use the Interactive and helpful hints. • Discussion Questions -  Teacher Notes for Discussion Questions  provide possible answers and/or possible questions to ask to further the discussion. • Extension Activities—some of the lessons give extension activity ideas that can enhance the learning. ##### Indicator {{'3b' | indicatorName}} Materials contain adult-level explanations and examples of the more complex grade-level/course-level concepts and concepts beyond the current course so that teachers can improve their own knowledge of the subject. The materials reviewed for CK-12 Middle School Interactive Math 7 for CCSS partially meet expectations for containing adult-level explanations, examples of the more complex grade-level concepts, and concepts beyond the current course so that teachers can improve their own knowledge of the subject. The Teacher Edition does not contain any adult-level explanations and examples of the more complex grade-level concepts so that teachers can improve their own knowledge of the subject.  In the Subjects Menu, Math Flexlets are available for 6th, 7th and 8th Grade Math Essentials. These are shortened versions of some key lessons intended for review.  For example, Interactive 6th Grade Math Essentials states, “This Flexlet is a great resource to prepare for or review Middle School Math 6. It is a collection of only the 'key' lessons in CK-12 Interactive Middle School Math 6. Additional detailed support for concepts introduced here is available in the full CK-12 FlexBook 2.0.” This resource does not offer adult-level explanations and examples of the more complex grade-level concepts since it addresses only key lessons and not more complex concepts. Additionally, Study Guides can be found under the Explore menu, and are intended as a “Quick review with key information for each concept.” The math content covered in the Study Guides is beyond the current course and offers math high school courses Algebra and Geometry. These Study Guides can be used so that teachers can improve their own knowledge of the subject. However, not all Study Guides are connected to High School standards or standards at all. ##### Indicator {{'3c' | indicatorName}} Materials include standards correlation information that explains the role of the standards in the context of the overall series. The materials reviewed for CK-12 Middle School Interactive Math 7 for CCSS partially meet expectations for including standards and correlation information that explains the role of the standards in the context of the overall series. Correlation information is present for the mathematics standards addressed throughout the grade level. However, there are few, if any, explanations of the role of the specific grade-level mathematics in the context of the series and no connection to future learning. Previous learning objectives are listed on most of the lessons. There are limited instances of objectives connecting to previous grade levels, and the remaining previous learning objectives listed are related to grade-level standards. Examples include: • Lesson 2.7, Multiplying Rational Numbers, lists the following as Previous Learning Objectives: Use positive and negative numbers to represent quantities in real-world contexts (6.NS.5). The Introductory Teacher Notes state, “Students should know how to multiply fractions and decimals, having learned it in 6th grade. This text will review this a little, but the main focus is on multiplying when the numbers have different signs. Students should use their knowledge of positive and negative values when working through some of the examples. For instance, when adding, the order does not matter.” • Lesson 3.1, Combining Like Terms, lists the following as Previous Learning Objectives: Identify terms, factors, and coefficients in an express (6.EE.2b) and Apply properties of operations as strategies to add and subtract rational numbers (7.NS.1d). • Lesson 6.1, Area and Perimeter of Scale Drawings, lists the following as Previous Learning Objectives: Solve real-world problems involving area of a special quadrilateral (6.G.1); Solve mathematical problems involving area of a special quadrilateral (6.G.1); Understand for what values of k will a scaled drawing be bigger or smaller than the original (7th); and Compute the actual lengths from a scale drawing (7.G.1). Future learning objectives are seldom present and are usually referred to later in the grade level and not to a concept in future grade levels or courses. For example, in Lesson 6.11, Volume of Composite 3D Solids, the Introductory Teacher Notes state, “This lesson combines the volumes of different types of prisms.  Even though the surface area of pyramids was explored earlier, the volume of one is not covered until 8th grade.” Mathematics standards, practices and learning objectives are listed throughout the grade level at the beginning of each lesson. Examples include: • In Lesson 3.11, Using Two-Step Inequalities to Solve Problems, the Focus Standard is 7.EE4b, and the standards for mathematical practice listed with the lesson is MP4. The Learning Objectives are the following: Use inverse operations to solve two-step inequalities, solve inequalities in the context of real-world problems and interpret the solution of an inequality within the context of the problem. • In Lesson 4.7, Changing Percents, the Focus Standard is 7.RP.3 and the standards for mathematical practice listed with the lesson are MP2 and MP4. The Learning Objective is, “Solve problems involving a percent where quantities change in a real-world context.” • In Lesson 7.3, Discerning Between Equally Likely and Non-Equally Likely Outcomes, the Focus Standards are given as 7.SP.7a and 7.SP.7b, and the standards for mathematical practice listed with the lesson are MP2 and MP6. The Learning Objectives are the following: Understand what an outcome is, understand what an event is, develop a uniform probability model by assigning equal probability to all outcomes, and use this model to determine probabilities of events, understand the difference between a fair and unfair event and develop a non-uniform probability model by observing frequencies in data generated from a chance process. ##### Indicator {{'3d' | indicatorName}} Materials provide strategies for informing all stakeholders, including students, parents, or caregivers about the program and suggestions for how they can help support student progress and achievement. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS do not provide strategies for informing all stakeholders, including students, parents, or caregivers, about the program and suggestions for how they can help support students’ progress and achievement. Although the materials support teachers with planning, instructions, and analysis of student progress, there are no specific resources for parents or caregivers. While students are able to access their reports, there are no strategies provided to assist their progress or achievement. There are no explanations for parents or caretakers on the place to create an account to help support in-class learning or home instruction. ##### Indicator {{'3e' | indicatorName}} Materials provide explanations of the instructional approaches of the program and identification of the research-based strategies. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for providing explanations of the instructional approaches of the program and identification of the research-based strategies. Instructional approaches of the program and identification of the research-based strategies can be found on the homepage, the citations for this research can be found under the “Resources” tab on the homepage. The materials state the following, “The CK-12 Interactive Middle School Math series promotes exploratory learning (Stein 2010). Each lesson contains interactive applets which actively engage students in the learning process and allow them to explore concepts in an open-ended environment (Cocea & Magoulas, 2015; Hoyles, 2018; NCTM, 20115). Inline question sets Socratically guide students to discover connections present in the interactive applets, and a list of Works Cited includes: • Stein, R. G. (2010). Math for Teachers: An Exploratory Approach. Kendall Hunt Publishing Company. • Cocea, M., & Magoulas, G. D. (2015). Participatory learner modeling design: a methodology for iterative learner models development. Information Sciences, 321, 48-70. • Schunk, D. H. (2012). Learning Theories: An Educational Perspective. Pearson. • Hoyles, C. (2018). Transforming the mathematical practices of learners and teachers through digital technology. Research in Mathematics Education. • Hoyles, C., & Lagrange, J. B. (Eds.). (2010). Mathematics education and technology: Rethinking the terrain. New York: Springer. • National Council of Teachers of Mathematics. (2014). Access and equity in mathematics education: A position of the national council of teachers of mathematics. National Council of Teachers of Mathematics. • National Council of Teachers of Mathematics. (2015). Strategic use of technology in teaching and learning mathematics: A position of the national council of teachers of mathematics. National Council of Teachers of Mathematics. • Wolf, D., Lindeman, P., Wolf, T., & Dunnerstick, R. (2011). Integrate Technology with Student Success. Mathematics Teaching in the Middle School, 16(9), 556-560.” ##### Indicator {{'3f' | indicatorName}} Materials provide a comprehensive list of supplies needed to support instructional activities. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for providing a comprehensive list of supplies needed to support instructional activities. The Interactives in the lessons are designed to replace any extra materials. There is a comprehensive list of supplies needed for the optional activities included at the beginning of the Teacher Edition under the Resources tab. The materials listed are provided for the lessons that need extra supplies for optional activities  (Note: the quantity listed is per student unless otherwise noted.) Examples include: • In Lesson 1.7, Interpreting Graphs of Proportional Relationships, Warm-up Activity, Proportional Relationships in Art: Ruler and Tape Measure. • In Lesson 6.2, Area and Circumference of Circles, Activity 1, Parts of a Circle: Ruler Access, Tape Measure Access, and Access to other measuring tools if desired. ##### Indicator {{'3g' | indicatorName}} This is not an assessed indicator in Mathematics. ##### Indicator {{'3h' | indicatorName}} This is not an assessed indicator in Mathematics. #### Criterion 3.2: Assessment The program includes a system of assessments identifying how materials provide tools, guidance, and support for teachers to collect, interpret, and act on data about student progress towards the standards. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for Assessment. The materials include an assessment system that provides multiple opportunities throughout the grade to determine students' learning and sufficient guidance to teachers for interpreting student performance and suggestions for follow-up, and the materials provide assessments that include opportunities for students to demonstrate the full intent of grade-level standards and practices.  The materials partially include assessment information in the materials to indicate which standards are assessed. ##### Indicator {{'3i' | indicatorName}} Assessment information is included in the materials to indicate which standards are assessed. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially meet expectations for having assessment information included in the materials to indicate which standards are assessed. The materials sometimes identify the standards and mathematical practices addressed by formal assessments. Formative assessments, including Inline Questions, Review Questions/Quiz, and Adaptive Practice are located in each lesson, however the materials only identify the standards and practices assessed for some of the formal assessments. In the Teacher Edition, at the beginning of each lesson, standards and mathematical practices are clearly listed, but specific standards and practices are not listed for each question on the Inline Questions, Adaptive Practice and Review Questions/Quizzes. The end of chapter assessments identify the standards for each question, but do not identify the mathematical practices. Examples include, but are not limited to: • In Lesson 3.1, Combining Like Terms,  Activity 2, Inline Question 4, “Simplify the expression: 8x+3x+6x-7 ” Answer choices a) 10x+7, b) 17x+7, c) 17x-7, d) 11x-7 • Chapter 2, Operations with Rational Numbers, Question 2: “(7.NS.A.1.a, 7.NS.A.1.c) The value of A is 5. a. What value, call it B, can you add to 5 to make 0? Write an equation using subtraction to show this. How could you represent this using addition?” • Chapter 7, Probability, Question 4: “(7.SP.C.8) Alex, Bianca, and Cameron are racing canoes. Assume that each racer has an equal chance of winning. a. (7.SP.C.8.b) Create a tree diagram to represent the number of ways each racer can finish.” ##### Indicator {{'3j' | indicatorName}} Assessment system provides multiple opportunities throughout the grade, course, and/or series to determine students' learning and sufficient guidance to teachers for interpreting student performance and suggestions for follow-up. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for including an assessment system that provides multiple opportunities throughout the grade to determine students' learning and sufficient guidance to teachers for interpreting student performance and suggestions for follow-up. The assessment system provides multiple opportunities to determine students' learning and sufficient guidance to teachers for interpreting student performance and most of the assessments provide sufficient suggestions for following-up with students. Examples include: • Every lesson has Adaptive Practice Questions which generate a report with the number correct, difficulty of the questions, time spent and mastery level. • Answer keys are provided for all Inline Questions, Discussion Questions, and End of Chapter assessments. • Each of the End of Chapter Assessments contains a rubric to assist the teacher in scoring student work. Each problem is given a 1-5 score and is correlated with the CCSS. Rubrics are provided for End of Chapter Assessments only. Scoring rubrics provide information on student performance but do not include suggestions for the teacher to follow up. • Statistics are given through reports to the teacher on each assessment component students take. The Skill Meter gauges student understanding and skill based question-by-question and is color-coded so teachers can quickly ascertain student understanding: • Beginning - new to concept (red) • Exploring - starting to understand (orange) • Developing - demonstrating familiarity (yellow) • Proficient - understands core concept (light green) • Mastery - deep, demonstrated understanding (dark green) • The Class Insights function uses the Skill Meter to give information on individual students and the entire class, by placing students on a quadrant analysis graph based on their skill level and engagement. The Class Insights function also has the Teacher Assistant which, “uncovers your students’ learning gaps and misconceptions, giving you (the Teacher) personalized insights on where you (the Teacher) can intervene effectively.” The Teacher Assistant provides suggestions for following-up with students through the “Insights and Recommendations” section. Examples of suggestions include: • Recommending specific “PLIX” activities to help students with low skill levels improve their skill levels. • Noting which students are doing exceptionally well on the current concept, and suggesting new concepts to keep those students challenged. • Information about the top question(s) students answered incorrectly, with the recommendations for students to review the following question(s) and the related paragraphs. • Information on which students are not reaching the goal of 10 correct answers on the Adaptive Practices, and recommending to remind students to complete that goal. ##### Indicator {{'3k' | indicatorName}} Assessments include opportunities for students to demonstrate the full intent of grade-level/course-level standards and practices across the series. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS meet expectations for providing assessments that include opportunities for students to demonstrate the full intent of grade-level standards and practices across the series. The assessments regularly provide opportunities for students to demonstrate the full intent of grade-level standards and practices through a wide variety of assessment types, such as multiple choice, drag and drop, matching, short answer, true/false, computational response, and discussion response. Students use different types of modalities to demonstrate their understanding in assessment, including short answer explanations and multi-layered questions. The Inline Questions and Review/Quiz Questions are connected to standards and practices. The End of Chapter Assessments have the content standards identified on the answer keys. ##### Indicator {{'3l' | indicatorName}} Assessments offer accommodations that allow students to demonstrate their knowledge and skills without changing the content of the assessment. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially provide assessments which offer accommodations that allow students to demonstrate their knowledge and skills without changing the content of the assessment. The materials have accommodations that are built into every Review Questions/Quiz. Teachers can set the number of attempts allowed, adjust the time limit, allow students to pause and resume, show hints, show solutions, or shuffle the questions. Teachers are able to alter these quizzes by choosing from item sets or adding their own questions. As a result, these items have the potential to alter grade-level expectations due to the fact that these are teacher-created itemsA Word version of the End of Chapter Assessments is included, making these assessments customizable. Both assessments are only available in English. #### Criterion 3.3: Student Supports The program includes materials designed for each child’s regular and active participation in grade-level/grade-band/series content. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS do not meet expectations for Student Supports. The materials provide manipulatives, both virtual and physical, that are accurate representations of the mathematical objects they represent and, when appropriate, are connected to written methods. The materials partially provide strategies and supports for students who read, write, and/or speak in a language other than English to regularly participate in learning grade-level mathematics, and partially provide extensions and/or opportunities for students to engage with grade-level mathematics at higher levels of complexity. The materials do not provide strategies and supports for students in special populations to support their regular and active participation in learning grade-level mathematics. ##### Indicator {{'3m' | indicatorName}} Materials provide strategies and supports for students in special populations to support their regular and active participation in learning grade-level/series mathematics. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS do not meet expectations for providing strategies and supports for students in special populations to support their regular and active participation in learning grade-level mathematics. The materials have some general strategies, but they do not explicitly provide specific strategies and supports for differentiating instruction to meet the needs of students in special populations or support their regular and active participation in the learning of grade-level mathematics. ##### Indicator {{'3n' | indicatorName}} Materials provide extensions and/or opportunities for students to engage with grade-level/course-level mathematics at higher levels of complexity. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS partially meet expectations for providing extensions and/or opportunities for students to engage with grade-level mathematics at higher levels of complexity. The program provides occasional opportunities for students to engage with grade-level mathematics at higher levels of complexity through Adaptive Practice and Review Questions.  However, these are additional to the lesson so not all advanced students would be provided access to them. The advanced students would be completing more assignments than their classmates. Examples include, but are not limited to: • In Lesson 2.3, Adding Integers and Other Rational Numbers,  Review Questions, the Teacher Notes state, “To customize the questions click here:” Under Assign to Class, Customize, Add question set, various “hard” questions may be assigned. For example, in “Set 1, Question 4, Simplify \frac{14}{70}+\frac{74}{35}.” (7.NS.1) • In Lesson 7.4,  Theoretical Probability Models,Review Questions, the Teacher Notes state, “To customize the questions click here:” Under Assign to Class, Customize, Add question set, various “hard” questions may be assigned. For example, in “Set 1, Question 1, Steve was asked to pick a number from 1 to 100. The probability of obtaining a prime number is 0.12.” (7.SP.5 & 7.SP.6) ##### Indicator {{'3o' | indicatorName}} Materials provide varied approaches to learning tasks over time and variety in how students are expected to demonstrate their learning with opportunities for students to monitor their learning. The materials reviewed for CK-12  Interactive Middle School Math 7 for CSS provide varied approaches to learning tasks over time and variety in how students are expected to demonstrate their learning with opportunities for students to monitor their learning. Students can demonstrate learning through Inline Questions, Review Questions, and Adaptive Practice. The Interactives offer additional opportunities for students to demonstrate their learning. Some of the Discussion Questions offer multiple solution paths, and the Inline and Review Questions give immediate feedback to the student. Student reports provide levels of mastery: beginning, exploring, developing, proficient or mastery. These reports give the students an idea of how well they are doing on a specific concept. Throughout the materials students work through Interactives that have a variety of outcomes. Students also answer Inline and Review Questions and have discussions that build off of the Interactives. For example, in Lesson 1.3, Identifying Proportional Relationships in Tables, Activity 2, the Teacher Notes state, “This Interactive shows 12 tables, each with 4 x-y values. Students are instructed to click on the tables that show proportional relationships. Once students click the tables they believe are correct, there is a button at the bottom labeled ‘Check.’ Once students click this button, a message will appear telling them whether or not they selected the correct tables, correct tables will turn green, and incorrect tables will turn red. At this point, you can encourage students to persevere in solving the Interactive. Students can click the red tables again to un-select them and can ‘check’ again if they wish.” (7.RP.2) Students have opportunities to share and compare their thinking with others. In many lessons, students discuss their findings during the Interactives and following Inline Questions. Sometimes students are asked to compare their thinking with others. Examples include: • In Lesson 1.5, Unit Rate and The Constant of Proportionality, Activity 1, the Discussion Question asks, “Jamal comes to the conclusion that the unit rate and the constant of proportionality are the same thing. Do you agree with Jamal or disagree? How are the constant of proportionality and the unit rate of a relationship similar and how are they different?” The Teacher Notes state, “This question has been designed to help students analyze the arguments of others and construct a viable argument. Encourage the students to look through previous activities and lessons to build evidence that supports the differences and similarities that they will use to either support or disprove Jamal's argument.” (7.RP.2) • In Lesson 2.1, Combining Opposite Numbers,  Activity 3, the Discussion Question directs students to, “Discuss other types of situations that you could use a number line to help you with.” The teacher notes state, “This question is open-ended, designed to get students to think beyond the typical number line.“ (7.NS.1) Students are able to reflect on their work and understand where they are in their learning through different reports, like the Heat Map. The reports that the student receives on the Adaptive Practice give feedback based on the difficulty level of each question answered, but there is no self-reflection. ##### Indicator {{'3p' | indicatorName}} Materials provide opportunities for teachers to use a variety of grouping strategies. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS partially provide opportunities for teachers to use a variety of grouping strategies. The program does include materials designed for each child’s regular and active participation in grade-level content. However, the majority of the lessons are based on individual instruction. Lesson instructions in the Teacher Notes provide teachers with suggestions for grouping strategies that include small-group options, working with partners and individual instruction. However, there is no guidance provided to the teacher on how to assign partners or on how to form the group based on the different needs of the students. Examples include, but are not limited to: • In Lesson 3.5, Solving Multi-Step with Rational Numbers, Activity 1, Shopping Discounts, Discussion Question, Teacher Notes suggest, “The students should begin with a turn and talk or small group discussion to help form their arguments. The discussion should then transition into a full class discussion to allow the students the opportunity to analyze the arguments of their classmates.” (7.EE.B) There is no guidance as to how to form the small groups. • In Lesson 4.9, Solving Problems Involving Taxes, Commissions, and Fees, the Warm Up’s Extension Activity gives a recommendation in the Teacher Edition. It states, “Allow students to pair up and write questions in the format of the previous two questions where they provide a sales price and either an amount of tax or a price after taxes. Their partner will have to guess in which state they could have purchased the item. Students can take turns.” There is no guidance as to how to form the pairs. (7.RP.3) ##### Indicator {{'3q' | indicatorName}} Materials provide strategies and supports for students who read, write, and/or speak in a language other than English to regularly participate in learning grade-level mathematics. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS partially meet expectations for providing strategies and supports for students who read, write, and/or speak in a language other than English to regularly participate in learning grade-level mathematics. The materials provide a means to change the language of the main text to any of the supported languages, which includes the directions for the Interactives. However, the text within the Interactive will not change, and the video content will still be in English. Additionally, the Adaptive Practice, which is expected to be homework, is available in two languages: English and Spanish. The materials do not provide any other strategies or support for students who read, write, and/or speak in a language other than English beyond changing the language of the text. ##### Indicator {{'3r' | indicatorName}} Materials provide a balance of images or information about people, representing various demographic and physical characteristics. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS partially provide a balance of images or information about people, representing various demographic and physical characteristics. The materials do not contain many images depicting people. The Interactives have images of things or shapes.  Students with disabilities are not included. Since this is a digital series, the names in the text can be changed to make it more relatable to students. Many of the questions do not use names, just non-specific gender terms such as the following: you, the student, the class, ... etc.  Although athletes in pictures are generally male, an equal number of male and female names are used. However, only a few names appear to represent different races. Examples include: • In Lesson 2.1, Combining Opposite Numbers, Activity 2, it begins, “Peter is painting his wooden fence white.” (7.NS.1) • In Lesson 2.6, Multi-Step Addition & Subtraction Problems Involving Rational Numbers, Activity 2, it explains, “Cheryl's friend Annika is visiting her in Washington DC.” (7.NS.2) • In Lesson 7.3, Discerning Between Equally Likely and Non-Equally Likely Outcomes, Activity 1, it states, “Sai and his friend Kai are rolling dice to see who gets the higher number.” (7.SP.C.7) ##### Indicator {{'3s' | indicatorName}} Materials provide guidance to encourage teachers to draw upon student home language to facilitate learning. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS do not provide guidance to encourage teachers to draw upon student home language to facilitate learning. There is no evidence of promoting home language knowledge as an asset to engage students or purposefully utilizing student home language in context with the materials. ##### Indicator {{'3t' | indicatorName}} Materials provide guidance to encourage teachers to draw upon student cultural and social backgrounds to facilitate learning. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS do not provide guidance to encourage teachers to draw upon student cultural and social backgrounds to facilitate learning. While there is some culture implied by names or problem contexts, specific guidance on how to connect students' cultural and/or social background to facilitate learning or motivate students is not found. ##### Indicator {{'3u' | indicatorName}} Materials provide supports for different reading levels to ensure accessibility for students. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS do not provide supports for different reading levels to ensure accessibility for students. While there are some videos and other tools available under the Related Content section, they do not identify strategies to engage students of different reading levels to ensure accessibility. Some of the Teacher Notes suggest that teachers encourage the students to use the proper vocabulary, but the materials provide no specific strategies for supporting students at different reading levels or grouping students by reading levels. ##### Indicator {{'3v' | indicatorName}} Manipulatives, both virtual and physical, are accurate representations of the mathematical objects they represent and, when appropriate, are connected to written methods. The materials reviewed for CK-12  Interactive Middle School Math 7 for CCSS meet expectations for providing manipulatives, both virtual and physical. They are accurate representations of the mathematical objects they represent and, when appropriate, are connected to written methods. The materials provide suggestions and/or links for virtual and physical manipulatives that support the understanding of grade-level concepts. Manipulatives are accurate representations of the mathematical objects they represent and are sometimes connected to written methods. Physical manipulatives, while not included with the series, are listed in the beginning of the Teacher Edition under the Resource tab. The use of physical manipulatives is minimal. Each lesson contains several Interactives where students use virtual manipulatives to gain an understanding of the math standard they are learning. They include a variety of manipulatives such as: graphs, x-y tables, number lines, coordinate planes, GeoGebra Interactives, word matching problems, tape diagrams, dice and playing cards. Examples include: • In Lesson 1.1, Unit Rate and Proportionality, Activity 3, students are understanding rates as a ratio. It states, “If a car travels 300 miles in 5 hours, how far does it travel in 1 hour? How far does a car travel in 1 hour if it travels 350 miles in 5 hours? Drag the car back and forth to determine how many miles the car drives in one hour.  When looking at the point value for the car, the x-value is the hours, and the y-value is the miles.“ (7.RP.1) • In Lesson 6.2, Area and Circumference of Circles, Activity 1, students use the Interactive to discover parts of a circle. It states, “Students will be practicing labeling the parts of a circle: radius, center, diameter, and circumference. At the bottom are draggable words that students can place on the black line next to the appropriate part. If a work is placed in the wrong spot, it will turn red.” (7.G.4) #### Criterion 3.4: Intentional Design The program includes a visual design that is engaging and references or integrates digital technology, when applicable, with guidance for teachers. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS integrate technology such as interactive tools, virtual manipulatives/objects, and/or dynamic mathematics software in ways that engage students in the grade-level standards, and the materials include or reference digital technology that provides opportunities for teachers and/or students to collaborate with each other. The materials have a visual design that supports students in engaging thoughtfully with the subject, and is neither distracting nor chaotic, and the materials provide teacher guidance for the use of embedded technology to support and enhance student learning. ##### Indicator {{'3w' | indicatorName}} Materials integrate technology such as interactive tools, virtual manipulatives/objects, and/or dynamic mathematics software in ways that engage students in the grade-level/series standards, when applicable. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS integrate technology such as interactive tools, virtual manipulatives/objects, and/or dynamic mathematics software in ways that engage students in the grade-level standards, when applicable. The materials integrate technology in ways that engage students in the grade-level standards and are aligned to the standards and the Mathematical Practices. Third party programs such as Geogebra are used to assist with simulations and the data collection tool. Insight is available for teachers to use to gauge engagement and performance. Each lesson includes Interactives that relate to the concept and engage students in the process of learning. However, the Interactives cannot be customized. ##### Indicator {{'3x' | indicatorName}} Materials include or reference digital technology that provides opportunities for teachers and/or students to collaborate with each other, when applicable. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS include or reference digital technology that provides opportunities for teachers and/or students to collaborate with each other, when applicable. Students can collaborate with other students through the CK-12 Cafe, Math, and PLIX Corner. The Math Corner is for students to ask questions or help other students. The PLIX Corner is where students can discover and discuss the Interactives found throughout CK-12 concepts. Teachers are also able to collaborate with students through the Math and PLIX Corner. Teachers can collaborate with other teachers through the CK-12 Cafe, Jumpstart for Educators, which allows all teachers with access to the materials, to “ask questions, collaborate, and explore CK-12 in this forum for educators.” ##### Indicator {{'3y' | indicatorName}} The visual design (whether in print or digital) supports students in engaging thoughtfully with the subject, and is neither distracting nor chaotic. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS have a visual design that supports students in engaging thoughtfully with the subject and is neither distracting nor chaotic. The lessons follow a consistent format and the print, as well as any graphics, are easy to follow and do not detract from the math. Each lesson starts with a Warm Up and is followed by activities that contain Interactives with Inline Questions and sometimes Discussion Questions. At the end of each lesson is a set of Review Questions for students. This format is consistent in each chapter throughout all grade levels. The graphics are visually appealing and support student understanding of the concepts. The font size, directions and text are appropriate for the grade level.  The format is engaging, and the Interactives have clear directions that make them easy to use. ##### Indicator {{'3z' | indicatorName}} Materials provide teacher guidance for the use of embedded technology to support and enhance student learning, when applicable. The materials reviewed for CK-12 Interactive Middle School Math 7 for CCSS provide teacher guidance for the use of embedded technology to support and enhance student learning, when applicable. All lessons include embedded technology in the form of Interactives. The Teacher Notes give guidance on how to use the technology to enhance student learning. Inline and Discussion Questions often follow these Interactives. Examples include: • In Lesson 4.8, Calculating Simple Interest, Activity 3, the Teacher Notes on the Interactive state, “This Interactive helps students practice finding the interest, given the principle amount, rate and amount of time. First is the scenario, next the breakdown of what each value is, and lastly, the text box for the answer. When the student types in their answer and presses Check, they will find out if their answer was correct or not and if they are correct, they can either move on or try a new number.” (7.RP.3) • In Lesson 8.1, Analyzing Data Sets Visually and Numerically: Review, Activity 2, the Teacher Notes on the Interactive state, “In this Interactive, students will get to experiment with how changing data affects the mean and mean absolute deviation. In this data set, there are five values that can range from one to nine. Below that are the mean and mean absolute deviation of the current data set. Students can use the sliders above each number to change its value. The mean and MAD will automatically be calculated as each value changes. “ (7.SP.4) ## Report Overview ### Summary of Alignment & Usability for CK-12 Interactive Middle School Math for CCSS | Math #### Math 6-8 The materials reviewed for CK-12 Interactive Middle School Math 6-8 for CCSS meet expectations for Alignment to the CCSSM. In Gateway 1, the materials meet expectations for focus and coherence. In Gateway 2, the materials meet expectations for rigor and practice-content connections. In Gateway 3, the materials partially meet expectations for Usability. Within Gateway 3, the materials partially meet expectations for Teacher Supports (Criterion 1), meet expectations for Assessment (Criterion 2), and do not meet expectations for Student Supports (Criterion 3). ###### Alignment Meets Expectations ###### Usability Partially Meets Expectations ###### Alignment Meets Expectations ###### Usability Partially Meets Expectations ###### Alignment Meets Expectations ###### Usability Partially Meets Expectations ## Report for {{ report.grade.shortname }} ### Overall Summary ###### Alignment {{ report.alignment.label }} ###### Usability {{ report.usability.label }} ### {{ gateway.title }} ##### Gateway {{ gateway.number }} {{ gateway.status.label }}
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4307944178581238, "perplexity": 2421.614361292838}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103821173.44/warc/CC-MAIN-20220630122857-20220630152857-00181.warc.gz"}
http://mathoverflow.net/questions/93665/graphs-with-maximal-number-of-paths-of-given-length
## graphs with maximal number of paths of given length Hi, For a given number of edges, the non directed graph which maximises the number of paths of length 2 is the quasi-star or the quasi-complete graph. Does anyone know : 1- what is the non directed graph which maximises the number of paths of length 3 2- same question for length 4 3- same question for length 3 + 4 If you have any hint, thank you very much. - Is the number of vertices also given? – Felix Goldberg Apr 10 2012 at 16:21 yes, sorry I forgot to mention. The number of vertices is also given. But if it helps, consider it can be as large as you wish and you don't need to connect them all. And by the way, I just noticed my mistake: I need the graphs which maximise the number of walks, not paths ! Sorry again – unknown (google) Apr 10 2012 at 18:58 Just a random note. Let $A$ be the adjacency matrix of your graph. The number of walks of length $k$ between vertices $v_i,v_j$ is given by the $(v_i,v_j)$ entry of $A^k$. Maybe this can help you somehow. – Jernej Apr 10 2012 at 20:40 yes, I know that. But I don't see how I can use it. – unknown (google) Apr 10 2012 at 20:43
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8879719972610474, "perplexity": 243.00497693813543}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368703592489/warc/CC-MAIN-20130516112632-00015-ip-10-60-113-184.ec2.internal.warc.gz"}
http://mathhelpforum.com/advanced-math-topics/28696-solved-can-anyone-help.html
# Math Help - [SOLVED] Can anyone help? 1. ## [SOLVED] Can anyone help? I was wondering if there is any kind of formula that will allow me to do this: I have about 20 numbers. I want a combination of some of the numbers to add up to be (let's say) \$25,768.45. I don't have to use all the numbers to make them add up to that number. I can use 2 or I can use all 20. Is there any kind of formula I can use to help me do this? The numbers are all big and the final number I need them to add up to is big too so I've just been plugging in numbers and guessing but it is taking forever AND I cannot make it work. I realize it may not work, but I thought maybe there is a trick or a shortcut to this. Thanks. Sorry if this makes no sense! 2. Given a set of twenty numbers, there are a maximum of $2^{20} - 1 = 1048575$ possible sums. (There may be several equal sums, but that is the number of different sums.) Given the power of today’s computers, it should be possible to actually check out each possible sum. But there is no mathematical formula to do this for an arbitrary set.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 1, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7665759921073914, "perplexity": 138.85215423680697}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207928780.77/warc/CC-MAIN-20150521113208-00259-ip-10-180-206-219.ec2.internal.warc.gz"}
https://search.datacite.org/works/10.4230/LIPICS.STACS.2012.408
### On the treewidth and related parameters of random geometric graphs Dieter Mitsche & Guillem Perarnau We give asymptotically exact values for the treewidth tw(G) of a random geometric graph G(n,r) in [0,sqrt(n)]^2. More precisely, we show that there exists some c_1 > 0, such that for any constant 0 < r < c_1, tw(G)=Theta(log(n)/loglog(n)), and also, there exists some c_2 > c_1, such that for any r=r(n)> c_2, tw(G)=Theta(r sqrt(n)). Our proofs show that for the corresponding values of r the same asymptotic bounds also hold for the pathwidth and...
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9643430709838867, "perplexity": 1197.139850451982}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267867364.94/warc/CC-MAIN-20180625014226-20180625034226-00255.warc.gz"}
https://oldschoolrunescape.fandom.com/wiki/Justiciar_armour
## FANDOM 19,873 Pages Justiciar armour (pronounced "jus-TISH-er") is a set of armour requiring 75 Defence to equip. It can only be obtained as rare drops from the Theatre of Blood. It is the best armour in RuneScape to provide Melee and Ranged defence bonuses, as well as having a damage-reducing set effect. The justiciar armour also provides a significant amount of Prayer bonus. When the full set of Justiciar armour is equipped, the player gains a set effect: all combat damage taken is reduced (except in PvP). The exact formula for the amount of damage reduced is $\frac{bonus}{3000}$, where bonus is the player's defence bonus for that particular style. For example, if an enemy was using a crush style attack, and the player has a crush defence bonus of +450, then 15% of the damage (450/3,000) is reduced. The set effect does stack with the elysian spirit shield's passive effect. The armour can be stored in the armour case of a player-owned house costume room. The set pieces count as Saradominist items in both the normal and Wilderness God Wars Dungeons. ## Components Item Exchange price Total price Justiciar faceguard 31,313,915 67,648,509 Justiciar chestguard 18,269,480 Justiciar legguards 18,065,114 ## Bonuses Item Attack Bonuses Defence Bonuses Other Justiciar faceguard 0 0 0 -6 -2 +60 +63 +59 -6 +67 0 0 0 +2 Justiciar chestguard 0 0 0 -40 -20 +132 +130 +117 -16 +142 0 0 0 +4 Justiciar legguards 0 0 0 -31 -17 +95 +92 +93 -14 +102 0 0 0 +4 Totals 0 0 0 -77 -39 +287 +285 +269 -36 +311 0 0 0 +10 ## Trivia Community content is available under CC-BY-SA unless otherwise noted.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.18631993234157562, "perplexity": 7471.115823696395}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370500426.22/warc/CC-MAIN-20200331084941-20200331114941-00234.warc.gz"}
https://www.vocal.com/beamforming-2/high-resolution-far-field-direction-of-arrival-using-three-microphones/
Recent advances in technology has led to requirements for finding direction of arrivals of acoustic disturbances in large open areas such as a car park or a stadium. Activities such as a car crash, an altercation or a discharge of a fire arm requires narrowed view of search to shorten the resolution time. In this regard, we presents  a solution to direction of arrival estimate of sound sources with a full circle view and attaining a resolution of less than $7^{\circ}$ for sources that are at distances of up to $100m$ away. Without loss of generality, our solution utilizes three non co-linear microphones but the approach easily scales to more than three microphones. Suppose $F_s \ge \frac{c}{d}$ , where $F_s$ is the sampling rate, $c$ is the speed of sound in free space, and $d$ is the microphone pairwise separation. This assumption ensures there is at least one full sample delay possible between pairwise microphones. Suppose also, for the purpose of this article that  $F_s = 16kHz$ with the separation distance between the three microphones  $d = 42.5 ~mm$. The orientation of the three microphones is as shown in Figure 1 below. Figure 1: Three microphone array topology. It can be shown that the direction of arrival for a far field acoustic source impinging on the triangular array topology in Figure 1 will obey: $\theta = \arctan{\left( \frac{\sqrt{3}(\tau_{1,2} -\tau_{1,3})}{\tau_{1,2} +\tau_{1,3}+2 \tau_{2,3}}\right)}$ where $\tau_{i,j}$ is the delay between microphones $i$ and $j$, $\theta$ is the angle to be estimated, and $d$ is the pairwise microphone distances apart with the signs of the numerator and denominator determining the quadrant in which the angle is located. The resolved angles will be as shown in Figure 2 below. Figure 2:  Direction of arrival resolution. The maximum resolution deviation falls within $\pm 7^{\circ}$. This bound can be further reduced if inter sample delays are employed for the pairwise microphone signal delays. VOCAL Technologies offers custom designed solutions for beamforming with a direction of arrival estimation, robust voice activity detector, acoustic echo cancellation and noise suppression. Our custom implementations of such systems are meant to deliver optimum performance for your specific beamforming task. Contact us today to discuss your solution!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 14, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6656613945960999, "perplexity": 727.1810067615036}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400201826.20/warc/CC-MAIN-20200921143722-20200921173722-00466.warc.gz"}
https://computing.ece.vt.edu/~lingjialiu/doku.php?id=kucourses:eecs965:schedules13
# Lingjia Liu's Webpage ### Site Tools kucourses:eecs965:schedules13 # EECS 965 Course Schedule ## Spring 2013 ### Detailed Schedule Date Topic Assignment 01/22/2013 Lecture 1: course objective and review of probability 01/24/2013 Lecture 2: review of vector space and matrix analysis 01/29/2013 Lecture 3: introduction to statistical signal processing 01/31/2013 Lecture 4: binary hypothesis testing 02/05/2013 Lecture 5: Bayesian hypothesis testing and Neyman-Pearson hypothesis testingHomework 1 is assigned. 02/07/2013 Lecture 6: minimum variance unbiased estimation 02/12/2013 Lecture 7: Cramer-Rao lower bound (CRLB)Homework 1 is due, Homework 2 is assigned. 02/14/2013 Lecture 8: general CRLB for singlas in white Gaussian noise 02/19/2013 Lecture 9: linear models Homework 2 is due, Homework 3 is assigned. 02/28/2013 Lecture 10: system identification problem Homework 3 is due. 03/05/2013 Lecture 11: general minimum variance unbiased estimation 03/07/2013 Lecture 12: best linear unbiased estimator (BLUE) 03/12/2013 Lecture 13: maximum likelihood estimation (MLE) 03/14/2013 Lecture 14: least square estimation (LSE) 03/26/2013 Lecture 15: geometric intepretation, order-recursive and sequential LSE Homework 4 is assigned. 03/28/2013 Lecture 16: constrained LSE and nonlinear LSE 04/02/2013 Lecture 17: methods of moments Homework 4 is due. 04/04/2013 Lecture 18: review of classic estimation 04/09/2013 Mid-term exam. 04/11/2013 Lecture 19: Bayesian estimation
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8519525527954102, "perplexity": 18568.08329788807}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-43/segments/1634323585246.50/warc/CC-MAIN-20211019074128-20211019104128-00385.warc.gz"}
http://community.wolfram.com/groups/-/m/t/248113?p_p_auth=W55OM9Iq
# How to multiply several matrices? GROUPS: The code Sum[(Cos)^2 - (Cos)^2, {i, 1, n}] provides one with a sum from 1 to n. Similarly, Product[] gives a sum from 1 to n.I would be interested in a similar command for a product of n matrices; I have to specify that I really wanna use the Dot multiplication, not the componentwise one. And n is not specified, so I am exactly in the same situation as for Sum and Product.Does anybody know how to do this?Many thanks in advance! Answer 1 year ago 1 Reply In[2]:= mat[n_] := {{n, n + 1}, {n + 2, n + 3}}  In[8]:= mat[1].mat[2].mat[3]  Out[8]= {{95, 118}, {211, 262}}  In[9]:= Dot[Sequence @@ Table[mat[i], {i, 3}]]  Out[9]= {{95, 118}, {211, 262}} Answer 1 year ago
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8587815761566162, "perplexity": 1771.4882197204045}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-40/segments/1443737940794.79/warc/CC-MAIN-20151001221900-00175-ip-10-137-6-227.ec2.internal.warc.gz"}
http://questionbank4you.blogspot.com/2012/01/circuit-theory-two-mark-questions-with.html
## Friday, 6 January 2012 ### Circuit Theory two mark questions with answer Department of Electrical and Electronics Engineering. Electric Circuits UNIT 1 (Basic Circuit concepts) 1. What do you mean by an electric network and an electric circuit Interconnection of two or more simple circuit elements (voltage sources, resistors, inductors and capacitors) is called an electric network. If a network contains at least one closed path, it is called an electric circuit. 1. Classify the network elements i)                    Active or passive ii)                  Unilateral or bilateral iii)                Linear or non linear iv)                Lumped or distributed 1. Differentiate active and passive elements Energy sources (voltage or current sources) are active elements, capable of delivering power to some external device. The passive elements are those which are capable only of receiving power. 1. Differentiate bilateral and unilateral elements In the bilateral element, the voltage - current relation is the same for current flowing in either direction. In contrast, a unilateral element has different relations between voltage and current for the two possible directions of current. 1. Differentiate linear and nonlinear elements An element is said to be linear, if it satisfies the linear voltage current relationship. That means the current through the element is multiplied by some constant ‘α’, and results in the multiplication of voltage across the element by the same constant. An element which does not satisfy this relation is called a nonlinear element. 1. Define lumped and distributed elements in electric circuits The elements which are separated physically are known as lumped elements, like resistors, capacitors and inductors. Distributed elements, on the other hand, are those which are not separable for analytical purposes. 1. Write the equations of energy stored by an inductor and capacitor Energy stored by an inductor is, W = LI2 / 2 I à Current flow through the inductor Energy stored by a capacitor W = CV2 / 2 V à Voltage across capacitor 1. State Kirchhoff’s voltage and current law Kirchhoff’s voltage law: It states that the algebraic sum of voltages around any closed path in a circuit is always zero. Kirchhoff’s current law: It states that the sum of currents entering into any point is equal to the sum of the currents leaving that point. 1. Write the equation for voltage for the given circuit using voltage division technique 1. Write the equation for current for the given circuit using current division technique 1. How can a practical voltage source, having an ideal voltage VS and internal series resistance RV be replaced by a current source? The practical voltage source, having an ideal voltage Vs and internal series resistance RV be replaced by a current source I = VS / RV in parallel with an internal resistance       RI = RV UNIT II   (Sinusoidal steady state analysis) 1. Write the equation for Instantaneous active power, Average power, Apparent power Instantaneous active power, P(t) = VmIm[Cos(2ωt+θ)+Cos θ] / 2 Average power, Pav = VmIm Cos θ / 2 Apparent power, Papp = Veff Ieff 1. Define Power factor It is defined as the ratio of average power to the apparent power, whereas the apparent power is the product of the effective values of the current and the voltage. Power factor = Pav / (Veff Ieff) It is also defined as the factor with which the volt amperes are to be multiplied to get true power in the circuit 1. Draw the power triangle 1. Give any one application of resonance In the area of communications, the ability of a radio receiver to select a certain frequency, transmitted by a station and to eliminate frequencies from other stations is based on the principle of resonance. 1. Define resonance frequency for a series RLC circuit The frequency at which resonance occurs is called the resonant frequency. fr = 1 / (2π√LC) L à Inductance of the circuit C à Capacitance of the circuit 1. Define Bandwidth of a RLC circuit It is the range of frequencies for which the current or output voltage is equal to 70.7 % of its value at the resonant frequency. 1. Define the Q factor of a RLC series circuit The Quality factor, Q, is the ratio of the reactive power in the inductor or capacitor to the true power in the resistance in series with the coil or capacitor. Q = 2π * (maximum energy stored / energy dissipated per cycle) 1. Define Magnification in resonance The ratio of voltage across either L or C to the voltage applied at resonance can be defined as magnification. Magnification = (VL / V) or (VC / V) 1. What is tank circuit? The parallel resonant circuit is generally called a tank circuit, because of the fact that, the circuit stores energy in the magnetic field of the coil and in the electric field of the capacitor. The stored energy is transferred back and forth between the capacitor and coil, and vice-versa. 1. Write the types of balanced and unbalanced loads in a three phase system? ii)                  Unbalanced three wire star connected load iii)                Unbalanced four wire star connected load 1. The neutral point of an unbalanced three wire star connected load is called as a floating neutral point. Why? The potential of the load star point is different from that of the supply star point. The result is that the load phase voltages are not equal to the supply phase voltage and they are not only unequal in magnitude, but also subtend angles other than 120o with one another. The magnitude of each phase voltages depends upon the individual phase loads. The potential of the load neutral point changes according to changes in the impedances of the phases, that is why sometimes the load neutral is also called floating neutral point. 1. When the wattmeter will give down scale reading? How to measure the power under such condition? If the phase angle between the line voltage and current is greater than 90o, the corresponding wattmeter would indicate downscale reading. To obtain upscale reading, the connections of either the current coil or the pressure coil have to be interchanged. Reading obtained after reversal of coil connection should be taken as negative. Then, the algebraic sum of the two wattmeter readings gives the total power. Unit – III (Circuits and theorems) 1.                  What do you mean by a planar and a non planar circuit? A circuit is said to be planar, if it can be drawn on a plane surface without crossovers. A non planar circuit cannot be drawn on a plane surface without a crossover. 2.                  State Superposition theorem It states that in any linear network containing two or more sources, the response in any element is equal to the algebraic sum of the responses caused by individual sources acting alone, while the other sources are non operative; that is, while considering the effect of individual sources, other ideal voltage sources and ideal current sources in the network are replaced by short circuit and open circuit across their terminals. This theorem is valid only for linear systems. The Superposition theorem is not valid for power responses. It is applicable only for computing voltage and current responses. 3.                  State Thevenin’s theorem It states that any circuit having a number of voltage sources, resistances (impedances) and open output terminals can be replaced by a simple equivalent circuit consisting of a single voltage source in series with a resistance (impedances), where the value of the voltage source is equal to the open circuit voltage across the output terminals, and resistance (impedance) is equal to the resistance seen in to the network across the output terminals. 4.                  State Norton’s theorem It states that any circuit with voltage sources, resistances (impedances) and open output terminals can be replaced by a single current source in parallel with single resistance (impedance), where the value of current source is equal to the current passing through the short circuit output terminals and the value of the resistance (impedance) is equal to the resistance seen into the output terminals. 5.                  State Reciprocity theorem According to this theorem, if we apply some input( from either a current source, or a voltage source) to a circuit which consists of resistors, inductors, capacitors and transformers, including coupled circuits, the ratio of response(output) in any element to the input is constant even when the position of input and output terminals are interchanged. This is called the Reciprocity theorem. 6.                  State Compensation(Substitution) theorem It states that any element in the network, linear or non linear, may be replaced by a voltage source of magnitude equal to line current, passing through the element multiplied by the value of the element, provided the currents and voltages in other parts of the circuit remain unaltered. This theorem is useful to find the changes in current or voltage when the value of resistance is changed in the circuit. 7.                  State Reciprocity theorem It states that maximum power is delivered from a source to a load when the load resistance is equal to the source resistance. 8.                  State Millman’s theorem Unit – IV (Response of Electric Circuits) 1. What is transient state? If a network contains energy storage elements, with change in excitation, the current and voltages change from one state to other state. The behaviour of the voltage (or) current when it is changed from one state to another state is called transient state. 1. What is natural response? If a circuit containing storage elements which are independent of sources, the response depends upon the nature of the circuit, it is called natural response. 1. What is meant by natural frequency? If the damping is made zero then the response oscillates with natural frequency without any opposition, such a frequency is called natural frequency of oscillation, denoted as ωn. 1. Define Damping ratio. It is the ratio of actual resistance R in the circuit to the critical resistance RC. It is denoted by greek letter zeta ζ. . 1. Write down few applications of RL, RC and RLC circuits. The few applications of RL,RC and RLC circuits are given below. Coupling circuits, Phase-shift circuits, Filters, Resonant circuits, AC bridge circuits, and Transformers. 1. What is the time-constant for RL and RC circuits? Time constant for RL circuit, =L/R Time constant for RC circuit, =1/RC 1. How does an inductor act at t=0+ and t=∞? An inductor will act as a open circuit at t=0+ and will act as a short circuit at t=∞. 1. Obtain the natural frequency and time constant of an RLC series circuit with R=1K, L=100mH and C=0.1F. 1. What is meant by over damped and under damped system? 1. When do we get transient response? Whenever a circuit is switched from one condition to another either by a change in the applied source (or) change in the circuit elements there is a transitional period during which the branch currents and voltage change from their values to new ones, this period is called transient. 1. Distinguish between steady state and transient response of an electric circuit. SL.No STEADY STATE RESPONSE TRANSIENT RESPONSE 1 Amplitude will not change. Amplitude may change. 2 Frequency will not change. Frequency may change. 3 Constant voltage and current with time. Change from one steady state to another. 4 Algebraic equations are used. Integro-differential equations are used 1. What are the causes of transient behavior occurring in a circuit? The causes are a)      It may be due to the sudden change of applied voltage. b)      When the voltage source is shorted. c)      When a circuit is connected or disconnected and d)     Due to storage elements in the circuit. 1. Define Decrement factor or logarithmic decrement. 1. What is the Laplace transform  ? 1. What is the cause of free response in electric circuits? The presence of storage element such as inductors or capacitors or both in a network in the absence of external energy source is the cause of free response. 1. Find the poles and zeros for the function , and also obtain the pole-zero plot. Unit – V (Two port network and Filters) 1. Define one-port network. A pair of terminals at which a signal may enter (or) leave a terminal is called a port. A network having only one such pair of terminals is called one-port network. 1. Define transmission parameters. 1. Write ABCD parameters in terms of Z-parameters. ABCD parameters in terms of Z-parameters is given below. A circuit consisting of constant sources is said to be in steady state if the voltages and currents do not change with time. 1. What is meant by free response? Free response is due to the internal energy stored in the network. It depends upon the type of elements, their size etc. This response is independent of the source. This  response dies gradually, i.e., it approaches zero as time becomes infinity. Free response is also known as natural (or) transient response. 1. Distinguish between free and forced response. When a circuit contains storage elements which are independent of the sources, the response depends upon the nature of circuit. This response is called natural (or) free response. The storage elements deliver the energy to the resistances. So the response changes with time, gets saturated after some time. It is referred to as the transient response. When we consider sources acting on a circuit, the response depends on the nature of such sources. This response is called forced response. 10. What is a Low pass filter? 11. Draw the characteristics of ideal low pass filter. 12. What is a high pass filter? A high-pass filter is a filter that passes high frequencies well, but attenuates (reduces the amplitude of) frequencies lower than the cutoff frequency. 13. Draw a high pass and low pass filter using RC. 14. Give the applications of filter. High-pass and low-pass filters are also used in digital image processing to perform transformations in the spatial frequency domain. Most high-pass filters have zero gain (-inf dB) at DC. Such a high-pass filter with very low cutoff frequency can be used to block DC from a signal that is undesired in that signal (and pass nearly everything else). These are sometimes called DC blocking filters. 1. A constant k high pass filter is required for a cut-off frequency of 2500Hz. The resistance of load circuit is 600Ω. Determine the value of the components required. SOLUTION: 1. Each of the series elements of a T-type Low pass filter consists of an inductance of 30mH having negligible resistance and the shunt element has a capacitance of 0.2µF. Calculate the cut-off frequency and determine the iterative impedance at a frequency of 1000Hz. GIVEN: L/2=30mHL=60mH, C=0.2µF SOLUTION: CONFIGURATION OF T-SECTION Iterative impedance (or) Characteristic impedance,
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.903038501739502, "perplexity": 1049.7466621814847}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-23/segments/1406510260734.19/warc/CC-MAIN-20140728011740-00448-ip-10-146-231-18.ec2.internal.warc.gz"}
https://ftapia.dev/es/publication/white-2020/
# The MESAS Project: ALMA Observations of the F-type Stars γ Lep, γ Vir A, and γ Vir B White et al. 2020 ### Resumen The spectrum of stars in the submillimeter to centimeter wavelength range remains poorly constrained due to a lack of data for most spectral types. An accurate characterization of stellar emission in this regime is needed to test stellar atmosphere models, and is also essential for revealing emission associated with unresolved circumstellar debris. We present Atacama Large Millimeter/submillimeter Array observations of the three nearby, main-sequence, debris-poor, F-type stars γ Lep, γ Vir A, and γ Vir B at 0.87 and 1.29 mm. We use these data to constrain semiempirical atmospheric models. We discuss the atmospheric structure of these stars, explore potential short-term variability, and the potential impact on debris disk studies. These results are part of an ongoing campaign to obtain long wavelength observations of debris-poor stars, entitled Measuring the Emission of Stellar Atmospheres at Submillimeter/millimeter wavelengths. Publicación The Astrophysical Journal
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9932803511619568, "perplexity": 7079.439188790008}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335254.72/warc/CC-MAIN-20220928113848-20220928143848-00110.warc.gz"}
http://mathematica.stackexchange.com/questions/46507/how-can-i-get-a-piechart-to-be-colored-as-i-wish
# How can I get a PieChart to be colored as I wish? How to use the value of a funtion[label] to assign color to a piece of the pie? (edited after answer of kguler ) My dude is how to control the colors in ranges of a function of [labels] In this way I can show: 1.- the percentage of Data(labels), using pieces of pie and 2.- the value f of labels, with range colors (similar to TemperatureMap but customized) ------ Original Post ------ I want to use a "gradient" or "a reduce (5-6) list of colors"(similar to DarkRainbow with a little colors)? I was trying with ColorData and ColorDataFunction ( here you can see that I obtain errors with the same code as the reference doc ) but I have no succes!! percent = {12, 18, 24, 46} labels = {1, 9, 15, 22} f = {8, 7, 1, 4} PieChart[percent, ChartLabels -> labels, ImageSize -> 200, PlotLabel -> "Test-Grad", ChartStyle -> ColorData["TemperatureMap" {0, 8}]] PieChart[percent, ChartLabels -> labels, ImageSize -> 200, PlotLabel -> "Test-Comic", ChartStyle -> ColorData[9, "ColorList"]] In the image you can see any more about explanation: - Update: I think you after something like pc3b = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-3b", ImageSize -> 300, ChartStyle -> {ColorData[{"TemperatureMap", {0, 8}}] /@ f}] Similar trick with ColorFunction should work too... rule = Thread[percent -> f]; PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-4b", ImageSize -> 300, ColorFunctionScaling -> False, ColorFunction -> (ColorData[{"TemperatureMap", {0, 8}}][#1 /. rule] &)] Original post: percent = {12, 18, 24, 46}; labels = {1, 9, 15, 22}; f = {8, 7, 1, 4}; pc1 = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-1", ImageSize -> 200, ColorFunction -> "TemperatureMap"]; pc2 = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-2", ImageSize -> 200, ColorFunction -> Function[{angle}, ColorData["TemperatureMap"][angle]]]; (*the following based on guessing how you might want to use color gradients with f *) pc3 = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-3", ImageSize -> 200, ChartStyle -> {ColorData["TemperatureMap"][#] & /@ Rescale[f]}]; pc4 = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Grad-4", ImageSize -> 200, ColorFunction -> Function[{angle}, ColorData["TemperatureMap"][Nearest[Rescale[f], angle]]]]; pc5 = PieChart[percent, ChartLabels -> labels, PlotLabel -> "Test-Comic", ImageSize -> 200, ChartStyle -> 9]; Row[{pc1, pc2, pc3, pc4, pc5}, Spacer[5]] // Panel - Yes I think that´s the good way. f must be a parameter because the color depends on its value. I´m making a image to concrete more. I post in a few minutes – Mika Ike Apr 21 '14 at 8:29 Here are any aditional precisions, sensa.square7.ch/gdfgdfgd10.jpg – Mika Ike Apr 21 '14 at 8:41 Yes :-) , your answer is the cuasi-solution!! It works well and gives similar colors to similar values, and equale colors to equal values of f[label]. But the only detail I would like to control (if possible) is: 1.- define the colors (with CMYK or RGB code) and 2.- the ranges of values of f. – Mika Ike Apr 21 '14 at 9:24 I was seeing that red (how much red, f much high) and warm colors indicate that f is over the mean; and blue(how much blue, much low) and cold colors indicate the lowest values. And I can Reverse colors. If you can show me how to control colors and ranges(as explains in a previous image-link), fantastic; if not, this may be the solution for me, and I´m mark the question solved by you. – Mika Ike Apr 21 '14 at 9:46 @MikaIke, it would be a good idea to avoid questions and answers that keep getting larger/longer :) If the answer is ok for the original version of your question, I suggest you post a new question focusing on the new parts. – kglr Apr 21 '14 at 10:05 E.g.: PieChart[{10, 20, 40}, ChartLabels -> {"Blah", "blah", "blahhh"}, ImageSize -> 200, PlotLabel -> "Test-Grad", ColorFunction -> ColorData["DarkRainbow"]] Note there is no color scheme called "gradient" - there is a class of schemes called gradients, you can use ColorData["Gradients"] to see the particular named schemes. If you want more control over colors assigned, point ColorFunction to your own function that does whatever is needed to the data and returns the appropriate color(s). - No, The colors depends on the value os a function of the labels. In your example, if F(Blah)=100,F(blahhh)=95,F(blah)=50 . The colors of the piece of Blah has the same (or close/similar) of the piece of blahhhh, because the value of F(Blah) and F(blahhhh) is similar. If you use this colors sensa.square7.ch/hgfhgfg.jpg , Blah and blahh have, the two, a darker color , while the other is ligther. In this way you can show 2 dimensions of data (one with the percent and other with the color.) – Mika Ike Apr 21 '14 at 7:37 @MikaIke as I said, point it to your own color function definition. The above is just showing correct usage of the basic form which you have completely wrong in your example image. Read the documentation, everything needed is there. – ciao Apr 21 '14 at 7:45 I was seeing the documentation (ColorData, colorFunction, and any more) but I can´t understand how to do that question. I continue reading about the issue but Only to put any colors, but in your example how do you put similar colors to "Blah" and "blahhh"(supossing that f[Blah] and f[blahhh] is close)?. using Color functions. – Mika Ike Apr 21 '14 at 8:09
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.21318891644477844, "perplexity": 3345.9413327881707}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257826736.89/warc/CC-MAIN-20160723071026-00294-ip-10-185-27-174.ec2.internal.warc.gz"}
https://cocotec.io/fdr/manual/introduction.html
# Introduction¶ FDR is a tool for analysing programs written in Hoare’s CSP notation, in particular machine-readable CSP namely CSPM, which combines the operators of CSP with a functional programming language. The original FDR was written in 1991 by Formal Systems (Europe) Ltd, and a completely revised version FDR2 was released in the mid-1990s by the same organisation. The current version of the tool is FDR4, first released in October 2016 following FDR3 which was first released in 2013. Both of these versions were released by the University of Oxford, which also released FDR2 versions 2.90 and above in the period 2008-12. FDR4.0 has extremely similar functionality to FDR2.94, but is completely re-written. The main differences are: 1. The user interface has been completely revised. 2. The debugger has been completely revised and gives simultaneous information about all components of a system, rather than one at a time. 3. There is an integrated type checker for CSPM. 4. It now uses multi-core parallelism to speed up its operation. 5. A version of the ProBE CSP animator has been integrated. 6. There is a utility for drawing graphical representations of the labelled transition systems that represent processes within FDR. The only significant functionality of FDR2.94 that FDR4.0 lacks is support for the revivals and refusal testing models of CSP and their divergence-strict versions (i.e. [V=, [VD=, [R= and [RD=). Note that the batch mode of FDR2.94 has been replaced by a new machine-readable interface based on standard formats (JSON, XML and YAML are supported). FDR uses many algorithms and data structures. The ones used in FDR4 are in some cases the same, in some cases mildly modified, and in other cases completely new. Papers about FDR4 and its development can be found in References. Many books and papers have been written about CSP and earlier versions of FDR. ## Citing FDR¶ When citing FDR, please refer to the following paper: @inproceedings{fdr, title={{FDR3 --- A Modern Refinement Checker for CSP}}, author={Thomas Gibson-Robinson, Philip Armstrong, Alexandre Boulgakov, A.W. Roscoe}, booktitle={Tools and Algorithms for the Construction and Analysis of Systems}, year = {2014}, pages = {187-201}, volume={8413}, series={Lecture Notes in Computer Science}, editor={Ábrahám, Erika and Havelund, Klaus}, } The manual may be cited as: @manual{fdrmanual, title={{Failures Divergences Refinement (FDR) Version 3}}, author={Thomas Gibson-Robinson, Philip Armstrong, Alexandre Boulgakov, A.W. Roscoe}, year={2013}, url={https://www.cs.ox.ac.uk/projects/fdr/}, }
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.344181090593338, "perplexity": 3676.2715695699508}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141747774.97/warc/CC-MAIN-20201205104937-20201205134937-00182.warc.gz"}
https://www.physicsforums.com/threads/how-do-i-show-two-planes-are-neither-coincident-or-parallel.302337/
# How do I show two planes are neither coincident, or parallel? 1. Mar 24, 2009 ### Random-Hero- 1. The problem statement, all variables and given/known data Show that the two planes are neither coincident, parallel, nor distinct. Identify, geometrically, how the planes intersect and determine the angle between the two planes to the nearest degree. $$\pi$$1: x + 2y - 4z + 7 = 0 $$\pi$$2: 2x - 2y - 5z + 10 = 0 I found the normals to be n1 = (1, 2, -4) n2 = (2, -2, -5) 3. The attempt at a solution Two planes, if they intersect, intersect in a line. An equation of a plane has the form nxx + nyy + nzz = d where \vec n = <nx,ny,nz> is the normal vector to the plane. It can be shown that the angle between two planes' normal vector equals the angle between the planes. You can use the dot product \vec n1 dot \vec n2 = |n1||n2|cos(\theta) But how do I like......SHOW that they aren't either of those without drawing a diagram? Because it looks like like the first part of the question wants me to SHOW they are neither parallel or coincident, and the second part is finding the angle. So how would I show that they are neither parallel or coincident without proving the angle? All I can really think of is the fact that n2 $$\neq$$kn1$$\pi$$ Last edited: Mar 24, 2009 2. Mar 24, 2009 ### Chaos2009 Well, parallel planes don't intersect, so finding an intersection proves they're not parallel, right? Coincident planes lie on top of one another so showing that they intersect in a line proves that they aren't coincident. But, aren't these two planes distinct, or am I thinking of something else. 3. Mar 24, 2009 ### Random-Hero- I think that maybe my professor just made a typo, because that doesn't make sense, from everything I've found so far, suggest they are distinct. But how do I show that? Do I use the dot/cross product for the parallel and perpendicular things? 4. Mar 25, 2009 ### Staff: Mentor I agree. The problem statement is incorrect. Two planes in three-dimensional space (as yours are) must be either parallel or coincident. If they are parallel, they are distinct. If they intersection in a line, I would say that they are distinct, since we can tell them apart, but maybe your prof means distinct in the sense that they share no points at all. The two planes you are given have different normals, so they must intersect in a line, so it's not possible to prove that they are not parallel, not coincident, and not distinct. It's a little like saying, n is a positive integer - prove that it is not even and not odd. 5. Mar 25, 2009 ### xaos an inhomogenous system of two equations in three unknowns of the form Ax=b. look at the solution set of this system. if there are no solutions, then the two planes are parallel. if the solution set is a plane, then the two planes are coincident. if the solution set is a line, then the two planes intersect in a line and cannot be distinct (no points of intersection, if i interpret this correctly). it is impossible for the solution set to be a point for a 2x3 system. hope this helps. Similar Discussions: How do I show two planes are neither coincident, or parallel?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8017929792404175, "perplexity": 392.639637983919}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187828411.81/warc/CC-MAIN-20171024105736-20171024125736-00262.warc.gz"}
http://math.stackexchange.com/questions/19347/how-do-you-generalize-the-compression-of-data-into-n-dimensional-space
How do you generalize the compression of data into N dimensional space? I have a theory that in order to best compress 1D data between 0 and 1 in a lossy way, you should map each data point $i$ to: $\int_0^i \! P(x) \, \mathrm{d}x$ Where $P(x)$ is the probability density function of the data. The probability density function over the mapped data should be level and being level when you store the values to a finite precision, the amount of redundant information should be minimized. My question is: how can I generalize this to more dimensions? Say I have a set of points which fit a given distribution and I want to compress points that fit the same distribution. My initial assessment is that I should incrementally move the points so that they follow a level probability density function, then (with lots of arm waving) it should be possible to traverse the mapping using the proximity to the reference points. A further constraint is that given data that fits something that looks like a spiral, rather than squashing it together, it should unwind it. Can anyone shed any light on this? • I assume that I'm not the first person to talk about such a concept. What is its name? • Is this a good approach? • How do you incrementally move the points so that they fit a level probability density function? • How do you traverse the mapping? • How do I modify the above approach to fit the added constraint? edit: To clarify, essentially what I want to do is take a set of points, then warp the space the points are in so that the points are equally spaced. - The map in your first paragraph is called the probability integral transform. – Rahul Narain Jan 28 '11 at 10:05 @Rahul, Thanks. – dan_waterworth Jan 28 '11 at 10:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8567156195640564, "perplexity": 265.31023902580984}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368705956734/warc/CC-MAIN-20130516120556-00045-ip-10-60-113-184.ec2.internal.warc.gz"}
https://forum.allaboutcircuits.com/threads/i2c-not-working.118974/
i2c not working kingsloth Joined Dec 28, 2015 12 i'm trying to read values with the following routine C: void read() { unsigned int lsb = 0; unsigned char msb = 0; char str[20]; if (I2CBusIsIdle(I2C1)) { I2CStart(I2C1); } I2CSendByte(I2C1, REG_WRITE); IdleI2C1(); I2CSendByte(I2C1, DATA_OUTPUT_X); IdleI2C1(); I2CRepeatStart(I2C1); IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); lsb = I2CGetByte(I2C1); } IdleI2C1(); X_axis=((msb<<8) | lsb); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); lsb = I2CGetByte(I2C1); } IdleI2C1(); Z_axis=((msb<<8) | lsb); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, FALSE); lsb = I2CGetByte(I2C1); } IdleI2C1(); Y_axis=((msb<<8) | lsb); I2CStop(I2C1); } I always get 0.0, what am i doing wrong? I have SDA on pin RG3 and SCL on RG2, both with a resistence of 4k7 to 3.3V. Moderators note : used code tags for C Last edited: sailorjoe Joined Jun 4, 2013 361 What i2c library are you using, please? Since I don't have the library information, I'm just guessing here. At line 24, consider what happens if no data is available. Does the i2c function I2CReceivedDataIsAvailable() wait for data, or return false if data hasn't arrived yet? spinnaker Joined Oct 29, 2009 7,837 i'm trying to read values from hmc5833l with the following routine C: void read() { unsigned int lsb = 0; unsigned char msb = 0; char str[20]; if (I2CBusIsIdle(I2C1)) { I2CStart(I2C1); } I2CSendByte(I2C1, REG_WRITE); IdleI2C1(); I2CSendByte(I2C1, DATA_OUTPUT_X); IdleI2C1(); I2CRepeatStart(I2C1); IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); lsb = I2CGetByte(I2C1); } IdleI2C1(); X_axis=((msb<<8) | lsb); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); lsb = I2CGetByte(I2C1); } IdleI2C1(); Z_axis=((msb<<8) | lsb); IdleI2C1(); { I2CAcknowledgeByte(I2C1, TRUE); msb = I2CGetByte(I2C1); } IdleI2C1(); IdleI2C1(); { I2CAcknowledgeByte(I2C1, FALSE); lsb = I2CGetByte(I2C1); } IdleI2C1(); Y_axis=((msb<<8) | lsb); I2CStop(I2C1); } I always get 0.0, what am i doing wrong? I have SDA on pin RG3 and SCL on RG2, both with a resistence of 4k7 to 3.3V. IF i use pic32 i still need to read to msb and lsb and then do the shift?(i just followed what i read on hmc data sheet) Moderators note : used code tags for C Have you scoped the output of the mcu to see if you are even putting out data? kingsloth Joined Dec 28, 2015 12 What i2c library are you using, please? Since I don't have the library information, I'm just guessing here. At line 24, consider what happens if no data is available. Does the i2c function I2CReceivedDataIsAvailable() wait for data, or return false if data hasn't arrived yet? It only returns true or false. Last edited: kingsloth Joined Dec 28, 2015 12 Have you scoped the output of the mcu to see if you are even putting out data? This is the only thing i get, ch1 - SCL ch2- SDA. RRITESH KAKKAR Joined Jun 29, 2010 2,829 From where you got the library for MPlab kingsloth Joined Dec 28, 2015 12 From where you got the library for MPlab don't know if i'm doing it right or not but i just pressed F1 (help) and there was a bunch of routines to use i2c, i just read them and use them. It compiled so i thought i was doing good. I'm trying to re-write the code making my own i2c functions, maybe it will work... RRITESH KAKKAR Joined Jun 29, 2010 2,829 I have done but the code was not mine. spinnaker Joined Oct 29, 2009 7,837 I'm using MPLAB X so i'm just using their library. It only returns true or false. MPLabX is not a compiler it is a development enviroment.. What compiler are you using? It isn't XC8 because the documentation does not match your calls unless you are using your own wrapper around the XC8 calls. So which is it? spinnaker Joined Oct 29, 2009 7,837 This is the only thing i get, ch1 - SCL ch2- SDA. This is good. At least something is coming out. Can you see if you are getting a response from the slave? You don't happen to have a logic analyzer? They can be a huge hel in situations like this. Next thing to check after that is all of your settings are the way you need them RRITESH KAKKAR Joined Jun 29, 2010 2,829 May be compiler Hi Tech C ErnieM Joined Apr 24, 2011 8,019 The scope picture looks perfect for the address to write to this device, and since the 9th bit is zero the slave is acknowledging it is being addressed. Doesn't get any better than that. The next step is to grab an entire write sequence and then the read sequence. See if things keep jetting that nice ACK bit set. And don't forget when reading multiple bytes your master pic must do the ACK, your lib has that function. Getting these low level transactions to all play together is a major piece of my debugging time. The good news is you have a scope as a reality check. I would strongly suggest you find some example code on the interwebs where someone shows the sequence that worked for them to bang this device. I am not sure how I would do it after peeking at the operational examples of the device data sheet. spinnaker Joined Oct 29, 2009 7,837 The scope picture looks perfect for the address to write to this device, and since the 9th bit is zero the slave is acknowledging it is being addressed. Doesn't get any better than that. The next step is to grab an entire write sequence and then the read sequence. See if things keep jetting that nice ACK bit set. And don't forget when reading multiple bytes your master pic must do the ACK, your lib has that function. Getting these low level transactions to all play together is a major piece of my debugging time. The good news is you have a scope as a reality check. I would strongly suggest you find some example code on the interwebs where someone shows the sequence that worked for them to bang this device. I am not sure how I would do it after peeking at the operational examples of the device data sheet. Wow that is where experience pays off! The XC8 manual has examples. But as I mention this is not XC8 unless is it is a wrapper. What confuses me id there is no configuration in the code as posted. kingsloth Joined Dec 28, 2015 12 MPLabX is not a compiler it is a development enviroment.. What compiler are you using? It isn't XC8 because the documentation does not match your calls unless you are using your own wrapper around the XC8 calls. So which is it? I'm using XC32 dannyf Joined Sep 13, 2015 2,197 what am i doing wrong? You didn't read the PLIB manual / example code. kingsloth Joined Dec 28, 2015 12 Thing is the only thing i got from the scope was that, nothing more, so probably it's only writing to device, never reads it spinnaker Joined Oct 29, 2009 7,837 I'm using XC32 I cannot imagine the xc32 peripheral library calls are that much different than XC8. If they are similar then what you posted is not from the perphial library. ErnieM Joined Apr 24, 2011 8,019 Oh bleep, spinnaker talks me up then I have to take it back. Most of the transmission looks ok, but the address should begin with a START transaction, which is SCL hi while SDA goes low. If you only see this pattern over and over the sequence never sends a STOP signal. I am typing on a silly iPad and can't copy and paste code. If I was doing this I would forget all that code you have and begin with a simple config command, which btw I do not see in your code, do you initial the device? One trick is to use a spare I/O pin as a flag, set it hi when you start a transaction and set lo when done. This gives you a solid signal to trigger your (very nice) scope. kingsloth Joined Dec 28, 2015 12 Oh bleep, spinnaker talks me up then I have to take it back. Most of the transmission looks ok, but the address should begin with a START transaction, which is SCL hi while SDA goes low. If you only see this pattern over and over the sequence never sends a STOP signal. I am typing on a silly iPad and can't copy and paste code. If I was doing this I would forget all that code you have and begin with a simple config command, which btw I do not see in your code, do you initial the device? One trick is to use a spare I/O pin as a flag, set it hi when you start a transaction and set lo when done. This gives you a solid signal to trigger your (very nice) scope. yeah im doing: I2CSetFrequency(I2C1, FPB, FSCK); // it gives 100kHz, thats right I2CEnable(I2C1, TRUE); To me looks like the problem is I2CStop(I2C1) isn't working, dont know how to solve it though. ErnieM Joined Apr 24, 2011 8,019 yeah im doing: I2CSetFrequency(I2C1, FPB, FSCK); // it gives 100kHz, thats right Without more of your code no one could tell what that statement could possibly mean, though I do hope you enable the I2C before you attempt to use it.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.21686582267284393, "perplexity": 3098.574920506179}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370518622.65/warc/CC-MAIN-20200403190006-20200403220006-00072.warc.gz"}