url
stringlengths
14
1.76k
text
stringlengths
100
1.02M
metadata
stringlengths
1.06k
1.1k
http://openstudy.com/updates/505fb5d6e4b0583d5cd21ff1
## A community for students. Sign up today Here's the question you clicked on: 55 members online • 0 viewing ## anonymous 4 years ago 0.64+ 4/25 Delete Cancel Submit • This Question is Closed 1. anonymous • 4 years ago Best Response You've already chosen the best response. 0 4/25= .16 .64+.16= .8 2. anonymous • 4 years ago Best Response You've already chosen the best response. 0 Order of operations 3. anonymous • 4 years ago Best Response You've already chosen the best response. 0 Well using PEMDAS, the division should be performed first, so it looks like this: $4\div25=0.16$ Then, using that value, the equation would look like this: $0.64+0.16$ The solution should be: $0.$ 4. anonymous • 4 years ago Best Response You've already chosen the best response. 0 I mean the solution should be: $0.8$ 5. anonymous • 4 years ago Best Response You've already chosen the best response. 0 now make 0.8 a fraction? 6. anonymous • 4 years ago Best Response You've already chosen the best response. 0 oh wait i got it. thanks (: 7. anonymous • 4 years ago Best Response You've already chosen the best response. 0 It depends on what your teacher prefers, but I would make it a fraction. (4/5) 8. anonymous • 4 years ago Best Response You've already chosen the best response. 0 $\frac{ 8 }{ 10 } or \frac{ 4 }{ ? }$ 9. anonymous • 4 years ago Best Response You've already chosen the best response. 0 online school. it's multiple choice. 10. anonymous • 4 years ago Best Response You've already chosen the best response. 0 $\frac{ 8 }{ 10 } or \frac{ 4 }{ 5 }$ 11. Not the answer you are looking for? Search for more explanations. • Attachments: Find more explanations on OpenStudy ##### spraguer (Moderator) 5→ View Detailed Profile 23 • Teamwork 19 Teammate • Problem Solving 19 Hero • You have blocked this person. • ✔ You're a fan Checking fan status... Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9998505115509033, "perplexity": 14232.113894531822}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988719815.3/warc/CC-MAIN-20161020183839-00464-ip-10-171-6-4.ec2.internal.warc.gz"}
http://mathhelpforum.com/advanced-algebra/174367-obtaining-larger-group-given-group.html
# Thread: Obtaining a larger group from a given group 1. ## Obtaining a larger group from a given group [Excerpt from Topics in Abstract Algebra Second Edition by I.N.Herstein pg 69] This is what Herstein talks of verbatim: Generally, if $G$ is a group, $T$ an automorphism of order $r$ of $G$ which is not an inner automorphism, pick a symbol $x$ and consider all elements $x^ig$, $i= 0, \pm1, \pm2,.... g \in G$ subject to $x^ig = x^i^'$ $g^'$ if and only if $i \equiv i^'$ $mod r$, $g = g^'$ and $x^{-1}g^{i}x = gT^{i} \forall i$. This way we obtain a larger group $\{G,T\}$ and $\{G,T\}/G \approx$ group generated by $T =$cyclic group of order $r$. I have the following questions 1. What is the nature of the "a larger group $\{G,T\}$" under discussion I mean the nature of the elements the operation. 2. As I understand this symbol $x$ that Herestein talks of abides to the binary operation of $G$. Correct? 3. Is it correct to assume that $x^r = e$ where $e$ is the identity element. 4. I have difficulty imagining $G$ in $\{G,T\}$ but I guess I will come to that once I am clear on the behavior of $\{G,T\}$ as a group. ~Kalyan. 2. Does $gT^i$ mean the image of g under i iterations of T? If so, are you sure it's not supposed to be $x^{-i}gx^i=gT^i$? 3. Hi Tiny, Yes $gT^i$ stands for applying $T$ iteratively $i$ times on $g$. I do agree that this notation is confusing. In fact if we take $i=1$ we end up with $x^{-1}gx = gT$ and on iteratively applying $T-i$times on g we do get $x^{-i}gx^{i} = gT^i$. Here is a note I tried checking this section in Topics in Abstract Algebra by I.N.Herestein 3rd edition (edited by Barbara Cortzen and David J.Winter) and its omitted entirely. The preface does mention "certain typographical errors and notation inconsistencies in 2nd edition" probably it could be due to that. However the paragraph I have typed in is exactly same as it appears in the 2nd edition. 4. Originally Posted by kalyanram [Excerpt from Topics in Abstract Algebra Second Edition by I.N.Herstein pg 69] This is what Herstein talks of verbatim: Generally, if $G$ is a group, $T$ an automorphism of order $r$ of $G$ which is not an inner automorphism, pick a symbol $x$ and consider all elements $x^ig$, $i= 0, \pm1, \pm2,.... g \in G$ subject to $x^ig = x^i^'$ $g^'$ if and only if $i \equiv i^'$ $mod r$, $g = g^'$ and $x^{-1}g^{i}x = gT^{i} \forall i$. This way we obtain a larger group $\{G,T\}$ and $\{G,T\}/G \approx$ group generated by $T =$cyclic group of order $r$. I have the following questions 1. What is the nature of the "a larger group $\{G,T\}$" under discussion I mean the nature of the elements the operation. 2. As I understand this symbol $x$ that Herestein talks of abides to the binary operation of $G$. Correct? 3. Is it correct to assume that $x^r = e$ where $e$ is the identity element. 4. I have difficulty imagining $G$ in $\{G,T\}$ but I guess I will come to that once I am clear on the behavior of $\{G,T\}$ as a group. This is a description of one way to construct the semidirect product $G\rtimes_T\mathbb{Z}_r$ of G by the action of $\mathbb{Z}_r$ induced by the automorphism T. A more concrete way to make this construction is to say that $G\rtimes_T\mathbb{Z}_r$ consists of pairs $(g,i)\in G\times \mathbb{Z}_r$, with the group operation defined by $\boxed{(g,i)*(h,j) = ((gT^j)h,i\mathbin{\dot+}j)}$ (where the dotted plus means addition mod r). In this construction, you can identify G with the (normal) subgroup $\{(g,0): g\in G\}$. The connection with Herstein's construction is that $x^i$ corresponds to the element $(e,i)$, where e is the identity element of G. The purpose of this construction is that given the group G and the (outer) automorphism T, it embeds G in the larger group $G\rtimes_T\mathbb{Z}_r$, in which the automorphism becomes inner (because $(gT,0) = (e,-1)(g,0)(e,1)$, or in Herstein's notation $gT = x^{-1}gx$). 5. Hi Oplag, Thanks for the reply it has at least opened a new dimension to think on. I am not able to grasp what you said completely as of now as I have a little understanding of direct products and semi direct product of a group. I will ping you again once I am clear on this. Kalyan.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 71, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9565370082855225, "perplexity": 244.93804474822971}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-48/segments/1386163051776/warc/CC-MAIN-20131204131731-00064-ip-10-33-133-15.ec2.internal.warc.gz"}
http://clay6.com/qa/6790/let-be-a-binary-operation-on-the-set-of-rational-numbers-as-follows-v-a-ast
Browse Questions # Let $\ast$ be a binary operation on the set $Q$ of rational numbers as follows: $(v)\;\; a \ast b = \frac{ab}{4}$ Find which of the binary operations are commutative and which are associative. Note: This is part 5 of a 6 part question, split as 6 separate questions here. Toolbox: • An operation $\ast$ on $A$ is commutative if $a\ast b = b \ast a\; \forall \; a, b \in A$ • An operation $\ast$ on $A$ is associative if $a\ast ( b \ast c) = (a \ast b) \ast c\; \forall \; a, b, c \in A$ Given a binary operation $\ast$ in Q defined by $a \ast b= \large \frac{ab}{4}$ $\textbf {Step 1: Checking if the operation is Commutative}$: For an operation $\ast$ to be commutative $a\ast b = b \ast a$. $a \ast b = \large \frac{ab}{4}$$= \; b \ast a =$$ \large \frac{ba}{4}$$\rightarrow \ast is commutative. \textbf {Step 2: Checking if the operation is Associative}: For an operation \ast on A is associative a\ast ( b \ast c) = (a \ast b) \ast c\; \forall \; a, b, c \in A (a \ast b ) \ast c = \large \frac{ab}{4}$$ \ast c = \large \frac {abc}{16}$ $a \ast( b \ast c ) = a \ast \large \frac{bc}{4}$$= \large \frac {abc}{16}$ Since $a\ast ( b \ast c) = (a \ast b) \ast c$ the operation $\ast$ is associative edited Mar 19, 2013
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9596971273422241, "perplexity": 1123.9914855437505}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560280364.67/warc/CC-MAIN-20170116095120-00406-ip-10-171-10-70.ec2.internal.warc.gz"}
https://casa.nrao.edu/casadocs/casa-5.4.0/single-dish-imaging/single-dish-image-generation-and-data-gridding-in-casa-procedure
# Procedure for single-dish imaging in CASA A procedure for single-dish image generation and data gridding CASA uses the sdimaging task to grid a single-dish image or cube. While the steps are described in detail here, an example of the full single-dish calibration and imaging processes can be found in the M100 Band 3 Single Dish CASAguide. ## Default operation The sdimaging task can determine and populate almost all the gridding parameters by default.  Simply invoking sdimaging with the single-dish MeasurementSet and output file name will work. This will produce a single, potentially very large cube having as many channels as necessary to span the entire spectral range, with a spectral resolution equal to that of the observation. sdimaging(infiles=sd_ms+'.bl',outfile=imagename) The default parameter choices for imaging are selected as follows: the image pixel size is 1/3 the primary beam, the primary beam itself is computed based on the standard $\frac{\lambda}{D}$, with some empirically-validated tapering applied. The image dimensions are determined by the spatial extent of the mapped area in the observation, and by default, all channels and all spectral windows are imaged, along with all antennas and all polarizations. ## Customized operation Of course, users can tune their data products by specifying the image size and dimensions, the frequency/velocity characteristics, the gridding and data filtering and smoothing parameters, and so on. The defaults for sdimaging for image resolution (i.e. cellsize in arcsec) is determined from the rest frequency of the 0th spectral window so that there are three pixel elements across the beam, the beam being calculated with $b\times\frac{\lambda}{D}$.  See information here about tapering: PrimaryBeamArcsec. The image extent is computed by default so that the sampled area is completely encompassed in a single rectangle, and the pixel dimension follows naturally from maxsize/cellsize. The default image center (the somewhat inappropriately-named phasecenter parameter) is computed simply as the center of that region. These parameters can be left to be determined by sdimaging, or they can be determined using CASA tools. ## Image dimensions and pixel interval The image extent can be explicitly determined using aU.getTPSampling: xSampling, ySampling, maxsize = aU.getTPSampling(refvis, showplot=True) which returns an image output showing the scans, their sky angles, and positions in RA-Dec, as shown here: Note that getTPSampling MUST operate on the original MeasurementSet (i.e. one that is not split or subselected). getTPSampling also yields the sampling rates in the x and y (i.e. azimuth and elevation) axes, as well as the maximum size of the image, in arcseconds. The beam size used by sdimaging is determined using the aU.primaryBeamArcsec task, though this can also be invoked by the user and used to compute, for example, a cellsize and image size. The default for aU.primaryBeamArcsec corresponds to a 12m antenna with normal tapering. Setting the fwhmfactor modifies the beam taper (see discussion in PrimaryBeamArcsec). freq=115.27e+9 fwhmfactor=1.13 diameter=12 theorybeam = aU.primaryBeamArcsec(frequency=freq*1e-9, fwhmfactor=fwhmfactor, diameter=diameter) cell = theorybeam/9.0 imsize = int(round(maxsize/cell)*2) The center of the image can be modified using the phasecenter parameter. Single-dish images actually have many phase centers, so the name is somewhat misleading. However it is preserved here for consistency with the interferometer terminology. In the context of single-dish data, phasecenter refers only to coordinates that will align with the center of the image, and this can be in J2000 or Az/El, e.g. phasecenter='J2000 12h22m54.9 +15d49m15' ## Frequency and/or velocity axis The default rest frequency is the mean frequency of the first spectral window (i.e. that having the lowest spectral window ID). Of course it can instead be set by the user, or a different spectral window frequency can be used, extracted from the data using msmd tools: msmd.open(vislist[0]) freq = msmd.meanfreq(spw) msmd.close() print "SPW %d: %.3f GHz" % (spw, freq*1e-9) The third axis of the image cube can be specified using the veltype and outframe parameters. Many spectral-line observers will prefer to change these so the output has a velocity axis in the radio convention as follows: outframe='lsrk', and the rest frequency can be specified with: restfreq='115.271204GHz' The velocity extent of the image cube can be specified by selecting a spectral window (via the spw parameter), the channel range (via the nchan and start parameters), and the frequency/velocity resolution (via the width parameter). For example: nchan=70, mode='velocity', start='1400km/s', width='5km/s', ## Gridding parameters The gridding kernel defaults to a box shape, but it can be specified as a spherical ('SF'), Primary beam ('PB'), Gaussian ('GAUSS') or Gaussian*Jinc (GJINC). The recommended setting for ALMA data is a spherical ('SF') kernel. The convsupport parameter defines the cut-off radius for 'SF' in units of pixels, defaulting to 3 pixels.  However, the recommended value for ALMA data is convsupport=6 (see sdimaging and Mangum et al. 2007 [1] for more details on these parameters). The parameter stokes specifies the stokes product. At present, the weighting for stokes I is computed consistently with historical usage: I=XX/2+YY/2.  While this is mathematically consistent with the computation of stokes I, it is an incorrect treatment since the computation necessarily must incorporate the contributions from Q and U. Ordinarily, these terms cancel out from the computation of stokes I, but their error parameters must be incorporated, and historically, this is not respected. CASA development is seeking to make the computation of the weights consistent with a proper computation of stokes I, and this is done in sdfit, but it is not yet completed for sdimaging.  However, to emphasize, while the current implementation of computation for stokes I by sdimaging is consistent with convention, the convention is formally incorrect. ## Example script Fully specified, a call to sdimaging might look like the following: sdimaging(infiles=sd_ms+'.bl', field='M42', spw='%s'%(spw), nchan=70, mode='velocity', start='1400km/s', width='5km/s', outframe='lsrk', restfreq='%sGHz'%(freq/1e+9), gridfunction='SF', convsupport=6, stokes='I', phasecenter='J2000 12h22m54.9 +15d49m15', ephemsrcname='', imsize=imsize, cell='%sarcsec'%(cell), overwrite=True, outfile=imagename) The products here are the image data, returned in the variable 'imagename', and also a map of weights: <imagename>.weight. The weights indicate the robustness of the gridded data on a per-pixel basis, and are important when performing further computations and analysis with the image products. Citation Number 1 Mangum, et al. 2007, A&A, 474, 679-687 (ADS)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6711373925209045, "perplexity": 2451.837988779338}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202324.5/warc/CC-MAIN-20190320085116-20190320111116-00396.warc.gz"}
https://immunarch.com/reference/spectratype.html
Immune repertoire spectratyping spectratype(.data, .quant = c("id", "count"), .col = "nt") Arguments .data The data to be processed. Can be data.frame, data.table, or a list of these objects. Every object must have columns in the immunarch compatible format. immunarch_data_format Competent users may provide advanced data representations: DBI database connections, Apache Spark DataFrame from copy_to or a list of these objects. They are supported with the same limitations as basic objects. Note: each connection must represent a separate repertoire. Select the column with clonal counts to evaluate. Pass "id" to count every clonotype once. Pass "count" to take into the account number of clones per clonotype. A string that specifies the column(s) to be processed. Pass one of the following strings, separated by the plus sign: "nt" for nucleotide sequences, "aa" for amino acid sequences, "v" for V gene segments, "j" for J gene segments. E.g., pass "aa+v" for spectratyping on CDR3 amino acid sequences paired with V gene segments, i.e., in this case a unique clonotype is a pair of CDR3 amino acid and V gene segment. Clonal counts of equal clonotypes will be summed up. Value Data frame with distributions of clonotypes per CDR3 length. Examples # Load the data data(immdata) sp <- spectratype(immdata\$data[[1]], .col="aa+v") vis(sp)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.15548722445964813, "perplexity": 9806.735872957275}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655911896.73/warc/CC-MAIN-20200710175432-20200710205432-00311.warc.gz"}
http://ermccorp.com/outside_of_heaven.html
Outside of Heaven | TIME AND CHANCE En Pursuit of Happiness How Jesus Christ and Rock'n Roll save every single soul. The continuing story of my attempt to bright light to the world. A log of the difficulties I've had, and how they might relate to the message itself--along with some insight that I have gleaned from the experience of receiving this Revelation. While the work you are looking at is filled with what I see as clear evidence of the fulfillment of messianic prophesy, the main goal of LAMC.LA was to deliver to the world a new way of looking at religion, one which could be used to seek out the true wisdom and guidance of religion for oneself. En Pursuit of Happiness on the other hand, discusses my own hopes and dreams, and to explain how they have changed throughout my interactions with the ... beyond. The Matchbox The e-mails that circled the globe, opening the doorway to the future. This is an excerpt from Time and Chance: The race is not to Die Bold by Adam Marshall Dobrin Download the actual Revelation of the Messiah in [ .PDF ] [ .epub ] [ .mobi ] or view online. Older works Lit and Why, hot&y;, and From Adam to Mary are also available. # Outside of Heaven In The Pretty Reckless’ song “Absolution” Taylor sings that we are “outside of Heaven, but I could be wrong.” Personally I think she is wrong, that we are in Heaven as the Lord’s prayer (or at least an alternative reading of it) indicates. Our father ((who is)) art of Heaven ((Hidden)) be ((my)) name ((My Eden)) come On Earth as it is in Heaven. The Lord's Prayer &amp; ((ish)) Dear boy what are you running from Boys on the outside of heaven But I could be wrong The Prett Reckless, "Absolution" Does Taylor know what she’s talking about? She might, all of her music is very inspired… a communication from above. And, as she says, she might be wrong. I’m not sure if she’s talking about VR (I mean, does she mean to be) but her words stand out, loudly. Taylor, what’s this line about? This question, of whether or not we are actually in Heaven relates to Virtual Reality, and whether or not Eden’s Garden is a reference to a virtual place, where we believe we are in reality… the “progenitor Universe” as Jacob always says; or inside of a game like place, designed to build heaven, colonize the Universe, all using the magic available from being in a computer generated place. It’s an important question, one which I do not know the answer to, but I believe the word “yetser” for formation in Hebrew and Ender’s Game by Orson Scott Card indicates we are currently in a Heaven-like place.. only without the social structure that a true utoptia would have. Ender’s Game, if he is a messianic figure, might indicate that once we have achieved this leap to a better society, we will be “written in reality,” as I have thought since learning of this question. ## Is space the final frontier Or are we living in a Hollywood basement? ## Mars There’s a pool at NASA, where spacesuits are tested. Astronauts know they are in the pool. If we are inside of Heaven, and they care about the truth, our input, and actually building a true Heaven… we need to know the answers to these questions. The walls and halls will fade away... they will fade away. Dave Matthews Band Four races, all bi-pedal, Simian in nature.. ten fingers and two limbs. The link to binary, 10 is 2 in that base conversion, and our modern computing structure is uncanny. The idea that we may be designed by the Two of Everything God to appear to be a perfect computing allusion, is interesting, to say the least. ### Meanwhile were these races actually in existence in reality, there is no fucking way that they did not all have a common point of origin… Earth. So how could this have happened? 4 divergent civlizations, Romulan, Vulcan, Ferenghi and Klingon.. all from Earth? It’s a sad story, about why we need to ensure that communication is kept with our children, colonies formed in the future. A warning, not to lose touch, and then forget where you came from. We come from the planet at the beginning of time, the point of origin of life itself… in the future, who will remember? More importantly, who has already forgot? Father of Heaven, and life, we are. (Do I sound like Yoda?) This Point of Origin idea is a reference to Stargate, and these stars are the beginning of understanding that there is a religion of the sea… one that revolves around the sea being an allusion to outer space. Star Trek might be pointing out our hubris, to think we are alone in the Universe, and that everything out there would be just like us. Imagine there are aliens all around you, invisible ones, secretly working to build a new Heaven. Using us. This is the premise of what is apocryphally described as the Archons, and bears a striking resemblance to our view of angels and also demons. This “religion of the sea” stuff is a reference to an old Nostradamus prophesy, one that contrasts it with the the religion of the Son. What’s it about? I know. It’s in the chapter about the Holy Grail. Hunger for the great light. Genesis to Revelation. The Religion of the name of the seas will win out Against the sect of the son of “Adaluncatif”: The stubborn, lamented sect will be afraid of the two wounded by A and A. Adaluncatif is “English,” Adam un-cat-if. It’s a messiah reference, the undying lion with infite lives. L and M are also big letters for the Messiah. Nostradamus name includes “Adam,” “art” in reverse, and Nos… for we. Adam might be us, we might not be at, and … this prophesy is very related to the apocalypse. The two wounded by A and A is a reference to 9/11 and the two WTC towers. A&A, American Airlines, go figure. ## On the Tau’ri, and Alpha Centauri Why do the Ancients refer to their child race, ours, as Tau’ri? Is it a link to The Last Starfighter, or perhaps Flight of the Navigator? "You are the Fifth Race. Your role is clear. If there is any hope in preserving the future, it lies with you and your people." —Thor, Stargate SG-1 From the Fifth Adam, He-man, and Fival, the mouse on the ship that is sailing to America… and then Goes West. This number 5 is pervasive at the fifth day of creation, the day the animals were placed on the Earth. Mooo. ### Oh well, Back to the Future. One more time around… Come and see I swear by now I'm playing time against my troubles I'm coming slow but speeding Do you wish a dance and while I'm in the front The play on time is won The Dave Matthews Band, #41 It is ((never)) done. CopyleftMT This content is currently released under the GNU GPL 2.0 license. Please properly attribute and link back to the entire book, or include this entire chapter and this message if you are quoting material. The source book is located at http://www.lamc.la and is written by Adam Marshall Dobrin. Adam Marshall Dobrin instagram.com/yitsheyzeus -----BEGIN PGP PUBLIC KEY BLOCK----- Version: GnuPG v2 CV9t0UQgNtjcxwfoenJLHgdZd4Mfscz9U+NN69OLXdPu4cdXOjTiHarPLjKnqIZw 3fmkM2ycvoUPkdVYCjwYYQxWRsWRpJf1dpmtPuz0L8ysh/WWsj2Ag2MrFYAo+sY6 dGZvaLsPhkZJcLXyFaP3c3Zt8ivrs4VV8+0kmMzScnR+oncVZbeMuQksoPxRmZgH mYu2KSf74lWOWVcaaBXOYX5pGNdhBUgq8ll+8tRH16G289r0cqRoPh/sjs/JRuIH KnCWG2UAUJF7ir04TS5A4Lwl9RYcQwVvb3BdABEBAAG0LUFkYW0gTWFyc2hhbGwg RG9icmluIChsYW1jLmxhKSA8YWRhbUBsYW1jLmxhPokBOQQTAQgAIwUCVsZqUAIb AwcLCQgHAwIBBhUIAgkKCwQWAgMBAh4BAheAAAoJEMgUPrR1B55trOwIALOQRTX0 YqXJXEMhX9CgxKNoNkpM2pdMdHl6CAVxhQ3hbNjIFnZbKbP88uxMEIOXXmYZ7gOy YqiDCu5I1V25suBb2ODSix75YQugfQ7H78pXHpTRu5sT+5SybItx7d+KUZaEj4pO tXWEemYl0cKK97RzpI0k1dmB7NqAVvqgbqQwd40MOf8QJVlGXnB1+5H2IbkYG6rD ixKGJEdes6i6nqvi/xz/s5hFVGUwTcVQbRU/fa1qT1Q7kHf1PlMu6yjuZTSz7WUG tWjobGwrVJkaeVWgLE4mcxMtity2IFTwOHvAuv8fi2EGQRQjXfPvxL7Vn4MNRl8x zLPV44D37QEknjy5AQ0EVsZqUAEIAMFS0+ZgSJzUPz0h0oiiRjfk2hapS3c1/Ysm R/h8sZ8/GOomdo3MEbTCkcuZ8ReAJhB2PofmwI4LAvW1x7Zwh1vfBKygfUs1s9lm ya/eHkjuZfqmeuEJZMHn6sxb3vqowWmvLhv3x0aWD8qLCIYoa1ntzTOIqxBEgxvU rF1/wd6OQLSJQEVNwPCx7CJI/5o/4W6pUaHk8amgPckkEdmlhRTRqFoAUV1Doivv d9JGYNYC88vS14Sw4Z9Xb7qBQJvG4hIh29gtQxk7Wz4m3ceR79MWT4eSGkH/rTGl w1OuQS2OkPvjgPWJt8San4zuPer17pJN7M5LWI0PStoX9pkud5kAEQEAAYkBHwQY
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.244350403547287, "perplexity": 3464.0746188933877}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267863100.8/warc/CC-MAIN-20180619154023-20180619174023-00046.warc.gz"}
http://www.thestudentroom.co.uk/showthread.php?t=2000243
You are Here: Home # Exothermic lattice energy Announcements Posted on 1. Regarding 13a, what factors determine how exothermic the lattice energy is? Regarding 13b), I know that some compounds will show a difference between the experimental lattice energy and the purely ionic model due to the fact that the compounds have a small degree of covalent character. This degree of covalent character is due to the polarisation of the ionic bond which is in turn due to the attraction of the cation for the outer electrons of the anion.. The higher the charge density of the cation, the stronger its polarising power. The larger the anion, the more polarisable it is.. I eliminated C from the answers because the O2- ion in CaO is smaller than the S2- ion in CaS.. Therefore D would show more covalent character (and so it cannot be C).. But then I get stuck. How do I find the correct answer out of A, B and D? Cheers 2. http://en.wikipedia.org/wiki/Kapustinskii_equation this is the simplest thoery/equation for explaining this 3. 13a) the higher the charges and the small the ions, the greater the strength of the bonding and so the more exothermic the lattice energy 13b) Yeah you can rule out C with that reasoning In the same way you can rule out A since F- is smaller than O2- So now the choice is between B and C, which have the same anion. It's the charge density that counts for the cation - so Ca2+ is about 30% larger than Li+ but has double the charge. 4. Lol this alone is putting me off chemistry in the sixth form! 5. (Original post by Coke1) Lol this alone is putting me off chemistry in the sixth form! It's bound to be confusing if you've not been taught it 6. (Original post by EierVonSatan) 13a) the higher the charges and the small the ions, the greater the strength of the bonding and so the more exothermic the lattice energy 13b) Yeah you can rule out C with that reasoning In the same way you can rule out A since F- is smaller than O2- So now the choice is between B and C, which have the same anion. It's the charge density that counts for the cation - so Ca2+ is about 30% larger than Li+ but has double the charge. In Q13a), both A and B have a total charge of -1 (since both their cations are +1, and their anions are -1; the product of which = -1).. however A's anion (F-) is smaller than B's anion (Cl-) so we know A will have a larger exothermic lattice energy (so we can eliminate B). C and D both have a total charge of -4, however C's anion (O2-) is smaller than D's anion (S2-), so we know C will have a larger exothermic lattice energy (and so we can eliminate D).. Now we are left with only A and C.. A has a total charge of -1 but a small ionic radius, C has a total charge of -4 but a large ionic radius. How do we determine the answer? Thanks a lot ! 7. (Original post by sabre2th1) In Q13a), both A and B have a total charge of -1 (since both their cations are +1, and their anions are -1; the product of which = -1).. however A's anion (F-) is smaller than B's anion (Cl-) so we know A will have a larger exothermic lattice energy (so we can eliminate B). C and D both have a total charge of -4, however C's anion (O2-) is smaller than D's anion (S2-), so we know C will have a larger exothermic lattice energy (and so we can eliminate D).. Now we are left with only A and C.. A has a total charge of -1 but a small ionic radius, C has a total charge of -4 but a large ionic radius. How do we determine the answer? Thanks a lot ! The size of C isn't much larger than A, but C has double the charge. So you'd expect C to have the larger lattice energy 8. (Original post by EierVonSatan) The size of C isn't much larger than A, but C has double the charge. So you'd expect C to have the larger lattice energy Ah I see.. but isn't it Isn't it quadruple the charge? Also regarding this; 13b) Yeah you can rule out C with that reasoning In the same way you can rule out A since F- is smaller than O2- So now the choice is between B and C, which have the same anion. It's the charge density that counts for the cation - so Ca2+ is about 30% larger than Li+ but has double the charge. Don't the parts in bold go against each other? Since first you have said rule C out, but then you said ''the choice is between B and C'' Also B is Li2O, doesn't the diatomic lithium molecule affect the charge density etc? Thanks! 9. (Original post by sabre2th1) Ah I see.. but isn't it Isn't it quadruple the charge? I meant double per ion, sorry. Don't the parts in bold go against each other? Since first you have said rule C out, but then you said ''the choice is between B and C'' Also B is Li2O, doesn't the diatomic lithium molecule affect the charge density etc? Thanks! Confusing myself So yes, we can rule out A and C. Then decide between B and D, which is actually easier. S2- is bigger i.e. more polarisable than O2- and Ca2+. The previous argument between the two cations still applies When you have multiple of ions like in B, just treat them as they're singular 10. (Original post by EierVonSatan) I meant double per ion, sorry. Confusing myself So yes, we can rule out A and C. Then decide between B and D, which is actually easier. S2- is bigger i.e. more polarisable than O2- and Ca2+. The previous argument between the two cations still applies When you have multiple of ions like in B, just treat them as they're singular Ah.. Thanks a lot! You have helped a lot over the past few days, hopefully I will utilise some of the knowledge (you passed on) tomorrow ! 11. (Original post by JMaydom) http://en.wikipedia.org/wiki/Kapustinskii_equation this is the simplest thoery/equation for explaining this Thanks but its quite complicated for an AS student lol, but I understand now anyways so no worries 12. (Original post by sabre2th1) Ah.. Thanks a lot! You have helped a lot over the past few days, hopefully I will utilise some of the knowledge (you passed on) tomorrow ! Good luck ## Register Thanks for posting! You just need to create an account in order to submit the post 1. this can't be left blank this is what you'll be called on TSR 2. this can't be left blank never shared and never spammed 3. this can't be left blank 6 characters or longer with both numbers and letters is safer 4. this can't be left empty 1. By joining you agree to our Ts and Cs, privacy policy and site rules 2. Slide the button to the right to create your account Updated: May 14, 2012 New on TSR ### Moving on from GCSEs What advice would you give someone starting A-levels? Study resources
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8207947611808777, "perplexity": 1707.913763496996}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-23/segments/1406510264270.11/warc/CC-MAIN-20140728011744-00055-ip-10-146-231-18.ec2.internal.warc.gz"}
https://nag.co.uk/demos
# NAG Library Interactive Demos Click on the mathematical area of interest and try out NAG routines for yourself!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 19, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.27048537135124207, "perplexity": 6036.691998214879}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-17/segments/1524125945624.76/warc/CC-MAIN-20180422154522-20180422174522-00079.warc.gz"}
http://gmatclub.com/forum/simple-q-94276.html?kudos=1
Find all School-related info fast with the new School-Specific MBA Forum It is currently 25 Aug 2016, 05:19 GMAT Club Daily Prep Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email. Customized for You we will pick new questions that match your level based on your Timer History Track every week, we’ll send you an estimated GMAT score based on your performance Practice Pays we will pick new questions that match your level based on your Timer History Events & Promotions Events & Promotions in June Open Detailed Calendar Simple Q Author Message TAGS: Hide Tags Senior Manager Joined: 24 Jul 2009 Posts: 297 Followers: 3 Kudos [?]: 118 [0], given: 0 Show Tags 15 May 2010, 11:50 00:00 Difficulty: (N/A) Question Stats: 67% (01:53) correct 33% (00:41) wrong based on 1 sessions HideShow timer Statistics The average of 4 different positive integers is 125. If the largest of these integers is 150, what is the least possible value of the smallest integer? * 1 * 2 * 12 * 53 * 100 Manager Joined: 20 Apr 2010 Posts: 153 Location: I N D I A Followers: 3 Kudos [?]: 19 [0], given: 16 Show Tags 17 May 2010, 01:36 i also think the ans is D... whats OA..? Senior Manager Joined: 25 Jun 2009 Posts: 306 Followers: 2 Kudos [?]: 117 [0], given: 6 Show Tags 17 May 2010, 02:45 nverma wrote: The average of 4 different positive integers is 125. If the largest of these integers is 150, what is the least possible value of the smallest integer? * 1 * 2 * 12 * 53 * 100 D, Let the numbers be a,b,c and d then $$\frac{a+b+c+d}{4}$$ $$= 125$$ a+ b+c +d = 500 d = 150 a+b+c= 350 now 150 is the largest among the lot, then let c= 149 and b = 148 then equation becomes a+ 148+ 149 = 350 or a = 53 Manager Joined: 29 Oct 2009 Posts: 66 Followers: 1 Kudos [?]: 2 [0], given: 3 Show Tags 17 May 2010, 02:46 D should be the ans... Lets take other numbers to be 148 and 146. Adding these 2 numbers to 150 and then subtracting with 125*4 , ans comes to 56, so only D is closest to that. Manager Joined: 16 Feb 2010 Posts: 187 Followers: 2 Kudos [?]: 17 [0], given: 14 Show Tags 17 May 2010, 07:59 another way to solve in order to minimize one no we have to maximize all the other no.s to get the same average , the max. no. is 150 so the maximum limit of other no.s is 149 & 148 NOW (148 +149+150 +X)/4 = 125 so x=53 ans option D Re: Simple Q   [#permalink] 17 May 2010, 07:59 Similar topics Replies Last post Similar Topics: Simple PS 1 29 May 2011, 14:16 1 Simple Algebra 3 01 May 2011, 14:27 simple algebra 1 23 Mar 2011, 20:42 Simple question 1 09 Feb 2011, 00:02 1 simple q.. 8 28 Dec 2008, 21:28 Display posts from previous: Sort by
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4620794951915741, "perplexity": 9464.828246711986}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-36/segments/1471982293195.16/warc/CC-MAIN-20160823195813-00083-ip-10-153-172-175.ec2.internal.warc.gz"}
https://worldwidescience.org/topicpages/h/hanford+atomic+products.html
#### Sample records for hanford atomic products 1. Hanford Atomic Products Operation monthly report, January 1956 Energy Technology Data Exchange (ETDEWEB) 1956-02-24 This is the monthly report for the Hanford Atomic Laboratories Products Operation, February, 1956. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 2. Hanford Atomic Products Operation monthly report for June 1955 Energy Technology Data Exchange (ETDEWEB) 1955-07-28 This is the monthly report for the Hanford Atomic Products Operation, June, 1955. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 3. Hanford Atomic Products Operation monthly report, February 1954 Energy Technology Data Exchange (ETDEWEB) McCune, F.K. 1954-03-23 This is a progress report of the production reactors on the Hanford Reservation for the month of February 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 4. Hanford Atomic Products Operation monthly report, April 1953 Energy Technology Data Exchange (ETDEWEB) McCune, F.K. 1953-05-20 This is a progress report of the production reactors on the Hanford Reservation for the month of April 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 5. Hanford Atomic Products Operation monthly report, January 1954 Energy Technology Data Exchange (ETDEWEB) McCune, F.K. 1954-02-25 This is a progress report of the production reactors on the Hanford Reservation for the month of January 1954. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes the accomplishments and employee relations for that month. 6. Hanford Atomic Products Operation monthly report, March 1953 Energy Technology Data Exchange (ETDEWEB) McCune, F.K. 1953-04-22 This is a progress report of the production reactors on the Hanford Reservation for the month of March 1953. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 7. Hanford Atomic Products Operation monthly report, April 1954 Energy Technology Data Exchange (ETDEWEB) McCune, F.K. 1954-05-21 This is a progress report of the production reactors on the Hanford Reservation for the month of April 1954. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 8. Hanford Atomic Products Operation monthly report for March 1956 Energy Technology Data Exchange (ETDEWEB) 1956-04-20 This is the monthly report for the Hanford Laboratories Operation, March, 1956. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology; financial activities, visits, biology operation, physics and instrumentation research, employee relations, pile technology, safety and radiological sciences are discussed. 9. Hanford Atomic Products Operation monthly report for February 1956 Energy Technology Data Exchange (ETDEWEB) 1956-02-21 This is the monthly report for the Hanford Laboratories Operation, February, 1956. Metallurgy, reactors fuels, chemistry, dosimetry, separation processes, reactor technology financial activities, visits, biology operation, physics and instrumentation research, employee relations are discussed. 10. Hanford Atomic Products Operation monthly report, July 1955 Energy Technology Data Exchange (ETDEWEB) 1955-08-26 This document presents a summary of work and progress at the Hanford Engineer Works for July 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and services departments have sections presenting their monthly statistics, work, progress, and summaries. 11. Hanford Atomic Products for Operation monthly report, February 1955 Energy Technology Data Exchange (ETDEWEB) 1955-03-18 This document presents a summary of work and progress at the Hanford Engineer Works for February 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 12. Hanford Atomic Products Operation monthly report, October 1955 Energy Technology Data Exchange (ETDEWEB) 1955-11-30 This document presents a summary of work and progress at the Hanford Engineer works for October, 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 13. Hanford Atomic Products Operation monthly report, May 1955 Energy Technology Data Exchange (ETDEWEB) 1955-06-23 This document presents a summary of work and progress at the Hanford Engineer Works for May 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 14. Hanford Atomic Products Operation monthly report, May 1954 Energy Technology Data Exchange (ETDEWEB) 1954-06-22 This document presents a summary of work and progress at the Hanford Engineer Works for May 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Science, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 15. Hanford Atomic Products Operation monthly report, September 1955 Energy Technology Data Exchange (ETDEWEB) 1955-10-27 This document presents a summary of work and progress at the Hanford Engineer Works for September 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 16. Hanford Atomic Products Operation monthly report, October 1954 Energy Technology Data Exchange (ETDEWEB) 1954-11-24 This document presents a summary of work and progress at the Hanford Engineer Works for October 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 17. Hanford Atomic Products Operation monthly report, June 1953 Energy Technology Data Exchange (ETDEWEB) 1953-07-22 This document presents a summary of work and progress at the Hanford Engineer Works for June 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 18. Hanford Atomic Products Operation monthly report, August 1953 Energy Technology Data Exchange (ETDEWEB) 1953-09-18 This document presents a summary of work and progress at the Hanford Engineer Works for August, 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 19. Monthly report Hanford Atomic Products Operation, July 1954 Energy Technology Data Exchange (ETDEWEB) 1954-08-20 This document presents a summary of work and progress at the Hanford Engineer Works for July 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services Departments have sections presenting their monthly statistics, work, progress, and summaries. 20. Hanford Atomic Products Operation monthly report, August 1955 Energy Technology Data Exchange (ETDEWEB) 1955-09-27 This document presents a summary of work and progress at the Hanford Engineer Works for August 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Sciences, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 1. Hanford Atomic Products Operation monthly report, May 1953 Energy Technology Data Exchange (ETDEWEB) 1953-06-19 This document presents a summary of work and progress at the Hanford Engineer Works for May 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 2. Hanford Atomic Products Operation monthly report, April 1956 Energy Technology Data Exchange (ETDEWEB) 1956-05-21 This document presents a summary of work and progress at the Hanford Engineer Works for April 1956. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the technical, design and project sections. Costs for the various departments are presented in the financial departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Service departments have sections presenting their monthly statistics, work, progress, and summaries. 3. Hanford Atomic Products Operation monthly report, August 1956 Energy Technology Data Exchange (ETDEWEB) 1956-09-28 This document presents a summary of work and progress at the Hanford Engineer Works for August 1956. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Sciences, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 4. Hanford Atomic Products Operation monthly report, December 1954 Energy Technology Data Exchange (ETDEWEB) 1955-01-25 This document presents a summary of work and progress at the Hanford Engineer Works for December 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 5. Hanford Atomic Products Operation monthly report for May 1956 Energy Technology Data Exchange (ETDEWEB) 1956-06-21 This document presents a summary of work and progress at the Hanford Engineer Works for May, 1956. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 6. Hanford Atomic Products Operation monthly report, March 1954 Energy Technology Data Exchange (ETDEWEB) 1954-04-23 This document presents a summary of work and progress at the Hanford Engineer Works for March 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Service departments have sections presenting their monthly statistics, work, progress, and summaries. 7. Hanford Atomic Products Operation monthly report, June 1954 Energy Technology Data Exchange (ETDEWEB) 1954-07-26 This document presents a summary of work and progress at the Hanford Engineer Works for June 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 8. Hanford Atomic Products Operation monthly report, August 1954 Energy Technology Data Exchange (ETDEWEB) 1954-09-17 This document presents a summary of work and progress at the Hanford Engineer Works for August 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department report plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities, and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 9. Hanford Atomic Products Operation monthly report, January 1955 Energy Technology Data Exchange (ETDEWEB) 1955-02-21 This document presents a summary of work and progress at the Hanford Engineer Works for January 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical Design, and Project Sections. Costs for the various departments are presented in the Financial departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 10. Hanford Atomic Products Operation, monthly report, July 1956 Energy Technology Data Exchange (ETDEWEB) 1956-08-23 This document presents a summary of work and progress at the Hanford Engineer Works for July, 1956. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 11. Hanford Atomic Products Operation monthly report, March 1955 Energy Technology Data Exchange (ETDEWEB) 1955-04-20 This document presents a summary of work and progress at the Hanford Engineer Works for March 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 12. Hanford Atomic Products Operation monthly report, November 1955 Energy Technology Data Exchange (ETDEWEB) 1955-12-30 This document presents a summary of work and progress at the Hanford Engineer Works for November 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 13. Hanford Atomic Products Operation monthly report for April 1955 Energy Technology Data Exchange (ETDEWEB) 1955-05-23 This document presents a summary of work and progress at the Hanford Engineer Works for April 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 14. Hanford Atomic Products Operation monthly report, July 1953 Energy Technology Data Exchange (ETDEWEB) 1953-08-20 This document presents a summary of work and progress at the Hanford Engineer Works for July 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 15. Hanford Atomic Products Operation monthly report, October 1953 Energy Technology Data Exchange (ETDEWEB) 1953-11-20 This document presents a summary of work and progress at the Hanford Engineer Works for October 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services. Employee and Public Relations, and Community Real Estate and Service departments have sections presenting their monthly statistics, work, progress, and summaries. 16. Hanford Atomic Products Operation monthly report for September 1954 Energy Technology Data Exchange (ETDEWEB) 1954-10-25 This document presents a summary of work and progress at the Hanford Engineer Works for September 1954. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 17. Hanford Atomic Products Operation monthly report, December 1953 Energy Technology Data Exchange (ETDEWEB) 1954-01-22 This document presents a summary of work and progress at the Hanford Engineer Works for December 1953. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summaries work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 18. Hanford Atomic Products Operation monthly report for December 1955 Energy Technology Data Exchange (ETDEWEB) 1956-01-30 This document presents a summary of work and progress at the Hanford Engineer Works for December 1955. The report is divided into sections by department. A plant wide general summary is included at the beginning of the report, after which the departmental summaries begin. The Manufacturing Department reports plant statistics, and summaries for the Metal Preparation, Reactor and Separation sections. The Engineering Departments section summarizes work for the Technical, Design, and Project Sections. Costs for the various departments are presented in the Financial Departments summary. The Medical, Radiological Sciences, Utilities and General Services, Employee and Public Relations, and Community Real Estate and Services departments have sections presenting their monthly statistics, work, progress, and summaries. 19. History of Hanford Site Defense Production (Brief) International Nuclear Information System (INIS) GERBER, M.S. 2001-01-01 This paper acquaints the audience with the history of the Hanford Site, America's first full-scale defense plutonium production site. The paper includes the founding and basic operating history of the Hanford Site, including World War II construction and operations, three major postwar expansions (1947-55), the peak years of production (1956-63), production phase downs (1964-the present), a brief production spurt from 1984-86, the end of the Cold War, and the beginning of the waste cleanup mission. The paper also delineates historical waste practices and policies as they changed over the years at the Hanford Site, past efforts to chemically treat, ''fractionate,'' and/or immobilize Hanford's wastes, and resulting major waste legacies that remain today. This paper presents original, primary-source research into the waste history of the Hanford Site. Finally, the paper places the current Hanford Site waste remediation endeavors in the broad context of American and world history 20. History of Hanford Site Defense Production (Brief) Energy Technology Data Exchange (ETDEWEB) GERBER, M S 2001-02-01 This paper acquaints the audience with the history of the Hanford Site, America's first full-scale defense plutonium production site. The paper includes the founding and basic operating history of the Hanford Site, including World War II construction and operations, three major postwar expansions (1947-55), the peak years of production (1956-63), production phase downs (1964-the present), a brief production spurt from 1984-86, the end of the Cold War, and the beginning of the waste cleanup mission. The paper also delineates historical waste practices and policies as they changed over the years at the Hanford Site, past efforts to chemically treat, ''fractionate,'' and/or immobilize Hanford's wastes, and resulting major waste legacies that remain today. This paper presents original, primary-source research into the waste history of the Hanford Site. Finally, the paper places the current Hanford Site waste remediation endeavors in the broad context of American and world history. 1. Management of Hanford Site non-defense production reactor spent nuclear fuel, Hanford Site, Richland, Washington International Nuclear Information System (INIS) 1997-03-01 The US Department of Energy (DOE) needs to provide radiologically, and industrially safe and cost-effective management of the non-defense production reactor spent nuclear fuel (SNF) at the Hanford Site. The proposed action would place the Hanford Site's non-defense production reactor SNF in a radiologically- and industrially-safe, and passive storage condition pending final disposition. The proposed action would also reduce operational costs associated with storage of the non-defense production reactor SNF through consolidation of the SNF and through use of passive rather than active storage systems. Environmental, safety and health vulnerabilities associated with existing non-defense production reactor SNF storage facilities have been identified. DOE has determined that additional activities are required to consolidate non-defense production reactor SNF management activities at the Hanford Site, including cost-effective and safe interim storage, prior to final disposition, to enable deactivation of facilities where the SNF is now stored. Cost-effectiveness would be realized: through reduced operational costs associated with passive rather than active storage systems; removal of SNF from areas undergoing deactivation as part of the Hanford Site remediation effort; and eliminating the need to duplicate future transloading facilities at the 200 and 400 Areas. Radiologically- and industrially-safe storage would be enhanced through: (1) removal from aging facilities requiring substantial upgrades to continue safe storage; (2) utilization of passive rather than active storage systems for SNF; and (3) removal of SNF from some storage containers which have a limited remaining design life. No substantial increase in Hanford Site environmental impacts would be expected from the proposed action. Environmental impacts from postulated accident scenarios also were evaluated, and indicated that the risks associated with the proposed action would be small 2. High-Volume Non-Destructive Test Applications at the Hanford Atomic Products Operation; Applications Industrielles des Essais Non Destructifs a l'Etablissement Nucleaire de Hanford; Provedenie bol'shogo chisla nedestruktivnykh ispytanii v ''khenford atomik prodakts opereishen''; Ensayos No Destructivos en Gran Escala Aplicados en Hanford Energy Technology Data Exchange (ETDEWEB) Worlton, D. C. [Pacific Northwest Laboratory, Battelle Memorial Institute, Richland, WA (United States) 1965-10-15 Safety and efficiency of critical Hanford processes are assured with rapid, reliable, and automatic non-destructive tests. High-sensitive eddy-current and ultrasonic inspection systems are in routine use in the field and in manufacturing production processes to provide maximum quality assurance of large volumes of material in minimum inspection time. This paper describes inspection systems being used to ensure quality of Hanford's production nuclear-fuel manufacturing processes. Operated as regular in-line manufacturing equipment, these systems employ ultrasonic attenuation measurements to monitor grain structure of bare uranium fuel cores, ultrasonic and eddy- current techniques to ensure adequate bonding and thickness of 0.040 in aluminium cladding on canned elements, and novel wide-band, high-resolution ultrasonic inspection techniques to detect defects in the fuel end-weld closures. Combined eddy-current and ultrasonic tests are applied simultaneously to perform a complete fuel- element inspection on a nine-second cycle; defective elements are automatically segregated from the process stream. Emphasis is given to advanced ultrasonic test methods of inspecting thin-walled, fuel-sheath tubing. Special highly focused transducers are used with wide-band circuitry to generate pure shear waves in 0.015-in-thick wall tubing. Lamb and other complicated wave motions are excluded so that tests results are readily interpreted and reproduced. Novel, economical methods of producing defect standards have been developed, as have critically important methods of ensuring uniform operating characteristics of the transducers themselves. Automatic tubing inspection equipment has been developed, and results of its routine use in testing some 30 000 ft of tubing are summarized. Finally, eddy-current techniques developed specifically for inspecting installed heat-exchanger tubing are reviewed. The technique employs novel read-out features which plot defect indications as oscilloscope 3. List of currently classified documents relative to Hanford Production Facilities Operations originated on the Hanford Site between 1961 and 1972 Energy Technology Data Exchange (ETDEWEB) 1993-04-01 The United States Department of Energy (DOE) has declared that all Hanford plutonium production- and operations-related information generated between 1944 and 1972 is declassified. Any documents found and deemed useful for meeting Hanford Environmental Dose Reconstruction (HEDR) objectives may be declassified with or without deletions in accordance with DOE guidance by Authorized Derivative Declassifiers. The September 1992, letter report, Declassifications Requested by the Technical Steering Panel of Hanford Documents Produced 1944--1960, (PNWD-2024 HEDR UC-707), provides an important milestone toward achieving a complete listing of documents that may be useful to the HEDR Project. The attached listing of approximately 7,000 currently classified Hanford-originated documents relative to Hanford Production Facilities Operations between 1961 and 1972 fulfills TSP Directive 89-3. This list does not include such titles as the Irradiation Processing Department, Chemical Processing Department, and Hanford Laboratory Operations monthly reports generated after 1960 which have been previously declassified with minor deletions and made publicly available. Also Kaiser Engineers Hanford (KEH) Document Control determined that no KEH documents generated between January 1, 1961 and December 31, 1972 are currently classified. Titles which address work for others have not been included because Hanford Site contractors currently having custodial responsibility for these documents do not have the authority to determine whether other than their own staff have on file an appropriate need-to-know. Furthermore, these documents do not normally contain information relative to Hanford Site operations. 4. International Atomic Energy Agency/Hanford Site shared use of calorimeters International Nuclear Information System (INIS) Welsh, T.L. 1997-01-01 Hanford Site operators combine gamma ray isotopic and calorimetry measurements for nondestructive plutonium assay. Such measurements offer lower variability (particularly for heterogeneous materials) and decreased radiation exposure, cost, waste, intrusiveness, and material handling compared to destructive analysis. Until now, the International Atomic Energy Agency (IAEA) has relied on destructive analysis to perform the most accurate verification requirements for plutonium stored under safeguards at the Hanford Site. It was recognized that using calorimetry could significantly reduce the need for the IAEA to perform destructive analysis. To authorize the operator's calorimeters for routine IAEA use, however, it was necessary to develop authentication features and perform independent 1558 testing. Authentication features include IAEA control of the hardware and calorimeter operating system software, measurement of certified IAEA standards, sealing of calorimeter chambers, and limited destructive analysis of IAEA selected items. A field test of these authentication features was performed at the Hanford Site in June 1997. The field test also was meant to enhance the credibility the IAEA imputes to calorimetry prior to its implementation. Progress in shared use of the Hanford Site calorimeters is reported 5. Hanford Immobilized Low-Activity Waste Product Acceptance Test Plan International Nuclear Information System (INIS) Peeler, D. 1999-01-01 'The Hanford Site has been used to produce nuclear materials for the U.S. Department of Energy (DOE) and its predecessors. A large inventory of radioactive and mixed waste, largely generated during Pu production, exists in 177 underground single- and double-shell tanks. These wastes are to be retrieved and separated into low-activity waste (LAW) and high-level waste (HLW) fractions. The DOE is proceeding with an approach to privatize the treatment and immobilization of Handord''s LAW and HLW.' 6. Hanford Immobilized Low-Activity Waste Product Acceptance Test Plan Energy Technology Data Exchange (ETDEWEB) Peeler, D. 1999-06-22 'The Hanford Site has been used to produce nuclear materials for the U.S. Department of Energy (DOE) and its predecessors. A large inventory of radioactive and mixed waste, largely generated during Pu production, exists in 177 underground single- and double-shell tanks. These wastes are to be retrieved and separated into low-activity waste (LAW) and high-level waste (HLW) fractions. The DOE is proceeding with an approach to privatize the treatment and immobilization of Handord''s LAW and HLW.' 7. Transmutation of fission products and actinide waste at Hanford Energy Technology Data Exchange (ETDEWEB) Daemen, L.L.; Pitcher, E.J.; Russell, G.J. [Los Alamos National Laboratory, NM (United States) 1995-10-01 The authors studied the neutronics of an ATW system for the transmutation of the fission products ({sup 99}Tc in particular) and the type of actinide waste stored in several tanks at Hanford. The heart of the system is a highly-efficient neutron production target. It is surrounded by a blanket containing a moderator/reflector material, as well as the products to be transmuted. The fission products are injected into the blanket in the form of an aqueous solution in heavy water, whereas an aqueous actinides slurry is circulated in the outer part of the blanket. For the sake of definiteness, the authors focussed on {sup 99}Tc (the most difficult fission product to transmute), and {sup 239}Pu, {sup 237}Np, and {sup 241}Am. Because of the low thermal neutron absorption cross-section of {sup 99}Tc, considerable care and effort must be devoted to the design of a very efficient neutron source. 8. Hanford regulated laundry: inventory control and production improvement study Energy Technology Data Exchange (ETDEWEB) Hostick, C. J.; Imhoff, C. H.; Levine, L. O. 1986-04-01 The purpose of this report is to assist the Hanford regulated laundry facility in reducing processing costs and in improving facility performance. Specific problem areas addressed were: no method for determining optimum manpower requirements, resulting in excessive amounts of employee overtime; no buffer inventory available to offset demand peaks, resulting in additional employee overtime and unmet demand; lack of adequate inventory control, resulting in unnecessary inventory costs; and no detailed analysis of the impact of 100% monitoring. 9. The economic and community impacts of closing Hanford's N Reactor and nuclear materials production facilities International Nuclear Information System (INIS) Scott, M.J.; Belzer, D.B.; Nesse, R.J.; Schultz, R.W.; Stokowski, P.A.; Clark, D.C. 1987-08-01 This study discusses the negative economic impact on local cities and counties and the State of Washington of a permanent closure of nuclear materials production at the Hanford Site, located in the southeastern part of the state. The loss of nuclear materials production, the largest and most important of the five Department of Energy (DOE) missions at Hanford, could occur if Hanford's N Reactor is permanently closed and not replaced. The study provides estimates of statewide and local losses in jobs, income, and purchases from the private sector caused by such an event; it forecasts impacts on state and local government finances; and it describes certain local community and social impacts in the Tri-Cities (Richland, Kennewick, and Pasco) and surrounding communities. 33 refs., 8 figs., 22 tabs 10. Reengineering Hanford Energy Technology Data Exchange (ETDEWEB) Badalamente, R.V.; Carson, M.L.; Rhoads, R.E. 1995-03-01 The Department of Energy Richland Operations Office is in the process of reengineering its Hanford Site operations. There is a need to fundamentally rethink and redesign environmental restoration and waste management processes to achieve dramatic improvements in the quality, cost-effectiveness, and timeliness of the environmental services and products that make cleanup possible. Hanford is facing the challenge of reengineering in a complex environment in which major processes cuts across multiple government and contractor organizations and a variety of stakeholders and regulators have a great influence on cleanup activities. By doing the upfront work necessary to allow effective reengineering, Hanford is increasing the probability of its success. 11. Reengineering Hanford International Nuclear Information System (INIS) Badalamente, R.V.; Carson, M.L.; Rhoads, R.E. 1995-03-01 The Department of Energy Richland Operations Office is in the process of reengineering its Hanford Site operations. There is a need to fundamentally rethink and redesign environmental restoration and waste management processes to achieve dramatic improvements in the quality, cost-effectiveness, and timeliness of the environmental services and products that make cleanup possible. Hanford is facing the challenge of reengineering in a complex environment in which major processes cuts across multiple government and contractor organizations and a variety of stakeholders and regulators have a great influence on cleanup activities. By doing the upfront work necessary to allow effective reengineering, Hanford is increasing the probability of its success 12. Environmental characterization of two potential locations at Hanford for a new production reactor Energy Technology Data Exchange (ETDEWEB) Watson, E.C.; Becker, C.D.; Fitzner, R.E.; Gano, K.A.; Imhoff, K.L.; McCallum, R.F.; Myers, D.A.; Page, T.L.; Price, K.R.; Ramsdell, J.V.; Rice D.G.; Schreiber D.L.; Skumatz L.A.; Sommer D.J.; Tawil J.J.; Wallace R.W.; Watson D.G. 1984-09-01 This report describes various environmental aspects of two areas on the Hanford Site that are potential locations for a New Production Reactor (NPR). The area known as the Skagit Hanford Site is considered the primary or reference site. The second area, termed the Firehouse Site, is considered the alternate site. The report encompasses an environmental characterization of these two potential NPR locations. Eight subject areas are covered: geography and demography; ecology; meteorology; hydrology; geology; cultural resources assessment; economic and social effects of station construction and operation; and environmental monitoring. 80 refs., 68 figs., 109 tabs. 13. Decommissioning planning and the assessment of alternatives for the Hanford production reactors International Nuclear Information System (INIS) Miller, C.E. Jr.; Potter, R.F. 1985-01-01 Several years ago, the US Department of Energy began assessing alternatives and planning the decommissioning of eight shut-down plutonium production reactors located on the DOE Hanford Site in Washington State. The first of these graphite-moderated, water-cooled, reactors was built and started up in 1944 as part of the World War II Manhattan Project. The last of them started up in 1955. The eight reactors each operated for 12 to 24 years, with all eight operating simultaneously for about 10 years. In the 1960's, production needs declined and the reactors were one-by-one permanently shut down, the last of them in 1971. (A ninth Hanford production reactor, N Reactor, was started up in 1963; it is still operating and is not within the scope of the decommissioning planning and alternatives assessment work reported in this paper). This paper provides an overview description of the decommissioning plan for the eight shut-down Hanford production reactors and their associated fuel storage basins. Included are descriptions of the decommissioning alternatives considered for the facilities, along with discussions of National Environmental Policy Act (NEPA) process activities applicable to the Hanford decommissioning work. The criteria used in assessing decommissioning alternatives and the assumptions used in the decommissioning planning are identified. 4 refs., 8 figs., 3 tabs 14. Radionuclide concentrations in agricultural products near the Hanford Site, 1982 through 1992 International Nuclear Information System (INIS) Antonio, E.J. 1994-06-01 The Pacific Northwest Laboratory reviewed monitoring data for agricultural products collected from 1982 through 1992 near the Hanford Site to determine radionuclide concentration trends. While samples were collected and analyzed, and results reported annual in Hanford Site environmental reports, an 11-year data set was reviewed for this report to increase the ability to assess trends and potential Hanford effects. Products reviewed included milk, chicken, eggs, beef, vegetables, fruit, wine, wheat, and alfalfa. To determine which radionuclides were detected sufficiently often to permit analysis for trends and effects, each radionuclide concentration and its associated uncertainty were ratioed. Radionuclides were considered routinely detectable if more than 50% of the ratios were between zero and one. Data for these radionuclides were then analyzed statistically, using analyses of variance. The statistical analyses indicated the following: for the most part, there were no measurable effects for Hanford operations; radionuclide concentrations in all products reviewed remained relatively low when compared to concentrations that would result in a 1-mrem effective dose equivalent to an individual; radionuclide concentrations are decreasing in general; however, 90 Sr concentrations in all media and 129 I in milk increased from 1982 to 1986, then decreased gradually for the remainder of the review period. The 129 I concentrations may be correlated with processing of irradiated reactor fuel at the Plutonium-Uranium Extraction (PUREX) Plant 15. Nuclear graphite development, operational problems, and resolution of these problems at the Hanford production reactors International Nuclear Information System (INIS) Morgan, W.C. 1996-01-01 This paper chronicles the history of the Hanford Production Reactor, from the initial design considerations for B, D, and F Reactors through the selection of the agreed method for safe disposal of the decommissioned reactors. The operational problems that challenged the operations and support staff of each new generation of production reactors, the engineering actions an operational changes that alleviated or resolved the immediate problems, the changes in reactor design and design-bases for the next generation of production reactors, and the changes in manufacturing variables that resulted in new ''improved'' grades of nuclear graphites for use in the moderators of the Hanford Production Reactors are reviewed in the context of the existing knowledge-base and the mission-driven priorities on the time. 14 refs, 6 figs, 3 tabs 16. Reinventing government: Reinventing Hanford International Nuclear Information System (INIS) Mayeda, J.T. 1994-05-01 The Hanford Site was established in 1943 as one of the three original Manhattan Project locations involved in the development of atomic weapons. It continued as a defense production center until 1988, when its mission changed to environmental restoration and remediation. The Hanford Site is changing its business strategy and in doing so, is reinventing government. This new development has been significantly influenced by a number of external sources. These include: the change in mission, reduced security requirements, new found partnerships, fiscal budgets, the Tri-Party agreement and stakeholder involvement. Tight budgets and the high cost of cleanup require that the site develop and implement innovative cost saving approaches to its mission. Costeffective progress is necessary to help assure continued funding by Congress 17. Hanford Immobilized LAW Product Acceptance: Tanks Focus Area Testing Data Package II International Nuclear Information System (INIS) Schulz, Rebecca L.; Lorier, Troy H.; Peeler, David K.; Brown, Kevin G.; Reamer, Irene A.; Vienna, John D.; Jiricka, Antonin; Jorgensen, Benaiah M.; Smith, Donald E. 2001-01-01 This report is a continuation of the Hanford Immobilized Low Activity Waste (LAW) Product Acceptance (HLP): Initial Tanks Focus Area Testing Data Package (Vienna (and others) 2000). In addition to new 5000-h product consistency test (PCT), vapor hydration test (VHT), and alteration products data, some previously reported data together with relevant background information are included for an easily accessible source of reference when comparing the response of the various glasses to different test conditions. A matrix of 55 glasses was developed and tested to identify the impact of glass composition on long-term corrosion behavior and to develop an acceptable composition region for Hanford LAW glasses. Of the 55 glasses, 45 were designed to systematically vary the glass composition, and 10 were selected because large and growing databases on their corrosion characteristics had accumulated. The targeted and measured compositions of these glasses are found in the Appendix A. All glasses were fabricated according to standard procedures and heat treated to simulate the slow cooling that will occur in a portion of the waste glass after vitrification in the planned treatment facility at Hanford 18. Inventory of chemicals used at Hanford Site production plants and support operations (1944-1980) Energy Technology Data Exchange (ETDEWEB) Klem, M. J. 1990-04-01 A complete list of chemicals used in the production facilities and support operations of the US Department of Energy Hanford Site is presented to aid development of plans for characterizing the radioactive liquid chemical wastes stored in the 149 single-shell tanks. The complete chemical list is compared to the list provided by the regulatory agencies to identify hazardous chemicals stored in the single-shell tanks. A reduced list has been developed by others and is used to identify the chemical constituents for analysis in the Waste Characterization Plan for the Hanford Site Single-Shell Tanks. The chemical list is based on chemical process flowsheets, essential material consumption records, letters, reports, and other historical data. 14 refs., 36 tabs. 19. Strategy for product composition control in the Hanford Waste Vitrification Plant International Nuclear Information System (INIS) Bryan, M.F.; Piepel, G.F. 1996-03-01 The Hanford Waste Vitrification Plant (HWVP) will immobilize transuranic and high-level radioactive waste in borosilicate glass. The major objective of the Process/Product Model Development (PPMD) cost account of the Pacific Northwest Laboratory HWVP Technology Development (PHTD) Project is the development of a system for guiding control of feed slurry composition (which affects glass properties) and for checking and documenting product quality. This document lays out the broad structure of HWVP's product composition control system, discusses five major algorithms and technical issues relevant to this system, and sketches the path of development and testing 20. Plutonium production story at the Hanford site: processes and facilities history Energy Technology Data Exchange (ETDEWEB) Gerber, M.S., Westinghouse Hanford 1996-06-20 This document tells the history of the actual plutonium production process at the Hanford Site. It contains five major sections: Fuel Fabrication Processes, Irradiation of Nuclear Fuel, Spent Fuel Handling, Radiochemical Reprocessing of Irradiated Fuel, and Plutonium Finishing Operations. Within each section the story of the earliest operations is told, along with changes over time until the end of operations. Chemical and physical processes are described, along with the facilities where these processes were carried out. This document is a processes and facilities history. It does not deal with the waste products of plutonium production. 1. Metal powder production by gas atomization Science.gov (United States) Ting, E. Y.; Grant, N. J. 1986-01-01 The confined liquid, gas-atomization process was investigated. Results from a two-dimensional water model showed the importance of atomization pressure, as well as delivery tube and atomizer design. The atomization process at the tip of the delivery tube was photographed. Results from the atomization of a modified 7075 aluminum alloy yielded up to 60 wt pct. powders that were finer than 45 microns in diameter. Two different atomizer designs were evaluated. The amount of fine powders produced was correlated to a calculated gas-power term. An optimal gas-power value existed for maximized fine powder production. Atomization at gas-power greater than or less than this optimal value produced coarser powders. 2. Atomic and molecular beams production and collimation CERN Document Server Lucas, Cyril Bernard 2013-01-01 Atomic and molecular beams are employed in physics and chemistry experiments and, to a lesser extent, in the biological sciences. These beams enable atoms to be studied under collision-free conditions and allow the study of their interaction with other atoms, charged particles, radiation, and surfaces. Atomic and Molecular Beams: Production and Collimation explores the latest techniques for producing a beam from any substance as well as from the dissociation of hydrogen, oxygen, nitrogen, and the halogens.The book not only provides the basic expressions essential to beam design but also offers 3. Isotope Production at the Hanford Site in Richland, Washington Energy Technology Data Exchange (ETDEWEB) Ammoniums 1999-06-01 This report was prepared in response to a request from the Nuclear Energy Research Advisory Committee (NERAC) subcommittee on ''Long-Term Isotope Research and Production Plans.'' The NERAC subcommittee has asked for a reply to a number of questions regarding (1) ''How well does the Department of Energy (DOE) infrastructure sme the need for commercial and medical isotopes?'' and (2) ''What should be the long-term role of the federal government in providing commercial and medical isotopes?' Our report addresses the questions raised by the NERAC subcommittee, and especially the 10 issues that were raised under the first of the above questions (see Appendix). These issues are related to the isotope products offered by the DOE Isotope Production Sites, the capabilities and condition of the facilities used to produce these products, the management of the isotope production programs at DOE laboratories, and the customer service record of the DOE Isotope Production sites. An important component of our report is a description of the Fast Flux Test Facility (FFTF) reactor at the Hbford Site and the future plans for its utilization as a source of radioisotopes needed by nuclear medicine physicians, by researchers, and by customers in the commercial sector. In response to the second question raised by the NERAC subcommittee, it is our firm belief that the supply of isotopes provided by DOE for medical, industrial, and research applications must be strengthened in the near future. Many of the radioisotopes currently used for medical diagnosis and therapy of cancer and other diseases are imported from Canada, Europe, and Asia. This situation places the control of isotope availability, quality, and pricing in the hands of non-U.S. suppliers. It is our opinion that the needs of the U.S. customers for isotopes and isotope products are not being adequately served, and that the DOE infrastructure and facilities devoted to the 4. Developing historical food production and consumption data for 131I dose estimates: The Hanford experience International Nuclear Information System (INIS) Anderson, D.M.; Marsh, T.L.; Deonigi, D.A. 1996-01-01 This paper describes the methods used to reconstruct the movement of commercial foods in and through the study area of the Hanford Environmental Dose Reconstruction Project. The most dose-relevant radionuclide released from Hanford separations plants was 131 I via the atmospheric pathway. As a result of atmospheric deposition of 131 I, commercial food supplies may have been contaminated. Because of the half-life of 131 I is relatively short, foods consumed soon after production, such as milk and produce, presented the highest risk. For that reason, this paper deals primarily with the reconstruction of milk and produce production, marketing, and consumption from 1945-1951, the period with the highest known 131 I releases. The reconstructed food production and consumption information was used as input to radiation dose estimates for representative individuals and as default values for real individuals who may not remember where they obtained food or how much they consumed during that period. Specific methods for tracing the movement of commercial milk and produce back from the point of human consumption, through commercial markets, to original production are presented. Results include the characteristics of food consumption exhibited by representative individuals, examples of commercial milk and produce market structures, and a review of commercial milk production and processing practices from 1945-1951 5. Trends in radionuclide concentrations for wildlife and food products near Hanford for the period 1971 through 1988 International Nuclear Information System (INIS) Eberhardt, L.E.; Cadwell, L.L.; Price, K.R.; Carlile, D.W. 1989-10-01 The objective of this summary investigation was to identify trends in radionuclide concentrations for wildlife and food products sampled from 1971 through 1988 as part of the Hanford Site Environmental Monitoring Program. No upward trends in radionuclide concentrations were detected for any wildlife or food products. Several sample types demonstrated significantly declining radionuclide concentrations. Three factors appeared to be responsible for the trends. First, the cessation of atmospheric testing by the United States and Soviet Union in 1971 contributed to the decline of radionuclides in some samples. Second, contaminants discharged to the Columbia River were reduced subsequent to the 1971 shutdown of the last Hanford nuclear reactor that used a once-through cooling water design. The reactor closing resulted in declines in activation products in oysters from Willapa Bay and in whitefish from the Hanford Reach of the Columbia River. Third, reductions in radionuclide concentrations in Hanford wildlife suggested a decreasing availability of environmental contaminants to wildlife. Remediation of areas having environmental surface contaminants on the Hanford Site was identified as a probable cause. 5 refs., 4 figs., 2 tabs 6. The economic and community impacts of closing Hanford's N Reactor and nuclear materials production facilities Energy Technology Data Exchange (ETDEWEB) Scott, M.J.; Belzer, D.B.; Nesse, R.J.; Schultz, R.W.; Stokowski, P.A.; Clark, D.C. 1987-08-01 This study discusses the negative economic impact on local cities and counties and the State of Washington of a permanent closure of nuclear materials production at the Hanford Site, located in the southeastern part of the state. The loss of nuclear materials production, the largest and most important of the five Department of Energy (DOE) missions at Hanford, could occur if Hanford's N Reactor is permanently closed and not replaced. The study provides estimates of statewide and local losses in jobs, income, and purchases from the private sector caused by such an event; it forecasts impacts on state and local government finances; and it describes certain local community and social impacts in the Tri-Cities (Richland, Kennewick, and Pasco) and surrounding communities. 33 refs., 8 figs., 22 tabs. 7. Production and detection of cold antihydrogen atoms CERN Multimedia Amoretti, M; Bonomi, G; Bouchta, A; Bowe, P; Carraro, C; Cesar, C L; Charlton, M; Collier, M; Doser, Michael; Filippini, V; Fine, K S; Fontana, A; Fujiwara, M C; Funakoshi, R; Genova, P; Hangst, J S; Hayano, R S; Holzscheiter, M H; Jørgensen, L V; Lagomarsino, V; Landua, Rolf; Landua, Rolf; Lindelöf, D; Lodi-Rizzini, E; Macri, M; Madsen, N; Manuzio, G; Marchesotti, M; Montagna, P; Pruys, H S; Regenfus, C; Riedler, P; Rochet, J; Rotondi, A; Rouleau, G; Testera, G; Van der Werf, D P; Variola, A; Watson, T L; CERN. Geneva 2002-01-01 A theoretical underpinning of the standard model of fundamental particles and interactions is CPT invariance, which requires that the laws of physics be invariant under the combined discrete operations of charge conjugation, parity and time reversal. Antimatter, the existence of which was predicted by Dirac, can be used to test the CPT theorem experimental investigations involving comparisons of particles with antiparticles are numerous. Cold atoms and anti-atoms, such as hydrogen and anti-hydrogen, could form the basis of a new precise test, as CPT invariance implies that they must have the same spectrum. Observations of antihydrogen in small quantities and at high energies have been reported at the European Organization for Nuclear Research (CERN) and at Fermilab, but were not suited to precision comparison measurements. Here we demonstrate the production of antihydrogen atoms at very low energy by mixing trapped antiprotons and positrons in a cryogenic environment. The neutral anti-atoms have been detected... 8. Remedial Investigation of Hanford Site Releases to the Columbia River International Nuclear Information System (INIS) Lerch, J.A. 2009-01-01 In south-central Washington State, the Columbia River flows through the U.S. Department of Energy Hanford Site. A primary objective of the Hanford Site cleanup mission is protection of the Columbia River, through remediation of contaminated soil and groundwater that resulted from its weapons production mission. Within the Columbia River system, surface water, sediment, and biota samples related to potential Hanford Site hazardous substance releases have been collected since the start of Hanford operations. The impacts of Hanford Site hazardous substance releases to the Columbia River in areas upstream, within, and downstream of the Hanford Site boundary have been previously investigated as mandated by the U.S. Department of Energy requirements under the Atomic Energy Act. The impacts are now being assessed under the Comprehensive Environmental Response, Compensation, and Liability Act of 1980 via a remedial investigation. The Remedial Investigation Work Plan for Hanford Site Releases to the Columbia River has been developed and issued to initiate the remedial investigation. The work plan establishes a phased approach to characterize contaminants, assess current risks, and determine whether or not there is a need for any cleanup actions. Field investigation activities began in October 2008 and are anticipated to continue into Fall 2009 over a 120 mile stretch of the Columbia River. Information gained from performing this remedial investigation will ultimately be used to help make final regulatory decisions for cleaning up Hanford Site contamination that exists in and along the Columbia River. (authors) 9. Trends in radionuclide concentrations for wildlife and food products near Hanford for the period 1971-1988 International Nuclear Information System (INIS) Cadwell, L.L.; Eberhardt, L.E.; Price, K.R.; Carlile, D.W. 1990-01-01 We evaluated the Hanford environmental data base for trends in radionuclide concentrations in wildlife and food products sampled from 1971 through 1988 on or near the U.S. Department of Energy's Hanford Site in southeastern Washington. Although statistical analyses showed short-term changes, no upward trends in radionuclide concentrations were detected. Many samples showed a significant decline in some radionuclides, particularly for 137 Cs. Concentrations of 65 Zn also showed a downward trend in many samples. Cessation of atmospheric testing by the United States and the USSR in 1971 contributed to the decline in radionuclide levels in some samples. Contaminants discharged to the Columbia River at Hanford were reduced after shutdown of the last once-through cooling-water reactor in 1971. A decline in concentrations of 65 Zn in oysters from Willapa Bay and 60 Co and 65 Zn in mountain whitefish from the Hanford Reach of the Columbia River are attributable to reactor closure. There was also an apparent reduction in availability of radiological contamination to Hanford wildlife after decommissioning of waste-water disposal ponds and remediation of contaminated terrestrial sites 10. Hanford recycling Energy Technology Data Exchange (ETDEWEB) Leonard, I.M. 1996-09-01 This paper is a study of the past and present recycling efforts on the Hanford site and options for future improvements in the recycling program. Until 1996, recycling goals were voluntarily set by the waste generators: this year, DOE has imposed goals for all its sites to accomplish by 1999. Hanford is presently meeting the voluntary site goals, but may not be able to meet all the new DOE goals without changes to the program. Most of these new DOE goals are recycling goals: * Reduce the generation of radioactive (low-level) waste from routine operations 50 percent through source reduction and recycling. * Reduce the generation of low-level mixed waste from routine operations 50 percent through source reduction and recycling. * Reduce the generation of hazardous waste from routine operations 50 percent through source reduction and recycling. * Recycle 33 percent of the sanitary waste from all operations. * Increase affirmative procurement of EPA-designated recycled items to 100 percent. The Hanford recycling program has made great strides-there has been a 98 percent increase in the amount of paper recycled since its inception in 1990. Hanford recycles paper, chemicals cardboard, tires, oil, batteries, rags, lead weights, fluorescent tubes, aerosol products, concrete, office furniture, computer software, drums, toner cartridges, and scrap metal. Many other items are recycled or reused by individual groups on a one time basis without a formal contract. Several contracts are closed-loop contracts which involve all parts of the recycle loop. Considerable savings are generated from recycling, and much more is possible with increased attention and improvements to this program. General methods for improving the recycling program to ensure that the new goals can be met are: a Contract and financial changes 0 Tracking database and methods improvements 0 Expanded recycling efforts. Specifically, the Hanford recycling program would be improved by: 0 Establishing one overall 11. Preliminary Assessment of the Hanford Tank Waste Feed Acceptance and Product Qualification Programs Energy Technology Data Exchange (ETDEWEB) Herman, C. C.; Adamson, Duane J.; Herman, D. T.; Peeler, David K.; Poirier, Micheal R.; Reboul, S. H.; Stone, M. E.; Peterson, Reid A.; Chun, Jaehun; Fort, James A.; Vienna, John D.; Wells, Beric E. 2013-04-01 The U.S. Department of Energy Office of Environmental Management (EM) is engaging the national laboratories to provide the scientific and technological rigor to support EM program and project planning, technology development and deployment, project execution, and assessment of program outcomes. As an early demonstration of this new responsibility, Savannah River National Laboratory (SRNL) and Pacific Northwest National Laboratory (PNNL) have been chartered to implement a science and technology program addressing Hanford Tank waste feed acceptance and product qualification. As a first step, the laboratories examined the technical risks and uncertainties associated with the planned waste feed acceptance and qualification testing for Hanford tank wastes. Science and technology gaps were identified for work associated with 1) feed criteria development with emphasis on identifying the feed properties and the process requirements, 2) the Tank Waste Treatment and Immobilization Plant (WTP) process qualification program, and 3) the WTP HLW glass product qualification program. Opportunities for streamlining the accetpance and qualification programs were also considered in the gap assessment. Technical approaches to address the science and technology gaps and/or implement the opportunities were identified. These approaches will be further refined and developed as strong integrated teams of researchers from national laboratories, contractors, industry, and academia are brought together to provide the best science and technology solutions. Pursuing the identified approaches will have immediate and long-term benefits to DOE in reducing risks and uncertainties associated with tank waste removal and preparation, transfers from the tank farm to the WTP, processing within the WTP Pretreatment Facility, and in producing qualified HLW glass products. Additionally, implementation of the identified opportunities provides the potential for long-term cost savings given the anticipated 12. Decommissioning of eight surplus production reactors at the Hanford Site, Richland, Washington International Nuclear Information System (INIS) 1992-12-01 The first section of this volume summarizes the content of the draft environmental impact statement (DEIS) and this Addendum, which together constitute the final environmental impact statement (FEIS) prepared on the decommissioning of eight surplus plutonium production reactors at Hanford. The FEIS consists of two volumes. The first volume is the DEIS as written. The second volume (this Addendum) consists of a summary; Chapter 9, which contains comments on the DEIS and provides DOE's responses to the comments; Appendix F, which provides additional health effects information; Appendix K, which contains costs of decommissioning in 1990 dollars; Appendix L, which contains additional graphite leaching data; Appendix M, which contains a discussion of accident scenarios; Appendix N, which contains errata; and Appendix 0, which contains reproductions of the letters, transcripts, and exhibits that constitute the record for the public comment period 13. Milk production and distribution in low-dose counties for the Hanford Thyroid Disease Study International Nuclear Information System (INIS) Schimmel, J.G. 1992-06-01 This report identifies sources of milk consumed by residents of Ferry, Okanogan, and Stevens Counties. This information will be used by the Hanford thyroid Disease Study to determine whether thyroid disease has been increased among people exposed to past iodine--131 emissions from Hanford Site Facilities 14. Legend and legacy: Fifty years of defense production at the Hanford Site Energy Technology Data Exchange (ETDEWEB) Gerber, M.S. 1992-09-01 Today, the Hanford Site is engaged in the largest waste cleanup effort ever undertaken in human history. That in itself makes the endeavor historic and unique. The Hanford Site has been designated the flagship of Department of Energy (DOE) waste remediation endeavors. And, just as the wartime Hanford Project remains unmatched in history, no counterpart exists for the current waste cleanup enterprise. This report provides a summary of the extensive historical record, however, which does give a partial road map. The science of environmental monitoring pioneered at the Hanford Site, and records of this type are the most complete of any in the world, from private companies or public agencies, for the early years of Site operations. The Hanford Site was unique for establishing a detailed, scientific, and multi-faceted environmental monitoring program. 15. Legend and legacy: Fifty years of defense production at the Hanford Site International Nuclear Information System (INIS) Gerber, M.S. 1992-09-01 Today, the Hanford Site is engaged in the largest waste cleanup effort ever undertaken in human history. That in itself makes the endeavor historic and unique. The Hanford Site has been designated the ''flagship'' of Department of Energy (DOE) waste remediation endeavors. And, just as the wartime Hanford Project remains unmatched in history, no counterpart exists for the current waste cleanup enterprise. This report provides a summary of the extensive historical record, however, which does give a partial road map. The science of environmental monitoring pioneered at the Hanford Site, and records of this type are the most complete of any in the world, from private companies or public agencies, for the early years of Site operations. The Hanford Site was unique for establishing a detailed, scientific, and multi-faceted environmental monitoring program 16. SAFETY AT FLUOR HANFORD (A) CASE STUDY - PREPARED BY THUNDERBIRD SCHOOL OF GLOBAL MANAGEMENT Energy Technology Data Exchange (ETDEWEB) ARNOLD LD 2009-09-25 By November of 1997, Fluor Hanford (Fluor) had been the site manager of the Hanford nuclear reservation for a year. The Hanford site had been established as part of the Manhattan Project in the 1940s that gave birth to the atomic bomb. Hanford produced two thirds of U.S. plutonium during the Cold War period. The Hanford site was half the size of Rhode Island and occupied 586 square miles in southeastern Washington State. The production of plutonium for more than 40 years left a huge legacy of chemical and radiological contamination: 80 square miles of contaminated groundwater; 2,300 tons of spent nuclear fuel stored in underwater basins; 20 tons of plutonium-laced contaminated materials; and 500 contaminated facilities. The cleanup involved a challenging combination of radioactive material handling within an infrastructure constructed in the 1940s and 1950s. The cleanup that began in 1988 was expected to take 30 years or more. Improving safety at Hanford had already proven to be a significant challenge. As the new site manager at Hanford, Fluor Hanford inherited lower- and mid-level managers and thousands of unionized employees, many of whom were second or third generation Hanford employees. These employees had seen many contractors come and go over the years. Some of the managers who had worked with the previous contractor saw Fluor's emphasis on safety as getting in the way of operations. Union-management relations were fractious. Hanford's culture was described as 'production driven-management told everyone what to do, and, if you didn't do it, there were consequences'. Worker involvement in designing and implementing safety programs was negligible. Fluor Hanford also was having trouble satisfying its client, the Department of Energy (DOE). The DOE did not see a clear path forward for performance improvements at Hanford. Clearly, major change was necessary, but how and where should it be implemented? 17. SAFETY AT FLUOR HANFORD (A) CASE STUDY - PREPARED BY THUNDERBIRD SCHOOL OF GLOBAL MANAGEMENT International Nuclear Information System (INIS) Arnold, L.D. 2009-01-01 By November of 1997, Fluor Hanford (Fluor) had been the site manager of the Hanford nuclear reservation for a year. The Hanford site had been established as part of the Manhattan Project in the 1940s that gave birth to the atomic bomb. Hanford produced two thirds of U.S. plutonium during the Cold War period. The Hanford site was half the size of Rhode Island and occupied 586 square miles in southeastern Washington State. The production of plutonium for more than 40 years left a huge legacy of chemical and radiological contamination: 80 square miles of contaminated groundwater; 2,300 tons of spent nuclear fuel stored in underwater basins; 20 tons of plutonium-laced contaminated materials; and 500 contaminated facilities. The cleanup involved a challenging combination of radioactive material handling within an infrastructure constructed in the 1940s and 1950s. The cleanup that began in 1988 was expected to take 30 years or more. Improving safety at Hanford had already proven to be a significant challenge. As the new site manager at Hanford, Fluor Hanford inherited lower- and mid-level managers and thousands of unionized employees, many of whom were second or third generation Hanford employees. These employees had seen many contractors come and go over the years. Some of the managers who had worked with the previous contractor saw Fluor's emphasis on safety as getting in the way of operations. Union-management relations were fractious. Hanford's culture was described as 'production driven-management told everyone what to do, and, if you didn't do it, there were consequences'. Worker involvement in designing and implementing safety programs was negligible. Fluor Hanford also was having trouble satisfying its client, the Department of Energy (DOE). The DOE did not see a clear path forward for performance improvements at Hanford. Clearly, major change was necessary, but how and where should it be implemented? 18. Hanford immobilized LAW product acceptance: Initial Tanks Focus Area testing data package Energy Technology Data Exchange (ETDEWEB) JD Vienna; A Jiricka; BP McGrail; BM Jorgensen; DE Smith; BR Allen; JC Marra; DK Peeler; KG Brown; IA Reamer; WL Ebert 2000-03-08 The Hanford Site's mission has been to produce nuclear materials for the US Department of Energy (DOE) and its predecessors. A large inventory of radioactive and mixed waste, largely generated during plutonium production, exists in 177 underground single- and double-shell tanks. These wastes are to be retrieved and separated into low-activity waste (LAW) and high-level waste (HLW) fractions. The total volume of LAW requiring immobilization will include the LAW separated from the tank waste, as well as new wastes generated by the retrieval, pretreatment, and immobilization processes. Per the Tri-Party Agreement (1994), both the LAW and HLW will be vitrified. It has been estimated that vitrification of the LAW waste will result in over 500,000 metric tons or 200,000 m{sup 3} of immobilized LAW (ILAW) glass. The ILAW glass is to be disposed of onsite in a near-surface burial facility. It must be demonstrated that the disposal system will adequately retain the radionuclides and prevent contamination of the surrounding environment. This report describes a study of the impacts of systematic glass-composition variation on the responses from accelerated laboratory corrosion tests of representative LAW glasses. A combination of two tests, the product consistency test and vapor-hydration test, is being used to give indictations of the relative rate at which a glass could be expected to corrode in the burial scenario. 19. Hanford immobilized LAW product acceptance: Initial Tanks Focus Area testing data package International Nuclear Information System (INIS) JD Vienna; A Jiricka; BP McGrail; BM Jorgensen; DE Smith; BR Allen; JC Marra; DK Peeler; KG Brown; IA Reamer; WL Ebert 2000-01-01 The Hanford Site's mission has been to produce nuclear materials for the US Department of Energy (DOE) and its predecessors. A large inventory of radioactive and mixed waste, largely generated during plutonium production, exists in 177 underground single- and double-shell tanks. These wastes are to be retrieved and separated into low-activity waste (LAW) and high-level waste (HLW) fractions. The total volume of LAW requiring immobilization will include the LAW separated from the tank waste, as well as new wastes generated by the retrieval, pretreatment, and immobilization processes. Per the Tri-Party Agreement (1994), both the LAW and HLW will be vitrified. It has been estimated that vitrification of the LAW waste will result in over 500,000 metric tons or 200,000 m 3 of immobilized LAW (ILAW) glass. The ILAW glass is to be disposed of onsite in a near-surface burial facility. It must be demonstrated that the disposal system will adequately retain the radionuclides and prevent contamination of the surrounding environment. This report describes a study of the impacts of systematic glass-composition variation on the responses from accelerated laboratory corrosion tests of representative LAW glasses. A combination of two tests, the product consistency test and vapor-hydration test, is being used to give indictations of the relative rate at which a glass could be expected to corrode in the burial scenario 20. Survey of Technetium Analytical Production Methods Supporting Hanford Nuclear Materials Processing International Nuclear Information System (INIS) TROYER, G.L. 1999-01-01 This document provides a historical survey of analytical methods used for measuring 99 Tc in nuclear fuel reprocessing materials and wastes at Hanford. Method challenges including special sludge matrices tested are discussed. Special problems and recommendations are presented 1. Phenomenology and modeling of particulate corrosion product behavior in Hanford N Reactor primary coolant International Nuclear Information System (INIS) Bechtold, D.B. 1983-01-01 The levels and composition of filterable corrosion products in the Hanford N Reactor Primary Loop are measurable by filtration. The suspended crud level has ranged from 0.0005 ppM to 6.482 ppM with a median 0.050 ppM. The composition approximates magnetite. The particle size distribution has been found in 31 cases to be uniformly a log normal distribution with a count median ranging from 1.10 to 2.31 microns with a median of 1.81 microns, and the geometric standard deviation ranging from 1.60 to 2.34 with a median of 1.84. An auto-correcting inline turbidimeter was found to respond to linearly to suspended crud levels over a range 0.05 to at least 6.5 ppM by direct comparison with filter sample weights. Cause of crud bursts in the primary loop were found to be power decreases. The crud transients associated with a reactor power drop, several reactor shutdowns, and several reactor startups could be modeled consistently with each other using a simple stirred-tank, first order exchange model of particulate between makeup, coolant, letdown, and loosely adherent crud on pipe walls. Over 3/10 of the average steady running particulate crud level could be accounted for by magnetically filterable particulate in the makeup feed. A simulation model of particulate transport has been coded in FORTRAN 2. Uncertainties in source term calculations generated by the ORIGEN2 computer code for Hanford Production Reactors International Nuclear Information System (INIS) Heeb, C.M. 1991-03-01 The ORIGEN2 computer code is the primary calculational tool for computing isotopic source terms for the Hanford Environmental Dose Reconstruction (HEDR) Project. The ORIGEN2 code computes the amounts of radionuclides that are created or remain in spent nuclear fuel after neutron irradiation and radioactive decay have occurred as a result of nuclear reactor operation. ORIGEN2 was chosen as the primary code for these calculations because it is widely used and accepted by the nuclear industry, both in the United States and the rest of the world. Its comprehensive library of over 1,600 nuclides includes any possible isotope of interest to the HEDR Project. It is important to evaluate the uncertainties expected from use of ORIGEN2 in the HEDR Project because these uncertainties may have a pivotal impact on the final accuracy and credibility of the results of the project. There are three primary sources of uncertainty in an ORIGEN2 calculation: basic nuclear data uncertainty in neutron cross sections, radioactive decay constants, energy per fission, and fission product yields; calculational uncertainty due to input data; and code uncertainties (i.e., numerical approximations, and neutron spectrum-averaged cross-section values from the code library). 15 refs., 5 figs., 5 tabs 3. Radionuclide inventory and source terms for the surplus production reactors at Hanford International Nuclear Information System (INIS) Miller, R.L.; Steffes, J.M. 1987-01-01 Radionuclide inventories have been estimated for the eight surplus production reactors at Hanford. The inventories listed represent more than 95% of the total curie burden; the remaining 5% is distributed in piping, tunnels, and various other locations within the reactor building and unaccounted for inventories within the reactors or fuel storage basins. Estimates are conservative as the methodology was designed to overestimate the radionuclide inventories in the facilities. The estimated inventory per reactor facility ranges from 13,000 curies to 58,000 curies. The majority of the present inventory consists of tritium, carbon-14, cobalt-60, and nickel-63. The information in this document combines data from past characterization efforts and introduces adjustments for added information and refinement. The inventory of hazardous materials in the reactor facilities is also addressed. This document has been revised to include new reduced inventory figures for chlorine-36. The new figures were derived from recent analysis of irradiated graphite from the 105-kW reactor 4. Hanford spent fuel inventory baseline International Nuclear Information System (INIS) Bergsman, K.H. 1994-01-01 This document compiles technical data on irradiated fuel stored at the Hanford Site in support of the Hanford SNF Management Environmental Impact Statement. Fuel included is from the Defense Production Reactors (N Reactor and the single-pass reactors; B, C, D, DR, F, H, KE and KW), the Hanford Fast Flux Test Facility Reactor, the Shipping port Pressurized Water Reactor, and small amounts of miscellaneous fuel from several commercial, research, and experimental reactors 5. Production of a High-Level Waste Glass from Hanford Waste Samples International Nuclear Information System (INIS) Crawford, C.L.; Farrara, D.M.; Ha, B.C.; Bibler, N.E. 1998-09-01 The HLW glass was produced from a HLW sludge slurry (Envelope D Waste), eluate waste streams containing high levels of Cs-137 and Tc-99, solids containing both Sr-90 and transuranics (TRU), and glass-forming chemicals. The eluates and Sr-90/TRU solids were obtained from ion-exchange and precipitation pretreatments, respectively, of other Hanford supernate samples (Envelopes A, B and C Waste). The glass was vitrified by mixing the different waste streams with glass-forming chemicals in platinum/gold crucibles and heating the mixture to 1150 degree C. Resulting glass analyses indicated that the HLW glass waste form composition was close to the target composition. The targeted waste loading of Envelope D sludge solids in the HLW glass was 30.7 wt percent, exclusive of Na and Si oxides. Condensate samples from the off-gas condenser and off-gas dry-ice trap indicated that very little of the radionuclides were volatilized during vitrification. Microstructure analysis of the HLW glass using Scanning Electron Microscopy (SEM) and Energy Dispersive X-Ray Analysis (EDAX) showed what appeared to be iron spinel in the HLW glass. Further X-Ray Diffraction (XRD) analysis confirmed the presence of nickel spinel trevorite (NiFe2O4). These crystals did not degrade the leaching characteristics of the glass. The HLW glass waste form passed leach tests that included a standard 90 degree C Product Consistency Test (PCT) and a modified version of the United States Environmental Protection Agency Toxicity Characteristic Leaching Procedure (TCLP) 6. Commercial milk distribution profiles and production locations. Hanford Environmental Dose Reconstruction Project Energy Technology Data Exchange (ETDEWEB) Deonigi, D.E.; Anderson, D.M.; Wilfert, G.L. 1994-04-01 The Hanford Environmental Dose Reconstruction (HEDR) Project was established to estimate radiation doses that people could have received from nuclear operations at the Hanford Site since 1944. For this period iodine-131 is the most important offsite contributor to radiation doses from Hanford operations. Consumption of milk from cows that ate vegetation contaminated by iodine-131 is the dominant radiation pathway for individuals who drank milk (Napier 1992). Information has been developed on commercial milk cow locations and commercial milk distribution during 1945 and 1951. The year 1945 was selected because during 1945 the largest amount of iodine-131 was released from Hanford facilities in a calendar year (Heeb 1993); therefore, 1945 was the year in which an individual was likely to have received the highest dose. The year 1951 was selected to provide data for comparing the changes that occurred in commercial milk flows (i.e., sources, processing locations, and market areas) between World War II and the post-war period. To estimate the doses people could have received from this milk flow, it is necessary to estimate the amount of milk people consumed, the source of the milk, the specific feeding regime used for milk cows, and the amount of iodine-131 contamination deposited on feed. 7. Hanford wells International Nuclear Information System (INIS) McGhan, V.L.; Myers, D.A.; Damschen, D.W. 1976-03-01 The Hanford Reservation contains about 2100 wells constructed from pre-Hanford Works to the present. As of Jan. 1976, about 1800 wells still exist, 850 of which were drilled to the groundwater table; 700 still contain water. This report provides the most complete documentation of these wells and supersedes all previous compilations, including BNWL-1739 8. Characterization of stored defense production spent nulcear fuel and associated materials at Hanford Site, Richland Washington: Environmental assessment International Nuclear Information System (INIS) 1995-03-01 There are about 2,100 tonnes (2,300 tons) of defense production spent nuclear fuel stored in the 100-K Area Basins located along the south shore of the Columbia River in the northern part of the Hanford Site. Some of the fuel which has been in storage for a number of years is in poor condition and continues to deteriorate. The basins also contain fuel fragments and radioactively contaminated sludge. The DOE needs to characterize defense production spent nuclear fuel and associated materials stored on the Hanford Site. In order to satisfy that need, the Department of Energy (DOE) proposes to select, collect and transport samples of spent nuclear fuel and associated materials to the 327 Building for characterization. As a result of that characterization, modes of interim storage can be determined that would be compatible with the material in its present state and alternative treatment processes could be developed to permit a broader selection of storage modes. Environmental impacts of the proposed action were determined to be limited principally to radiation exposure of workers, which, however, were found to be small. No health effects among workers or the general public would be expected under routine operations. Implementation of the proposed action would not result in any impacts on cultural resources, threatened, endangered and candidate species, air or water quality, socioeconomic conditions, or waste management 9. Hanford wells International Nuclear Information System (INIS) Chamness, M.A.; Merz, J.K. 1993-08-01 Records describing wells located on or near the Hanford Site have been maintained by Pacific Northwest Laboratory and the operating contractor, Westinghouse Hanford Company. In support of the Ground-Water Surveillance Project, portions of the data contained in these records have been compiled into the following report, which is intended to be used by those needing a condensed, tabular summary of well location and basic construction information. The wells listed in this report were constructed over a period of time spanning almost 70 years. Data included in this report were retrieved from the Hanford Envirorunental Information System (HEIS) database and supplemented with information not yet entered into HEIS. While considerable effort has been made to obtain the most accurate and complete tabulations possible of the Hanford Site wells, omissions and errors may exist. This document does not include data on lithologic logs, ground-water analyses, or specific well completion details 10. Mesonic atom production in high-energy nuclear collisions International Nuclear Information System (INIS) Wakai, M.; Bando, H.; Sano, M. 1987-08-01 The production probability of π-mesonic atom in high-energy nuclear collisions is estimated by a coalescence model. The production cross section is calculated for p + Ne and Ne + Ne systems at 2.1 GeV/A and 5.0 GeV/A beam energy. It is shown that nuclear fragments with larger charge numbers have the advantage in the formation of π-mesonic atoms. The cross section is proportional to Z 3 and of the order of magnitude of 1 ∼ 10 μb in all the above cases. The production cross sections of K-mesonic atoms are also estimated. (author) 11. The production and investigation of cold antihydrogen atoms International Nuclear Information System (INIS) Pittner, H. 2005-04-01 This work reports on experiments in which antihydrogen atoms have been produced in cryogenic Penning traps from antiproton and positron plasmas by two different methods and on experiments that have been carried out subsequently in order to investigate the antihydrogen atoms. By the first method antihydrogen atoms have been formed during the process of positron cooling of antiprotons in so called nested Penning traps and detected via a field ionization method. A measurement of the state distribution has revealed that the antihydrogen atoms are formed in highly excited states. This suggests along with the high production rate that the antihydrogen atoms are formed by three-body recombination processes and subsequent collisional deexcitations. However current theory cannot yet account for the measured state distribution. Typical radii of the detected antihydrogen atoms lie in the range between 0.4 μm and 0.15 μm. The deepest bound antihydrogen atoms have radii below 0.1 μm.The kinetic energy of the weakest bound antihydrogen atoms has been measured to about 200 meV. By the second method antihydrogen atoms have been synthesized in charge-exchange processes. Lasers are used to produce a Rydberg cesium beam within the cryogenic Penning trap that collides with trapped positrons so that Rydberg positronium atoms are formed via charge-exchange reactions. The Rydberg positronium atoms that collide with nearby stored antiprotons form antihydrogen atoms in charge-exchange reactions. So far, 14±4 antihydrogen atoms have been detected background-free via a field-ionization method. The antihydrogen atoms produced via the two-step charge-exchange mechanism are expected to have a temperature of 4.2 K, the temperature of the antiprotons from which they are formed 12. Production of hyperthermal hydrogen atoms by an arc discharge International Nuclear Information System (INIS) Samano, E.C. 1993-01-01 A magnetically confined thermal electric arc gas heater has been designed and built as a suitable source of heat for dissociating hydrogen molecules with energy in the range of a few eV. Specifically, the average beam kinetic energy is determined to be 1.5 eV, the dissociation rate is 0.5 atoms per molecule and the atom beam intensity in the forward direction is 1018 atoms/sr-sec. The working pressure in the arc discharge region is from 15 to 25 torr. This novel atom source has been successfully ignited and operated with pure hydrogen during several hours of continuous performance, maintaining its characteristics. The hyperthermal hydrogen atom beam, which is obtained from this source is analyzed and characterized in a high vacuum system, the characterization of the atom beam is accomplished by two different methods: calorimetry and surface ionization. Calorimetic sensor were used for detecting the atom beam by measuring the delivered power of the impinging atoms on the sensor surface. In the second approach an H-surface production backscattering experiment from a low work function surface was conducted. The validity of these two methods is discussed, and the results are compared. The different collision mechanisms to dissociate and ionize hydrogen molecules in the arch discharge are reviewed, as well as the physics of electric arcs. Finally, a Monte Carlo simulation program is used to calculate the ionization probability of low energy atoms perpendicularly reflected from a surface converter, as a model for atom surface ionization 13. Decommissioning of eight surplus production reactors at the Hanford Site, Richland, Washington. Addendum (Final Environmental Impact Statement) Energy Technology Data Exchange (ETDEWEB) 1992-12-01 The first section of this volume summarizes the content of the draft environmental impact statement (DEIS) and this Addendum, which together constitute the final environmental impact statement (FEIS) prepared on the decommissioning of eight surplus plutonium production reactors at Hanford. The FEIS consists of two volumes. The first volume is the DEIS as written. The second volume (this Addendum) consists of a summary; Chapter 9, which contains comments on the DEIS and provides DOEs responses to the comments; Appendix F, which provides additional health effects information; Appendix K, which contains costs of decommissioning in 1990 dollars; Appendix L, which contains additional graphite leaching data; Appendix M, which contains a discussion of accident scenarios; Appendix N, which contains errata; and Appendix 0, which contains reproductions of the letters, transcripts, and exhibits that constitute the record for the public comment period. 14. Explosion of cation exchange column in americium recovery service, Hanford plant, August 30, 1976 International Nuclear Information System (INIS) 1976-01-01 This document is a collection of thirty references related to the explosion of the cation exchange column in the Americium Recovery Service of the Hanford Atomic Products Operation, Richland, Washington, on August 30, 1976. Some of the documents are related to the design and safety studies, while others refer to the accident and resulting decontamination efforts, investigations, and legal consequences 15. Atomic spectral-product representations of molecular electronic structure: metric matrices and atomic-product composition of molecular eigenfunctions. Science.gov (United States) Ben-Nun, M; Mills, J D; Hinde, R J; Winstead, C L; Boatz, J A; Gallup, G A; Langhoff, P W 2009-07-02 Recent progress is reported in development of ab initio computational methods for the electronic structures of molecules employing the many-electron eigenstates of constituent atoms in spectral-product forms. The approach provides a universal atomic-product description of the electronic structure of matter as an alternative to more commonly employed valence-bond- or molecular-orbital-based representations. The Hamiltonian matrix in this representation is seen to comprise a sum over atomic energies and a pairwise sum over Coulombic interaction terms that depend only on the separations of the individual atomic pairs. Overall electron antisymmetry can be enforced by unitary transformation when appropriate, rather than as a possibly encumbering or unnecessary global constraint. The matrix representative of the antisymmetrizer in the spectral-product basis, which is equivalent to the metric matrix of the corresponding explicitly antisymmetric basis, provides the required transformation to antisymmetric or linearly independent states after Hamiltonian evaluation. Particular attention is focused in the present report on properties of the metric matrix and on the atomic-product compositions of molecular eigenstates as described in the spectral-product representations. Illustrative calculations are reported for simple but prototypically important diatomic (H(2), CH) and triatomic (H(3), CH(2)) molecules employing algorithms and computer codes devised recently for this purpose. This particular implementation of the approach combines Slater-orbital-based one- and two-electron integral evaluations, valence-bond constructions of standard tableau functions and matrices, and transformations to atomic eigenstate-product representations. The calculated metric matrices and corresponding potential energy surfaces obtained in this way elucidate a number of aspects of the spectral-product development, including the nature of closure in the representation, the general redundancy or 16. Production and detection of atomic hexadecapole at Earth's magnetic field. Science.gov (United States) Acosta, V M; Auzinsh, M; Gawlik, W; Grisins, P; Higbie, J M; Jackson Kimball, D F; Krzemien, L; Ledbetter, M P; Pustelny, S; Rochester, S M; Yashchuk, V V; Budker, D 2008-07-21 Optical magnetometers measure magnetic fields with extremely high precision and without cryogenics. However, at geomagnetic fields, important for applications from landmine removal to archaeology, they suffer from nonlinear Zeeman splitting, leading to systematic dependence on sensor orientation. We present experimental results on a method of eliminating this systematic error, using the hexadecapole atomic polarization moment. In particular, we demonstrate selective production of the atomic hexadecapole moment at Earth's magnetic field and verify its immunity to nonlinear Zeeman splitting. This technique promises to eliminate directional errors in all-optical atomic magnetometers, potentially improving their measurement accuracy by several orders of magnitude. 17. Hanford land disposal restrictions plan for mixed wastes International Nuclear Information System (INIS) 1990-10-01 Since the early 1940s, the Hanford Site has been involved in the production and purification of nuclear defense materials. These production activities have resulted in the generation of large quantities of liquid and solid radioactive mixed waste. This waste is subject to regulation under authority of both the Resource Conservation and Recovery Act of 1976 (RCRA) and the Atomic Energy Act. The State of Washington Department of Ecology (Ecology), the US Environmental Protection Agency (EPA), and the US Department of Energy (DOE) have entered into an agreement, the Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement) to bring Hanford Site Operations into compliance with dangerous waste regulations. The Tri-Party Agreement was amended to require development of the Hanford Land Disposal Restrictions Plan for Mixed Wastes (this plan) to comply with land disposal restrictions requirements for radioactive mixed waste. The Tri-Party Agreement requires, and the this plan provides, the following sections: Waste Characterization Plan, Storage Report, Treatment Report, Treatment Plan, Waste Minimization Plan, a schedule, depicting the events necessary to achieve full compliance with land disposal restriction requirements, and a process for establishing interim milestones. 34 refs., 28 figs., 35 tabs 18. Hanford land disposal restrictions plan for mixed wastes Energy Technology Data Exchange (ETDEWEB) 1990-10-01 Since the early 1940s, the Hanford Site has been involved in the production and purification of nuclear defense materials. These production activities have resulted in the generation of large quantities of liquid and solid radioactive mixed waste. This waste is subject to regulation under authority of both the Resource Conservation and Recovery Act of 1976 (RCRA) and the Atomic Energy Act. The State of Washington Department of Ecology (Ecology), the US Environmental Protection Agency (EPA), and the US Department of Energy (DOE) have entered into an agreement, the Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement) to bring Hanford Site Operations into compliance with dangerous waste regulations. The Tri-Party Agreement was amended to require development of the Hanford Land Disposal Restrictions Plan for Mixed Wastes (this plan) to comply with land disposal restrictions requirements for radioactive mixed waste. The Tri-Party Agreement requires, and the this plan provides, the following sections: Waste Characterization Plan, Storage Report, Treatment Report, Treatment Plan, Waste Minimization Plan, a schedule, depicting the events necessary to achieve full compliance with land disposal restriction requirements, and a process for establishing interim milestones. 34 refs., 28 figs., 35 tabs. 19. Introduction to the Hanford Site Energy Technology Data Exchange (ETDEWEB) Cushing, C.E. 1995-06-01 This section of the 1994 Hanford Site Environmental Report discusses the Site mission and provides general information about the site. The U.S. DOE has established a new mission for Hanford including: Management of stored wastes, environmental restoration, research and development, and development of new technologies. The Hanford Reservation is located in south central Washington State just north of the confluence of the Snake and Yakima Rivers with the Columbia River. The approximately 1,450 square kilometers which comprises the Hanford Site, with restricted public access, provides a buffer for the smaller areas within the site which have historically been used for the production of nuclear materials, radioactive waste storage, and radioactive waste disposal. 20. Introduction to the Hanford Site International Nuclear Information System (INIS) Cushing, C.E. 1995-01-01 This section of the 1994 Hanford Site Environmental Report discusses the Site mission and provides general information about the site. The U.S. DOE has established a new mission for Hanford including: Management of stored wastes, environmental restoration, research and development, and development of new technologies. The Hanford Reservation is located in south central Washington State just north of the confluence of the Snake and Yakima Rivers with the Columbia River. The approximately 1,450 square kilometers which comprises the Hanford Site, with restricted public access, provides a buffer for the smaller areas within the site which have historically been used for the production of nuclear materials, radioactive waste storage, and radioactive waste disposal 1. Chemical composition analysis and product consistency tests to support enhanced Hanford waste glass models: Results for the January, March, and April 2015 LAW glasses Energy Technology Data Exchange (ETDEWEB) Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Riley, W. T. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Best, D. R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL) 2015-09-03 In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for several simulated low activity waste (LAW) glasses (designated as the January, March, and April 2015 LAW glasses) fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions. 2. Chemical composition analysis and product consistency tests to support Enhanced Hanford Waste Glass Models. Results for the Augusta and October 2014 LAW Glasses Energy Technology Data Exchange (ETDEWEB) Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Best, D. R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL) 2015-07-07 In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for several simulated low activity waste (LAW) glasses (designated as the August and October 2014 LAW glasses) fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions. 3. Annual Hanford seismic report - fiscal year 1996 International Nuclear Information System (INIS) Hartshorn, D.C.; Reidel, S.P. 1996-12-01 Seismic monitoring (SM) at the Hanford Site was established in 1969 by the US Geological Survey (USGS) under a contract with the US Atomic Energy Commission. Since 1980, the program has been managed by several contractors under the US Department of Energy (USDOE). Effective October 1, 1996, the Seismic Monitoring workscope, personnel, and associated contracts were transferred to the USDOE Pacific Northwest National Laboratory (PNNL). SM is tasked to provide an uninterrupted collection and archives of high-quality raw and processed seismic data from the Hanford Seismic Network (HSN) located on and encircling the Hanford Site. SM is also tasked to locate and identify sources of seismic activity and monitor changes in the historical pattern of seismic activity at the Hanford Site. The data compiled are used by SM, Waste Management, and engineering activities at the Hanford Site to evaluate seismic hazards and seismic design for the Site 4. Fluor Hanford Project Focused Progress at Hanford International Nuclear Information System (INIS) HANSON, R.D. 2000-01-01 Fluor Hanford is making significant progress in accelerating cleanup at the Hanford site. This progress consistently aligns with a new strategic vision established by the U.S. Department of Energy's Richland Operations Office (RL) 5. Pollution prevention opportunity assessments at Hanford Energy Technology Data Exchange (ETDEWEB) Betsch, M.D., Westinghouse Hanford 1996-06-26 The Pollution Prevention Opportunity Assessment (PPOA) is a pro- active way to look at a waste generating activity and identify opportunities to minimize wastes through a cost benefit analysis. Hanfords PPOA process is based upon the graded approach developed by the Kansas City Plant. Hanford further streamlined the process while building in more flexibility for the individual users. One of the most challenging aspects for implementing the PPOA process at Hanford is one overall mission which is environmental restoration, Now that the facilities are no longer in production, each has a different non- routine activity making it difficult to quantify the inputs and outputs of the activity under consideration. 6. Historical genesis of Hanford Site wastes International Nuclear Information System (INIS) Gerber, M.S. 1991-01-01 This paper acquaints the audience with historical waste practices and policies as they changed over the years at the Hanford Site, and with the generation of the major waste streams of concern in Hanford Site clean-up today. The paper also describes the founding and basic operating history of the Hanford Site, including World War 11 construction and operations, three major postwar expansions (1947-55), the peak years of production (1956-63), production phase downs (1964-the present), and some past suggestions and efforts to chemically treat, open-quotes fractionate,close quotes and/or immobilize Hanford's wastes. Recent events, including the designation of the Hanford Site as the open-quotes flagshipclose quotes of Department of Energy (DOE) waste remediation efforts and the signing of the landmark Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement), have generated new interest in Hanford's history. Clean-up milestones dictated in this agreement demand information about how, when, in what quantities and mixtures, and under what conditions, Hanford Site wastes were generated and released. This paper presents original, primary-source research into the waste history of the Hanford Site. The earliest, 1940s knowledge base, assumptions and calculations about radioactive and chemical discharges, as discussed in the memos, correspondence and reports of the original Hanford Site (then Hanford Engineer Works) builders and operators, are reviewed. The growth of knowledge, research efforts, and subsequent changes in Site waste disposal policies and practices are traced. Finally, the paper places the current Hanford Site waste remediation endeavors in the broad context of American and world history 7. MANHATTAN PROJECT B REACTOR HANFORD WASHINGTON [HANFORD'S HISTORIC B REACTOR (12-PAGE BOOKLET) Energy Technology Data Exchange (ETDEWEB) GERBER MS 2009-04-28 The Hanford Site began as part of the United States Manhattan Project to research, test and build atomic weapons during World War II. The original 670-square mile Hanford Site, then known as the Hanford Engineer Works, was the last of three top-secret sites constructed in order to produce enriched uranium and plutonium for the world's first nuclear weapons. B Reactor, located about 45 miles northwest of Richland, Washington, is the world's first full-scale nuclear reactor. Not only was B Reactor a first-of-a-kind engineering structure, it was built and fully functional in just 11 months. Eventually, the shoreline of the Columbia River in southeastern Washington State held nine nuclear reactors at the height of Hanford's nuclear defense production during the Cold War era. The B Reactor was shut down in 1968. During the 1980's, the U.S. Department of Energy began removing B Reactor's support facilities. The reactor building, the river pumphouse and the reactor stack are the only facilities that remain. Today, the U.S. Department of Energy (DOE) Richland Operations Office offers escorted public access to B Reactor along a designated tour route. The National Park Service (NPS) is studying preservation and interpretation options for sites associated with the Manhattan Project. A draft is expected in summer 2009. A final report will recommend whether the B Reactor, along with other Manhattan Project facilities, should be preserved, and if so, what roles the DOE, the NPS and community partners will play in preservation and public education. In August 2008, the DOE announced plans to open B Reactor for additional public tours. Potential hazards still exist within the building. However, the approved tour route is safe for visitors and workers. DOE may open additional areas once it can assure public safety by mitigating hazards. 8. Production of dimeson atoms in high-energy collisions Energy Technology Data Exchange (ETDEWEB) Afanasyev, L.; Gevorkyan, S.; Voskresenskaya, O. [Joint Institute for Nuclear Research, Dubna (Russian Federation) 2017-04-15 The production of two-meson electromagnetic bound states and free meson pairs π{sup +}π{sup -}, K{sup +}K{sup -}, π{sup +}K{sup -+} in relativistic collisions has been considered. It is shown that using of exact Coulomb wave functions for dimeson atom (DMA) allows one to calculate the yield of discrete states with the desired accuracy. The relative probabilities of production of DMA and meson pairs in the free state are estimated. The amplitude of DMA transition from 1S to 2P state, which is essential for the pionium Lamb shift measurements, has been obtained. (orig.) 9. The Wahluke (North) Slope of the Hanford Site: History and present challenges International Nuclear Information System (INIS) Gerber, M.S. 1996-01-01 The Hanford Site was founded in early 1943 for the top secret government mission of producing plutonium for the world's first atomic weapons. A great deal of land was needed, both to separate various Site facilities from each other, and to provide buffer zones for safety and security purposes. In total, 640 square miles were occupied by the original Hanford Site and its buffer zones. Much of this land had been earmarked for inclusion in the Columbia Basin Irrigation Project (CBP). After World War II ended, a series of national decisions led to a long-term mission for the Hanford Site, and area residents learned that the Site lands they had hoped to farm would be withheld from agricultural production for the foreseeable future. A long set of negotiations commenced between the federal management agency responsible for Hanford (the Atomic Energy Commission -- AEC), and the Bureau of Reclamation (BOR), Department of the Interior that managed the CBP. Some lands were turned back to agriculture, and other compromises made, in the Site's far northern buffer lands known as the Wahluke Slope, during the 1950s. In the mid-1960s, further negotiations were about to allow farming on lands just north of the Columbia River, opposite Hanford's reactors, when studies conducted by the BOR found drainage barriers to irrigation. As a result of these findings, two wildlife refuges were created on that land in 1971. Today, after the Hanford Site plutonium production mission has ended and as Site cleanup goes forward, the possibility of total release of Wahluke Slope lands from the control of the Department of Energy (DOE -- a successor agency to the AEC) is under discussion. Such discussion encompasses not just objective and clearly visible criteria, but it resurrects historical debates about the roles of farming and government presence in the Columbia Basin 10. TECHNICAL ASSESSMENT OF BULK VITRIFICATION PROCESS/ PRODUCT FOR TANK WASTE TREATMENT AT THE DEPARTMENT OF ENERGY HANFORD SITE International Nuclear Information System (INIS) SCHAUS, P.S. 2006-01-01 At the U.S. Department of Energy (DOE) Hanford Site, the Waste Treatment Plant (WTP) is being constructed to immobilize both high-level waste (IUW) for disposal in a national repository and low-activity waste (LAW) for onsite, near-surface disposal. The schedule-controlling step for the WTP Project is vitrification of the large volume of LAW, current capacity of the WTP (as planned) would require 50 years to treat the Hanford tank waste, if the entire LAW volume were to be processed through the WTP. To reduce the time and cost for treatment of Hanford Tank Waste, and as required by the Tank Waste Remediation System Environmental Impact Statement Record of Decision and the Hanford Federal Facility Consent Agreement (Tn-Party Agreement), DOE plans to supplement the LAW treatment capacity of the WTP. Since 2002, DOE, in cooperation with the Environmental Protection Agency and State of Washington Department of Ecology has been evaluating technologies that could provide safe and effective supplemental treatment of LAW. Current efforts at Hanford are intended to provide additional information to aid a joint agency decision on which technology will be used to supplement the WTP. A Research, Development and Demonstration permit has been issued by the State of Washington to build and (for a limited time) operate a Demonstration Bulk Vitrification System (DBVS) facility to provide information for the decision on a supplemental treatment technology for up to 50% of the LAW. In the Bulk Vitrification (BV) process, LAW, soil, and glass-forming chemicals are mixed, dried, and placed in a refractory-lined box, Electric current, supplied through two graphite electrodes in the box, melts the waste feed, producing a durable glass waste-form. Although recent modifications to the process have resulted in significant improvements, there are continuing technical concerns 11. TECHNICAL ASSESSMENT OF BULK VITRIFICATION PROCESS & PRODUCT FOR TANK WASTE TREATMENT AT THE DEPARTMENT OF ENERGY HANFORD SITE Energy Technology Data Exchange (ETDEWEB) SCHAUS, P.S. 2006-07-21 At the U.S. Department of Energy (DOE) Hanford Site, the Waste Treatment Plant (WTP) is being constructed to immobilize both high-level waste (IUW) for disposal in a national repository and low-activity waste (LAW) for onsite, near-surface disposal. The schedule-controlling step for the WTP Project is vitrification of the large volume of LAW, current capacity of the WTP (as planned) would require 50 years to treat the Hanford tank waste, if the entire LAW volume were to be processed through the WTP. To reduce the time and cost for treatment of Hanford Tank Waste, and as required by the Tank Waste Remediation System Environmental Impact Statement Record of Decision and the Hanford Federal Facility Consent Agreement (Tn-Party Agreement), DOE plans to supplement the LAW treatment capacity of the WTP. Since 2002, DOE, in cooperation with the Environmental Protection Agency and State of Washington Department of Ecology has been evaluating technologies that could provide safe and effective supplemental treatment of LAW. Current efforts at Hanford are intended to provide additional information to aid a joint agency decision on which technology will be used to supplement the WTP. A Research, Development and Demonstration permit has been issued by the State of Washington to build and (for a limited time) operate a Demonstration Bulk Vitrification System (DBVS) facility to provide information for the decision on a supplemental treatment technology for up to 50% of the LAW. In the Bulk Vitrification (BV) process, LAW, soil, and glass-forming chemicals are mixed, dried, and placed in a refractory-lined box, Electric current, supplied through two graphite electrodes in the box, melts the waste feed, producing a durable glass waste-form. Although recent modifications to the process have resulted in significant improvements, there are continuing technical concerns. 12. Maximizing Production Capacity from an Ultrafiltration Process at the Hanford Department of Waste Treatment Facility International Nuclear Information System (INIS) Foust, Henry C.; Holton, Langdon K.; Demick, Laurence E. 2005-01-01 The Department of Energy has contracted Bechtel National, Inc. to design, construct and commission a Waste Treatment and Immobilization Plant (WTP) to treat radioactive slurry currently stored in underground waste storage tanks. A critical element of the waste treatment capacity for the WTP is the proper operation of an ultrafiltration process (UFP). The UFP separates supernate solution from radioactive solids. The solution and solid phases are separately immobilized. An oversight review of the UFP design and operation has identified several methods to improve the capacity of the ultrafiltration process, which will also improve the capacity of the WTP. Areas explored were the basis of design, an analysis of the WTP capacity, process chemistry within the UFP, and UFP process control. This article discusses some of the findings of this oversight review in terms of sodium and solid production, which supports the treatment of low activity waste (LAW) associated with the facility, and solid production, which supports the treatment of high level waste (HLW) associated with the facility 13. Fission product iodine during early Hanford-Site operations: Its production and behavior during fuel processing, off-gas treatment and release to the atmosphere International Nuclear Information System (INIS) Burger, L.L. 1991-05-01 The Hanford Environmental Dose Reconstruction (HEDR) Project was established to estimate the radiological dose impact that Hanford Site operations may have made on the local and regional population. This impact is estimated by examining operations involving radioactive materials that were conducted at the Hanford Site from the startup of the first reactor in 1944 to the present. HEDR Project work is divided among several technical tasks. One of these tasks, Source Terms, is designed to develop quantitative estimates of all significant emissions of radionuclides by Hanford Site operations since 1944. Radiation doses can be estimated from these emissions by accounting for specific radionuclide transport conditions and population demography. This document provides technical information to assist in the evaluation of iodine releases. 115 refs., 5 figs., 3 tabs 14. Fission product iodine during early Hanford-Site operations: Its production and behavior during fuel processing, off-gas treatment and release to the atmosphere Energy Technology Data Exchange (ETDEWEB) Burger, L.L. 1991-05-01 The Hanford Environmental Dose Reconstruction (HEDR) Project was established to estimate the radiological dose impact that Hanford Site operations may have made on the local and regional population. This impact is estimated by examining operations involving radioactive materials that were conducted at the Hanford Site from the startup of the first reactor in 1944 to the present. HEDR Project work is divided among several technical tasks. One of these tasks, Source Terms, is designed to develop quantitative estimates of all significant emissions of radionuclides by Hanford Site operations since 1944. Radiation doses can be estimated from these emissions by accounting for specific radionuclide transport conditions and population demography. This document provides technical information to assist in the evaluation of iodine releases. 115 refs., 5 figs., 3 tabs. 15. Public involvement in environmental surveillance at Hanford International Nuclear Information System (INIS) Hanf, R.W. Jr.; Patton, G.W.; Woodruff, R.K.; Poston, T.M. 1994-08-01 Environmental surveillance at the Hanford Site began during the mid-1940s following the construction and start-up of the nation's first plutonium production reactor. Over the past approximately 45 years, surveillance operations on and off the Site have continued, with virtually all sampling being conducted by Hanford Site workers. Recently, the US Department of Energy Richland Operations Office directed that public involvement in Hanford environmental surveillance operations be initiated. Accordingly, three special radiological air monitoring stations were constructed offsite, near hanford's perimeter. Each station is managed and operated by two local school teaches. These three stations are the beginning of a community-operated environmental surveillance program that will ultimately involve the public in most surveillance operations around the Site. The program was designed to stimulate interest in Hanford environmental surveillance operations, and to help the public better understand surveillance results. The program has also been used to enhance educational opportunities at local schools 16. Atoms International Nuclear Information System (INIS) Fuchs, Alain; Villani, Cedric; Guthleben, Denis; Leduc, Michele; Brenner, Anastasios; Pouthas, Joel; Perrin, Jean 2014-01-01 Completed by recent contributions on various topics (atoms and the Brownian motion, the career of Jean Perrin, the evolution of atomic physics since Jean Perrin, relationship between scientific atomism and philosophical atomism), this book is a reprint of a book published at the beginning of the twentieth century in which the author addressed the relationship between atomic theory and chemistry (molecules, atoms, the Avogadro hypothesis, molecule structures, solutes, upper limits of molecular quantities), molecular agitation (molecule velocity, molecule rotation or vibration, molecular free range), the Brownian motion and emulsions (history and general features, statistical equilibrium of emulsions), the laws of the Brownian motion (Einstein's theory, experimental control), fluctuations (the theory of Smoluchowski), light and quanta (black body, extension of quantum theory), the electricity atom, the atom genesis and destruction (transmutations, atom counting) 17. Productivity, mortality, and response to disturbance of nesting Swainson's hawks on the Hanford Site Energy Technology Data Exchange (ETDEWEB) Poole, L.D.; Marr, N.V.; McCorquodale, S.M. 1988-03-01 The objectives of this study were to characterize Swainson's hawk (Buteo swainson) use of the US Department of Energy (DOE) Hanford Site and to evaluate the potential for engineering and other human activities on the Hanford Site to negatively affect the nesting Swainson's hawk population. Activities associated with the Basalt Waste Isolation Project (BWIP) were used as the primary external stimuli in studying hawk responses to potential human disturbance. Parked and moving vehicles were the most common disturbance sources observed in Swainson's hawk territories. Hawks appeared to be sensitive to disturbance from pedestrians and slow-moving vehicles near nests. Novel stimuli were much more likely to evoke strong responses than were recurring events. Adult hawks reacted more frequently and vigorously than did juveniles. When disturbed, adult hawks usually flew toward the location of the disturbance; juvenile hawks usually flew away from disturbances. Human activity associated with BWIP may have had negative on one pair of nesting Swainson's hawks and may have precluded the use of an additional traditional nesting territory. Negative impacts to nesting Swainson's hawks from human activity could be minimized by confining activities to the non-nesting period or to distances greater than 2.2 km from nest sites. Tree groves and elevated perches, including utility poles, across the Hanford Site are probably critical to the success of nesting Swainson's hawks. Potential mitigation strategies associated with energy research and development activities on the Hanford Site could include provisions for maintenance and establishment of drought-tolerant trees and native vegetation. 22 refs., 5 figs., 3 tabs. 18. Vitrification of Hanford wastes in a joule-heated ceramic melter and evaluation of resultant canisterized product International Nuclear Information System (INIS) Chapman, C.C.; Buelt, J.L.; Slate, S.C.; Katayama, Y.B.; Bunnell, L.R. 1979-08-01 Experience gained in the week-long vitrification test and characterization of the glass produced in the run support the following conclusions: The Hanford waste simulated in this test can be readily vitrified in a joule-heated ceramic melter. Physical properties of the molten glass were entirely compatible with melter operation. The average feed rate of 106 kg/h is high enough to make the ceramic melter a feasible piece of equipment for vitrifying Hanford wastes. The glass produced in this trial had good chemical durability, 6(10) -5 g/cm 2 -d. When one of the canisters was purposely dropped onto a steel pad, the damage was limited to deformation of the steel can in the impact area, cracking of a weld, and fracturing of glass in the immediate vicinity of the impact area. No glass was released from the canister as a result of the drop test. The results of this vitrification test support the technical feasibility of vitrifying Hanford wastes by means of a joule-heated ceramic melter. Surface area for large glass castings is equivalent to the mass median particle diameters between 4.27 cm (1.75 in.) and 8.91 cm (3.51 in.) even when allowed to cool rapidly by standing in ambient air. Large canisters (up to 0.91 m in dia) can be cast without large voids while standing in air if the fill rate is over 100 kg/h. 34 figures, 10 tables 19. Hanford Site Development Plan International Nuclear Information System (INIS) Hathaway, H.B.; Daly, K.S.; Rinne, C.A.; Seiler, S.W. 1993-05-01 The Hanford Site Development Plan (HSDP) provides an overview of land use, infrastructure, and facility requirements to support US Department of Energy (DOE) programs at the Hanford Site. The HSDP's primary purpose is to inform senior managers and interested parties of development activities and issues that require a commitment of resources to support the Hanford Site. The HSDP provides an existing and future land use plan for the Hanford Site. The HSDP is updated annually in accordance with DOE Order 4320.1B, Site Development Planning, to reflect the mission and overall site development process. Further details about Hanford Site development are defined in individual area development plans 20. Atomic x-ray production by relativistic heavy ions International Nuclear Information System (INIS) Ioannou, J.G. 1977-12-01 The interaction of heavy ion projectiles with the electrons of target atoms gives rise to the production, in the target, of K-, L- or higher shell vacancies which are in turn followed by the emission of characteristic x-rays. The calculation of the theoretical value of the K- and L-shells vacancy production cross section was carried out for heavy ion projectiles of any energy. The transverse component of the cross section is calculated for the first time in detail and extensive tables of its numerical value as a function of its parameters are also given. Experimental work for 4.88 GeV protons and 3 GeV carbon ions is described. The K vacancy cross section has been measured for a variety of targets from Ti to U. The agreement between the theoretical predictions and experimental results for the 4.88 GeV protons is rather satisfactory. For the 3 GeV carbon ions, however, it is observed that the deviation of the theoretical and experimental values of the K vacancy production becomes larger with the heavier target element. Consequently, the simple scaling law of Z 1 2 for the cross section of the heavy ion with atomic number Z 1 to the proton cross section is not true, for the K-shell at least. A dependence on the atomic number Z 2 of the target of the form (Z 1 - αZ 2 ) 2 , instead of Z 1 2 , is found to give extremely good agreement between theory and experiment. Although the exact physical meaning of such dependence is not yet clearly understood, it is believed to be indicative of some sort of screening effect of the incoming fast projectile by the fast moving in Bohr orbits K-shell electrons of the target. The enhancement of the K-shell ionization cross section by relativistic heavy ions on heavy targets is also discussed in terms of its practical applications in various branches of science and technology 1. Hanford External Dosimetry Program International Nuclear Information System (INIS) Fix, J.J. 1990-10-01 This document describes the Hanford External Dosimetry Program as it is administered by Pacific Northwest Laboratory (PNL) in support of the US Department of Energy (DOE) and its Hanford contractors. Program services include administrating the Hanford personnel dosimeter processing program and ensuring that the related dosimeter data accurately reflect occupational dose received by Hanford personnel or visitors. Specific chapters of this report deal with the following subjects: personnel dosimetry organizations at Hanford and the associated DOE and contractor exposure guidelines; types, characteristics, and procurement of personnel dosimeters used at Hanford; personnel dosimeter identification, acceptance testing, accountability, and exchange; dosimeter processing and data recording practices; standard sources, calibration factors, and calibration processes (including algorithms) used for calibrating Hanford personnel dosimeters; system operating parameters required for assurance of dosimeter processing quality control; special dose evaluation methods applied for individuals under abnormal circumstances (i.e., lost results, etc.); and methods for evaluating personnel doses from nuclear accidents. 1 ref., 14 figs., 5 tabs 2. Personal recollections of radiation biology research at Hanford International Nuclear Information System (INIS) Thompson, R.C. 1995-01-01 This paper traces the evolution of the Hanford biology programme over a period of nearly five decades. The programme began in the 1940s with a focus on understanding the potential health effects of radionuclides such as 131 I associated with fallout from the atomic bomb. These studies were extended in the 1950s to experiments on the toxicity and metabolism of plutonium and fission products such as 90 Sr and 137 Cs. In the 1960s, a major long term project was initiated on the inhalation toxicology and carcinogenic effects of plutonium oxide and plutonium nitrate in dogs and rodents. The project remained a major effort within the overall Hanford biology programme throughout the 1970s and 1980s, during which time a broad range of new projects on energy-related pollutants, radon health effects, and basic radiation biology were initiated. Despite the many evolutionary changes that have occurred in the Hanford biology programme, the fundamental mission of understanding the effects of radiation on human health has endured for nearly five decades. (author) 3. Managing Hanford Site solid waste through strict acceptance criteria International Nuclear Information System (INIS) Jasen, W.G.; Pierce, R.D.; Willis, N.P. 1993-02-01 Various types of waste have been generated during the 50-year history of the Hanford Site. Regulatory changes in the last 20 years have provided the emphasis for better management of these wastes. Interpretations of the Atomic Energy Act of 1954 (AEA) and the Resource Conservation and Recovery Act of 1976 (RCRA) have led to the definition of a group of wastes called radioactive mixed wastes (RMW). As a result of the radioactive and hazardous properties of these wastes, strict management programs have been implemented for the management of these wastes. Solid waste management is accomplished through a systems performance approach to waste management that used best-demonstrated available technology (BDAT) and best management practices. The solid waste program at the Hanford Site strives to integrate all aspects of management relative to the treatment, storage and disposal (TSD) of solid waste. Often there are many competing and important needs. It is a difficult task to balance these needs in a manner that is both equitable and productive. Management science is used to help the process of making decisions. Tools used to support the decision making process include five-year planning, cost estimating, resource allocation, performance assessment, waste volume forecasts, input/output models, and waste acceptance criteria. The purpose of this document is to describe how one of these tools, waste acceptance criteria, has helped the Hanford Site manage solid wastes 4. The production and escape of nitrogen atoms on Mars Science.gov (United States) Fox, J. L. 1993-01-01 Updated rate coefficients and a revised ionosphere-thermosphere model are used to compute the production rates and densities of odd nitrogen species in the Martian atmosphere. Computed density profiles for N(4S), N(2D), N(2P), and NO are presented. The model NO densities are found to be about a factor of 2-3 less than those measured by the Viking 1 mass spectrometer. Revised values for the escape rates of N atoms from dissociative recombination and ionospheric reactions are also computed. Dissociative recombination is found to be comparable in importance to photodissociation at low solar activity, but it is still the most important escape mechanism for N-14 at high solar activity. 5. Researchers take up environmental challenge at Hanford International Nuclear Information System (INIS) Illman, D.L. 1993-01-01 The Hanford nuclear site, built to produce plutonium for the nation's first atomic weapons, occupies 560 square miles of desert in southeastern Washington State. Only 29 months after ground was broken at the site in March 1943, the Hanford project delivered the plutonium used in the bomb that was dropped on Nagasaki, Japan, at the end of World War II. Secrecy surrounding the nuclear weapons program continued through the Cold War years, concealing the fact that for decades, hazardous and radioactive wastes were discharged to the ground, water, and air at Hanford. Only in 1986 were documents finally declassified--tens of thousands of them--describing the construction, operation, and maintenance of the Hanford facilities, allowing a picture to be pieced together of the environmental cost there of the nuclear weapons buildup. That cost may never be completely tallied. But Westinghouse Hanford, Co., the principal operations contractor on the site, and Pacific Northwest Laboratories (PNL), operated by Battelle Memorial Institute for the Department of Energy (DOE), have now begun working together to develop new technologies that are needed to address the short-term and long-term challenges of environmental restoration at Hanford. The paper discusses the problems and possible solutions that are being investigated 6. TRIGA International - History of Training Research Isotope production General Atomics International Nuclear Information System (INIS) 2008-01-01 TRIGA conceived at GA in 1956 by a distinguished group of scientists including Edward Teller and Freeman Dyson. First TRIGA reactor Mk-1 was commissioned on 3 may 1958 at G.A. Characteristic feature of TRIGA reactors is inherent safety: Sitting can be confinement or conventional building. TRIGA reactors are the most prevalent in the world: 67 reactors in 24 countries. Steady state powers up to 14 MWt, pulsing up to 22,000 MWt. To enlarge the scope of its manufactured products, CERCA engaged in a Joint Venture with General Atomics, and in July 1995 a new Company was founded: TRIGA INTERNATIONAL SAS (50% GA, 50% CERCA; Head Office: Paris (France); Sales offices: GA San Diego (Ca, USA) and CERCA Lyon (France); Manufacturing plant: CERCA Romans. General Atomics ID: founded in 1955 at San Diego, California, by General Dynamics; status: Privately held corporation; owners: Neal and Linden Blue; business: High technology research, design, manufacturing, and production for industry and Government in the U.S. and overseas; locations: U.S., Germany, Japan, Australia, Thailand, Morocco; employees: 5,000. TRIGA's ID: CERCA is a subsidiary of AREVA, born in November 05, 1957. Activities: fuel manufacture for research reactor, equipment and components for high-energy physics, radioactive sources and reference sources; plants locations: Romans and Pierrelatte (France); total strength: 180. Since the last five years TRIGA has manufactured and delivered more than 800 fuel elements with a door to door service. TRIGA International has the experience to manufacture all types of TRIGA fuel: standard fuel elements, instrumented fuel elements, fuel followed control rods, geometry: 37.3 mm (1.47 in.), 35.8 mm (1.4 in), 13 mm (0.5 in), chemical Composition: U w% 8.5, 12, 20, 30 and 45 w/o, erbium and no erbium. TRIGA International is on INL's approved vendor list (ISO 9000/NQA) and is ready to meet any TRIGA fuel needs either in the US or worldwide 7. Giant atoms for the production of nuclear fuel International Nuclear Information System (INIS) Fahr, H.J. 1980-01-01 Neutral atoms can be blown up to the size of a football, if the electrons of the atomic shell are appropriately excited by photons just below the ionization energy. Such atoms, called Rydberg atoms, behave very differently to the usual ones in multiple respects. The fact that they can very easily be ionized by conventional electrostatic fields is being investigated as a method of isotope separation, for instance in the very important case of U 235 and U 238 . (orig.) [de 8. Remedial Investigation of Hanford Site Releases to the Columbia River - 13603 Energy Technology Data Exchange (ETDEWEB) Lerch, J.A.; Hulstrom, L.C. [Washington Closure Hanford, LLC, Richland, Washington 99354 (United States); Sands, J.P. [U.S Department of Energy, Richland Operations Office, Richland, Washington 99352 (United States) 2013-07-01 In south-central Washington State, the Columbia River flows through the U.S. Department of Energy Hanford Site. A primary objective of the Hanford Site cleanup mission is protection of the Columbia River, through remediation of contaminated soil and groundwater that resulted from its weapons production mission. Within the Columbia River system, surface water, sediment, and biota samples related to potential Hanford Site hazardous substance releases have been collected since the start of Hanford operations. The impacts from release of Hanford Site radioactive substances to the Columbia River in areas upstream, within, and downstream of the Hanford Site boundary have been previously investigated as mandated by the U.S. Department of Energy requirements under the Atomic Energy Act. The Remedial Investigation Work Plan for Hanford Site Releases to the Columbia River [1] was issued in 2008 to initiate assessment of the impacts under the Comprehensive Environmental Response, Compensation, and Liability Act of 1980 [2]. The work plan established a phased approach to characterize contaminants, assess current risks, and determine whether or not there is a need for any cleanup actions. Field investigation activities over a 120-mile stretch of the Columbia River began in October 2008 and were completed in 2010. Sampled media included surface water, pore water, surface and core sediment, island soil, and fish (carp, walleye, whitefish, sucker, small-mouth bass, and sturgeon). Information and sample results from the field investigation were used to characterize current conditions within the Columbia River and assess whether current conditions posed a risk to ecological or human receptors that would merit additional study or response actions under CERCLA. The human health and ecological risk assessments are documented in reports that were published in 2012 [3, 4]. Conclusions from the risk assessment reports are being summarized and integrated with remedial investigation 9. Atom International Nuclear Information System (INIS) Auffray, J.P. 1997-01-01 The atom through centuries, has been imagined, described, explored, then accelerated, combined...But what happens truly inside the atom? And what are mechanisms who allow its stability? Physicist and historian of sciences, Jean-Paul Auffray explains that these questions are to the heart of the modern physics and it brings them a new lighting. (N.C.) 10. Continuous production of nanostructured particles using spatial atomic layer deposition International Nuclear Information System (INIS) Ommen, J. Ruud van; Kooijman, Dirkjan; Niet, Mark de; Talebi, Mojgan; Goulas, Aristeidis 2015-01-01 In this paper, the authors demonstrate a novel spatial atomic layer deposition (ALD) process based on pneumatic transport of nanoparticle agglomerates. Nanoclusters of platinum (Pt) of ∼1 nm diameter are deposited onto titania (TiO 2 ) P25 nanoparticles resulting to a continuous production of an active photocatalyst (0.12–0.31 wt. % of Pt) at a rate of about 1 g min −1 . Tuning the precursor injection velocity (10–40 m s −1 ) enhances the contact between the precursor and the pneumatically transported support flows. Decreasing the chemisorption temperature (from 250 to 100 °C) results in more uniform distribution of the Pt nanoclusters as it decreases the reaction rate as compared to the rate of diffusion into the nanoparticle agglomerates. Utilizing this photocatalyst in the oxidation reaction of Acid Blue 9 showed a factor of five increase of the photocatalytic activity compared to the native P25 nanoparticles. The use of spatial particle ALD can be further expanded to deposition of nanoclusters on porous, micron-sized particles and to the production of core–shell nanoparticles enabling the robust and scalable manufacturing of nanostructured powders for catalysis and other applications 11. Hanford stakeholder participation in evaluating innovative technologies: VOC product line, Passive soil vapor extraction using borehole flux tunable hybrid plasma International Nuclear Information System (INIS) Peterson, T.; McCabe, G.; Niesen, K.; Serie, P. 1995-05-01 A three-phased stakeholder participation program was conducted to support the Volatile Organic Compounds Arid Site Integrated Demonstration (VOC-Arid ID). The US DOE's Office of Technology Development (OTD) sponsored and directed the VOC-Arid ID. Its purpose was to develop and demonstrate new technologies for remediating VOC contamination in soil and ground water. The integrated demonstration, hosted by the Hanford site in Washington State, is being transitioned into the Department of Energy's (DOE) Plume Focus Area. The Plume Focus Area has the same basic objectives as the ID, but is broader in scope and is a team effort with technology developers and technology users. The objective is to demonstrate a promising technology once, and if results warrant deploy it broadly across the DOE complex and in private sector applications 12. TRACKING CLEAN UP AT HANFORD International Nuclear Information System (INIS) CONNELL, C.W. 2005-01-01 The Hanford Federal Facility Agreement and Consent Order, known as the ''Tri-Party Agreement'' (TPA), is a legally binding agreement among the US Department of Energy (DOE), The Washington State Department of Ecology, and the US Environmental Protection Agency (EPA) for cleaning up the Hanford Site. Established in the 1940s to produce material for nuclear weapons as part of the Manhattan Project, Hanford is often referred to as the world's large environmental cleanup project. The Site covers more than 580 square miles in a relatively remote region of southeastern Washington state in the US. The production of nuclear materials at Hanford has left a legacy of tremendous proportions in terms of hazardous and radioactive waste. From a waste-management point of view, the task is enormous: 1700 waste sites; 450 billion gallons of liquid waste; 70 billion gallons of contaminated groundwater; 53 million gallons of tank waste; 9 reactors; 5 million cubic yards of contaminated soil; 22 thousand drums of mixed waste; 2.3 tons of spent nuclear fuel; and 17.8 metric tons of plutonium-bearing material and this is just a partial listing. The agreement requires that DOE provide the results of analytical laboratory and non-laboratory tests/readings to the lead regulatory agency to help guide then in making decisions. The agreement also calls for each signatory to preserve--for at least ten years after the Agreement has ended--all of the records in it, or its contractors, possession related to sampling, analysis, investigations, and monitoring conducted. The Action Plan that supports the TPA requires that Ecology and EPA have access to all data that is relevant to work performed, or to be performed, under the Agreement. Further, the Action Plan specifies two additional requirements: (1) that EPA, Ecology and their respective contractor staffs have access to all the information electronically, and (2) that the databases are accessible to, and used by, all personnel doing TPA 13. Hanford site environment International Nuclear Information System (INIS) Isaacson, R.E. 1976-01-01 A synopsis is given of the detailed characterization of the existing environment at Hanford. The following aspects are covered: demography, land use, meteorology, geology, hydrology, and seismology. It is concluded that Hanford is one of the most extensively characterized nuclear sites 14. Hanford defense waste studies International Nuclear Information System (INIS) Napier, B.A.; Zimmerman, M.G.; Soldat, J.K. 1981-01-01 PNL is assisting Rockwell Hanford Operations to prepare a programmatic environmental impact statement for the management of Hanford defense nuclear waste. The Ecological Sciences Department is leading the task of calculation of public radiation doses from a large matrix of potential routine and accidental releases of radionuclides to the environment 15. Simulation of N-atom production in dielectric-barrier discharge in nitrogen at atmospheric pressure International Nuclear Information System (INIS) Tsyganov, Dmitry; Pancheshnyi, Sergey 2012-01-01 A plasma-chemical model of atomic nitrogen production in a Townsend dielectric-barrier discharge in nitrogen at atmospheric pressure is presented. On the basis of the comparison with measured densities, a significant discrepancy between the calculated and the measured production rate of nitrogen atoms is observed and discussed. (paper) 16. Hanford Site Development Plan International Nuclear Information System (INIS) Hathaway, H.B.; Daly, K.S.; Rinne, C.A.; Seiler, S.W. 1992-05-01 The Hanford Site Development Plan (HSDP) provides an overview of land use, infrastructure, and facility requirements to support US Department of Energy (DOE) programs at the Hanford Site. The HSDP's primary purpose is to inform senior managers and interested parties of development activities and issues that require a commitment of resources to support the Hanford Site. The HSDP provides a land use plan for the Hanford Site and presents a picture of what is currently known and anticipated in accordance with DOE Order 4320.1B. Site Development Planning. The HSDP wig be updated annually as future decisions further shape the mission and overall site development process. Further details about Hanford Site development are defined in individual area development plans 17. Atomic absorption determination of vanadium in products of metallurgical production and mineral feed stock International Nuclear Information System (INIS) Polikarpova, N.V.; Panteleeva, E.Yu. 1983-01-01 Rapid and selective method of atomic absorption determination of vanadium in metallurgical process products and numerical feed stock is suggested. Buffering mixture of aluminium and phosphoric acid is used to suppress the effect of sample composition on the value of vanadium atomic absorption. The concentration of buffer components can vary from 400 up to 2000 μg/ml Al and from 2 up to 5% vol. H 3 PO 4 . The suggested mixture completely eli-- minates the strong chromium effect. The developed method was used for analyzing steels, alloys based on Mo, Ni, Ti, Cr, as well as titanium magnetite ores and concentrates. The method enables to determine from 0.05 up to 10% vanadium with 0.05-0.01 relative standard deviation, respectively 18. 1997 evaluation of tritium removal and mitigation technologies for Hanford Site wastewaters International Nuclear Information System (INIS) Jeppson, D.W.; Biyani, R.K.; Duncan, J.B.; Flyckt, D.L.; Mohondro, P.C.; Sinton, G.L. 1997-01-01 This report contains results of a biennial assessment of tritium separation technology and tritium nitration techniques for control of tritium bearing wastewaters at the Hanford Site. Tritium in wastewaters at Hanford have resulted from plutonium production, fuel reprocessing, and waste handling operations since 1944. this assessment was conducted in response to the Hanford Federal Facility Agreement and Consent Order 19. 1995 Report on Hanford site land disposal restrictions for mixed waste International Nuclear Information System (INIS) Black, D.G. 1995-04-01 This report was submitted to meet the requirements of Hanford Federal Facility Agreement and Consent Order Milestone M-26-01E. This milestone requires the preparation of an annual report that covers characterization, treatment, storage, minimization, and other aspects of land disposal restricted mixed waste at the Hanford Site. The U.S. Department of Energy, its predecessors, and contractors at the Hanford Site were involved in the production and purification of nuclear defense materials from the early 1940s to the late 1980s. These production activities have generated large quantities of liquid and solid radioactive mixed waste. This waste is subject to regulation under authority of both the Resource Conservation and Recovery Act of 1976 and Atomic Energy Act of 1954. This report covers mixed waste only. The Washington State Department of Ecology, U.S. Environmental Protection Agency, and U.S. Department of Energy have entered into an agreement, the Hanford Federal Facility Agreement and Consent Order (commonly referred to as the Tri-Party Agreement) to bring the Hanford Site operations into compliance with dangerous waste regulations. The Tri-Party Agreement required development of the original land disposal restrictions (LDRs) plan and its annual updates to comply with LDR requirements for radioactive mixed waste. This report is the fifth update of the plan first issued in 1990. Tri-Party Agreement negotiations completed in 1993 and approved in January 1994 changed and added many new milestones. Most of the changes were related to the Tank Waste Remediation System and these changes are incorporated into this report 20. 1998 report on Hanford Site land disposal restrictions for mixed waste International Nuclear Information System (INIS) Black, D.G. 1998-01-01 This report was submitted to meet the requirements of Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement) Milestone M-26-01H. This milestone requires the preparation of an annual report that covers characterization, treatment, storage, minimization, and other aspects of managing land-disposal-restricted mixed waste at the Hanford Facility. The US Department of Energy, its predecessors, and contractors on the Hanford Facility were involved in the production and purification of nuclear defense materials from the early 1940s to the late 1980s. These production activities have generated large quantities of liquid and solid mixed waste. This waste is regulated under authority of both the Resource Conservation and Recovery Act of l976 and the Atomic Energy Act of 1954. This report covers only mixed waste. The Washington State Department of Ecology, US Environmental Protection Agency, and US Department of Energy have entered into the Tri-Party Agreement to bring the Hanford Facility operations into compliance with dangerous waste regulations. The Tri-Party Agreement required development of the original land disposal restrictions (LDR) plan and its annual updates to comply with LDR requirements for mixed waste. This report is the eighth update of the plan first issued in 1990. The Tri-Party Agreement requires and the baseline plan and annual update reports provide the following information: (1) Waste Characterization Information -- Provides information about characterizing each LDR mixed waste stream. The sampling and analysis methods and protocols, past characterization results, and, where available, a schedule for providing the characterization information are discussed. (2) Storage Data -- Identifies and describes the mixed waste on the Hanford Facility. Storage data include the Resource Conservation and Recovery Act of 1976 dangerous waste codes, generator process knowledge needed to identify the waste and to make LDR determinations, quantities 1. 1998 report on Hanford Site land disposal restrictions for mixed waste Energy Technology Data Exchange (ETDEWEB) Black, D.G. 1998-04-10 This report was submitted to meet the requirements of Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement) Milestone M-26-01H. This milestone requires the preparation of an annual report that covers characterization, treatment, storage, minimization, and other aspects of managing land-disposal-restricted mixed waste at the Hanford Facility. The US Department of Energy, its predecessors, and contractors on the Hanford Facility were involved in the production and purification of nuclear defense materials from the early 1940s to the late 1980s. These production activities have generated large quantities of liquid and solid mixed waste. This waste is regulated under authority of both the Resource Conservation and Recovery Act of l976 and the Atomic Energy Act of 1954. This report covers only mixed waste. The Washington State Department of Ecology, US Environmental Protection Agency, and US Department of Energy have entered into the Tri-Party Agreement to bring the Hanford Facility operations into compliance with dangerous waste regulations. The Tri-Party Agreement required development of the original land disposal restrictions (LDR) plan and its annual updates to comply with LDR requirements for mixed waste. This report is the eighth update of the plan first issued in 1990. The Tri-Party Agreement requires and the baseline plan and annual update reports provide the following information: (1) Waste Characterization Information -- Provides information about characterizing each LDR mixed waste stream. The sampling and analysis methods and protocols, past characterization results, and, where available, a schedule for providing the characterization information are discussed. (2) Storage Data -- Identifies and describes the mixed waste on the Hanford Facility. Storage data include the Resource Conservation and Recovery Act of 1976 dangerous waste codes, generator process knowledge needed to identify the waste and to make LDR determinations, quantities 2. 1995 Report on Hanford site land disposal restrictions for mixed waste Energy Technology Data Exchange (ETDEWEB) Black, D.G. 1995-04-01 This report was submitted to meet the requirements of Hanford Federal Facility Agreement and Consent Order Milestone M-26-01E. This milestone requires the preparation of an annual report that covers characterization, treatment, storage, minimization, and other aspects of land disposal restricted mixed waste at the Hanford Site. The U.S. Department of Energy, its predecessors, and contractors at the Hanford Site were involved in the production and purification of nuclear defense materials from the early 1940s to the late 1980s. These production activities have generated large quantities of liquid and solid radioactive mixed waste. This waste is subject to regulation under authority of both the Resource Conservation and Recovery Act of 1976 and Atomic Energy Act of 1954. This report covers mixed waste only. The Washington State Department of Ecology, U.S. Environmental Protection Agency, and U.S. Department of Energy have entered into an agreement, the Hanford Federal Facility Agreement and Consent Order (commonly referred to as the Tri-Party Agreement) to bring the Hanford Site operations into compliance with dangerous waste regulations. The Tri-Party Agreement required development of the original land disposal restrictions (LDRs) plan and its annual updates to comply with LDR requirements for radioactive mixed waste. This report is the fifth update of the plan first issued in 1990. Tri-Party Agreement negotiations completed in 1993 and approved in January 1994 changed and added many new milestones. Most of the changes were related to the Tank Waste Remediation System and these changes are incorporated into this report. 3. Storage of non-defense production reactor spent nuclear fuel at the Department of Energy's Hanford Site International Nuclear Information System (INIS) Carlson, A.B. 1998-01-01 In 1992, the US Department of Energy (DOE) established a program at the Hanford Site for management of DOE-owned spent nuclear fuel (SNF) until final disposition. Currently, the DOE-owned SNF Program is developing and implementing plans to assure existing storage, achieve interim storage, and prepare DOE-owned SNF for final disposition. Program requirements for management of the SNF are delineated in the DOE-owned SNF Program Plan.(DOE 1995a) and the DOE Spent Fuel Program's Requirements Document (DOE 1994a). Major program requirements are driven by the following: commitments established in the Defense Nuclear Facilities Safety Board (DNFSB) Recommendation 94-1 Implementation Plan (DOE 1995b); corrective action plans for resolving vulnerabilities identified in the DOE Spent Fuel Working Group's Report on Health, Safety, and Environmental Vulnerabilities for Reactor Irradiated Nuclear Materials (DOE 1993); the settlement agreement between the US Department of Navy, the US Department of Energy, and the State of Idaho on the record of decision (ROD) from the DOE Programmatic SNF Management and Idaho National Engineering Laboratory Environmental Restoration and Waste Management Programs Environmental Impact Statement (DOE Programmatic SNF EIS) (Idaho, 1995) 4. Engineering report of plasma vitrification of Hanford tank wastes International Nuclear Information System (INIS) Hendrickson, D.W. 1995-01-01 This document provides an analysis of vendor-derived testing and technology applicability to full scale glass production from Hanford tank wastes using plasma vitrification. The subject vendor testing and concept was applied in support of the Hanford LLW Vitrification Program, Tank Waste Remediation System 5. Hanford Site Groundwater Monitoring for Fiscal Year 2006 Energy Technology Data Exchange (ETDEWEB) Hartman, Mary J.; Morasch, Launa F.; Webber, William D. 2007-03-01 This report presents the results of groundwater monitoring for FY 2006 on DOE's Hanford Site. Results of groundwater remediation, vadose zone monitoring, and characterization are summarized. DOE monitors groundwater at the Hanford Site to fulfill a variety of state and federal regulations, including the Atomic Energy Act (AEA), the Resource Conservation and Recovery Act (RCRA), the Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA), and Washington Administrative Code (WAC). 6. Positron production in heavy ion-atom collisions International Nuclear Information System (INIS) Kienle, P. 1985-08-01 Following an introduction into the concept of an atom with an overcritical field, established by a nucleus with a charge larger than 173, the spontaneous positron emission from such an atom with an 1s binding energy exceeding 2mc 2 is presented. Such a process, in which an electron is bound and a monoenergetic positron emitted turns the neutral into a charged vacuum. In a U-U di-nuclear system (Z=184) the spontaneous positron emission proceeds with an energy of about 300 keV corresponding to a decay time of 3x10 -19 sec. (orig./WL) 7. Hydrogeologic model for the old Hanford townsite International Nuclear Information System (INIS) MacDonald, Q.; Csun, C. 1994-01-01 The Hanford Site in southeastern Washington state produced the country's first plutonium during WW II, and production continued through the end of the cold war. This plutonium production generated significant volumes of chemical and radioactive wastes, some of which were discharged directly to the local sediments as wastewater. Artifical recharge is still the dominating influence on the uppermost and unconfined aquifer over much of the Hanford site. Groundwater from a portion of this aquifer, which is in excess of drinking water standards for tritium, discharges to the Columbia River in the vicinity of the old Hanford townsite. The Hanford site lies within the Pasco basin, which is a structural basin in the Columbia Plateau. Columbia River basalt is overlain by the fluvial and lacustrian Ringold formation. The Ringold is unconformably overlain by the informal Hanford formation. Relatively impermeable basalt outcrops and subcrops along a northwest-southeast-trending anticline across the study area. Hanford sediments include both fluvial and glacial flood deposits lying on an irregular surface of basalt and sedimentary rocks. The coarser flood deposits have very high hydraulic conductivity and probably are the most important conduit for contaminant transport within the aquifer. A finite element model (CFEST-SC) is being used to study the effect of changing river stage on baseflow to the Columbia River near the old Hanford townsite. A steady-state version of the model produces calculated head within 1 m of observed values. Transient flow and solute transport results are expected to help further define the relationship between the contaminated aquifer and the Columbia River 8. Hanford Site surface environmental surveillance International Nuclear Information System (INIS) Dirkes, R.L. 1998-01-01 Environmental surveillance of the Hanford Site and the surrounding region is conducted to demonstrate compliance with environmental regulations, confirm adherence to US Department of Energy (DOE) environmental protection policies, support DOE environmental management decisions, and provide information to the public. The Surface Environmental Surveillance Project (SESP) is a multimedia environmental monitoring program conducted to measure the concentrations of radionuclides and chemical contaminants in the environment and assess the integrated effects of these contaminants on the environment and the public. The monitoring program includes sampling air, surface water, sediments, soil, natural vegetation, agricultural products, fish, and wildlife. Functional elements inherent in the operation of the SESP include project management, quality assurance/control, training, records management, environmental sampling network design and implementation, sample collection, sample analysis, data management, data review and evaluation, exposure assessment, and reporting. The SESP focuses on those contaminant/media combinations calculated to have the highest potential for contributing to off-site exposure. Results of the SESP indicate that contaminant concentrations in the Hanford environs are very low, generally below environmental standards, at or below analytical detection levels, and indicative of environmental levels. However, areas of elevated contaminant concentrations have been identified at Hanford. The extent of these areas is generally limited to past operating areas and waste disposal sites 9. Hanford Site Development Plan Energy Technology Data Exchange (ETDEWEB) Rinne, C.A.; Curry, R.H.; Hagan, J.W.; Seiler, S.W.; Sommer, D.J. (Westinghouse Hanford Co., Richland, WA (USA)); Yancey, E.F. (Pacific Northwest Lab., Richland, WA (USA)) 1990-01-01 The Hanford Site Development Plan (Site Development Plan) is intended to guide the short- and long-range development and use of the Hanford Site. All acquisition, development, and permanent facility use at the Hanford Site will conform to the approved plan. The Site Development Plan also serves as the base document for all subsequent studies that involve use of facilities at the Site. This revision is an update of a previous plan. The executive summary presents the highlights of the five major topics covered in the Site Development Plan: general site information, existing conditions, planning analysis, Master Plan, and Five-Year Plan. 56 refs., 67 figs., 31 tabs. 10. Hanford Site Development Plan International Nuclear Information System (INIS) Rinne, C.A.; Curry, R.H.; Hagan, J.W.; Seiler, S.W.; Sommer, D.J.; Yancey, E.F. 1990-01-01 The Hanford Site Development Plan (Site Development Plan) is intended to guide the short- and long-range development and use of the Hanford Site. All acquisition, development, and permanent facility use at the Hanford Site will conform to the approved plan. The Site Development Plan also serves as the base document for all subsequent studies that involve use of facilities at the Site. This revision is an update of a previous plan. The executive summary presents the highlights of the five major topics covered in the Site Development Plan: general site information, existing conditions, planning analysis, Master Plan, and Five-Year Plan. 56 refs., 67 figs., 31 tabs 11. Production of pulsed atomic oxygen beams via laser vaporization methods International Nuclear Information System (INIS) Brinza, D.E.; Coulter, D.R.; Liang, R.H.; Gupta, A. 1987-01-01 Energetic pulsed atomic oxygen beams were generated by laser-driven evaporation of cryogenically frozen ozone/oxygen films and thin films of indium-tin oxide (ITO). Mass and energy characterization of beams from the ozone/oxygen films were carried out by mass spectrometry. The peak flux, found to occur at 10 eV, is estimated from this data to be 3 x 10(20) m(-2) s(-1). Analysis of the time-of-flight data indicates a number of processes contribute to the formation of the atomic oxygen beam. The absence of metastable states such as the 2p(3) 3s(1) (5S) level of atomic oxygen blown off from ITO films is supported by the failure to observe emission at 777.3 nm from the 2p(3) 3p(1) (5P/sub J/) levels. Reactive scattering experiments with polymer film targets for atomic oxygen bombardment are planned using a universal crossed molecular beam apparatus 12. Hanford Waste Vitrification Plant applied technology plan International Nuclear Information System (INIS) Kruger, O.L. 1990-09-01 This Applied Technology Plan describes the process development, verification testing, equipment adaptation, and waste form qualification technical issues and plans for resolution to support the design, permitting, and operation of the Hanford Waste Vitrification Plant. The scope of this Plan includes work to be performed by the research and development contractor, Pacific Northwest Laboratory, other organizations within Westinghouse Hanford Company, universities and companies with glass technology expertise, and other US Department of Energy sites. All work described in this Plan is funded by the Hanford Waste Vitrification Plant Project and the relationship of this Plan to other waste management documents and issues is provided for background information. Work to performed under this Plan is divided into major areas that establish a reference process, develop an acceptable glass composition envelope, and demonstrate feed processing and glass production for the range of Hanford Waste Vitrification Plant feeds. Included in this work is the evaluation and verification testing of equipment and technology obtained from the Defense Waste Processing Facility, the West Valley Demonstration Project, foreign countries, and the Hanford Site. Development and verification of product and process models and other data needed for waste form qualification documentation are also included in this Plan. 21 refs., 4 figs., 33 tabs 13. Hanford Tanks Initiative quality assurance implementation plan International Nuclear Information System (INIS) Huston, J.J. 1998-01-01 Hanford Tanks Initiative (HTI) Quality Assurance Implementation Plan for Nuclear Facilities defines the controls for the products and activities developed by HTI. Project Hanford Management Contract (PHMC) Quality Assurance Program Description (QAPD)(HNF-PRO599) is the document that defines the quality requirements for Nuclear Facilities. The QAPD provides direction for compliance to 10 CFR 830.120 Nuclear Safety Management, Quality Assurance Requirements. Hanford Tanks Initiative (HTI) is a five-year activity resulting from the technical and financial partnership of the US Department of Energy's Office of Waste Management (EM-30), and Office of Science and Technology Development (EM-50). HTI will develop and demonstrate technologies and processes for characterization and retrieval of single shell tank waste. Activities and products associated with HTI consist of engineering, construction, procurement, closure, retrieval, characterization, and safety and licensing 14. Calibrated atomic force microscope measurements of vickers hardness indentations and tip production and characterisation for scanning tunelling microscope DEFF Research Database (Denmark) Jensen, Carsten P. Calibrated atomic force microscope measurements of vickers hardness indentations and tip production and characterisation for scanning tunelling microscope......Calibrated atomic force microscope measurements of vickers hardness indentations and tip production and characterisation for scanning tunelling microscope... 15. Hanford Site Infrastructure Plan International Nuclear Information System (INIS) 1990-01-01 The Hanford Site Infrastructure Plan (HIP) has been prepared as an overview of the facilities, utilities, systems, and services that support all activities on the Hanford Site. Its purpose is three-fold: to examine in detail the existing condition of the Hanford Site's aging utility systems, transportation systems, Site services and general-purpose facilities; to evaluate the ability of these systems to meet present and forecasted Site missions; to identify maintenance and upgrade projects necessary to ensure continued safe and cost-effective support to Hanford Site programs well into the twenty-first century. The HIP is intended to be a dynamic document that will be updated accordingly as Site activities, conditions, and requirements change. 35 figs., 25 tabs 16. Hanford Emergency Response Plan International Nuclear Information System (INIS) Wagoner, J.D. 1994-04-01 The Hanford Emergency Response Plan for the US Department of Energy (DOE), Richland Operations Office (RL), incorporates into one document an overview of the emergency management program for the Hanford Site. The program has been developed in accordance with DOE orders, and state and federal regulations to protect worker and public health and safety and the environment in the event of an emergency at or affecting the Hanford Site. This plan provides a description of how the Hanford Site will implement the provisions of DOE 5500 series and other applicable Orders in terms of overall policies and concept of operations. It should be used as the basis, along with DOE Orders, for the development of specific contractor and RL implementing procedures 17. Hanford Emergency Response Plan Energy Technology Data Exchange (ETDEWEB) Wagoner, J.D. 1994-04-01 The Hanford Emergency Response Plan for the US Department of Energy (DOE), Richland Operations Office (RL), incorporates into one document an overview of the emergency management program for the Hanford Site. The program has been developed in accordance with DOE orders, and state and federal regulations to protect worker and public health and safety and the environment in the event of an emergency at or affecting the Hanford Site. This plan provides a description of how the Hanford Site will implement the provisions of DOE 5500 series and other applicable Orders in terms of overall policies and concept of operations. It should be used as the basis, along with DOE Orders, for the development of specific contractor and RL implementing procedures. 18. Hanford cultural resources laboratory International Nuclear Information System (INIS) Wright, M.K. 1995-01-01 This section of the 1994 Hanford Site Environmental Report describes activities of the Hanford Cultural Resources Laboratory (HCRL) which was established by the Richland Operations Office in 1987 as part of PNL.The HCRL provides support for the management of the archaeological, historical, and traditional cultural resources of the site in a manner consistent with the National Historic Preservation Act, the Native American Graves Protection and Repatriation Act, and the American Indian Religious Freedom Act 19. Hanford cultural resources laboratory Energy Technology Data Exchange (ETDEWEB) Wright, M.K. 1995-06-01 This section of the 1994 Hanford Site Environmental Report describes activities of the Hanford Cultural Resources Laboratory (HCRL) which was established by the Richland Operations Office in 1987 as part of PNL.The HCRL provides support for the management of the archaeological, historical, and traditional cultural resources of the site in a manner consistent with the National Historic Preservation Act, the Native American Graves Protection and Repatriation Act, and the American Indian Religious Freedom Act. 20. Hanford Facility contingency plan International Nuclear Information System (INIS) Sutton, L.N.; Miskho, A.G.; Brunke, R.C. 1993-10-01 The Hanford Facility Contingency Plan, together with each TSD unit-specific contingency plan, meets the WAC 173-303 requirements for a contingency plan. This plan includes descriptions of responses to a nonradiological hazardous materials spill or release at Hanford Facility locations not covered by TSD unit-specific contingency plans or building emergency plans. This plan includes descriptions of responses for spills or releases as a result of transportation activities, movement of materials, packaging, and storage of hazardous materials 1. Hanford work faces change International Nuclear Information System (INIS) Anon. 1992-01-01 This article is a discussion of DOE efforts in the awarding of a large engineering-construction contract at the Hanford Reservation. Though the announced winner was a group lead by J. A. Jones Construction/Duke Engineering Services, the incumbent (ICF-Kaiser Engineers) protested the announced award. The protest was dismissed by the GAO, but DOE officials still reopened the bidding. There was also a short note regarding the award of the ERMC at Hanford 2. Managing risk at Hanford International Nuclear Information System (INIS) Hesser, W.A.; Stillwell, W.G.; Rutherford, W.A. 1994-01-01 Clearly, there is sufficient motivation from Washington for the Hanford community to pay particular attention to the risks associated with the substantial volumes of radiological, hazardous, and mixed waste at Hanford. But there is also another reason for emphasizing risk: Hanford leaders have come to realize that their decisions must consider risk and risk reduction if those decisions are to be technically sound, financially affordable, and publicly acceptable. The 560-square miles of desert land is worth only a few thousand dollars an acre (if that) -- hardly enough to justify the almost two billion dollars that will be spent at Hanford this year. The benefit of cleaning up the Hanford Site is not the land but the reduction of potential risk to the public and the environment for future generations. If risk reduction is our ultimate goal, decisions about priority of effort and resource allocation must consider those risks, now and in the future. The purpose of this paper is to describe how Hanford is addressing the issues of risk assessment, risk management, and risk-based decision making and to share some of our experiences in these areas 3. Volatility and entrainment of feed components and product glass characteristics during pilot-scale vitrification of simulated Hanford site low-level waste International Nuclear Information System (INIS) Shade, J.W. 1996-01-01 Commercially available melter technologies were tested for application to vitrification of Hanford site low-level waste (LLW). Testing was conducted at vendor facilities using a non-radioactive LLW simulant. Technologies tested included four Joule-heated melter types, a carbon electrode melter, a cyclone combustion melter, and a plasma torch-fired melter. A variety of samples were collected during the vendor tests and analyzed to provide data to support evaluation of the technologies. This paper describes the evaluation of melter feed component volatility and entrainment losses and product glass samples produced during the vendor tests. All vendors produced glasses that met minimum leach criteria established for the test glass formulations, although in many cases the waste oxide loading was less than intended. Entrainment was much lower in Joule-heated systems than in the combustion or plasma torch-fired systems. Volatility of alkali metals, halogens, B, Mo, and P were severe for non-Joule-heated systems. While losses of sulfur were significant for all systems, the volatility of other components was greatly reduced for some configurations of Joule-heated melters. Data on approaches to reduce NO x generation, resulting from high nitrate and nitrite content in the double-shell slurry feed, are also presented 4. Chemical composition analysis and product consistency tests to support enhanced Hanford waste glass models. Results for the third set of high alumina outer layer matrix glasses Energy Technology Data Exchange (ETDEWEB) Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States) 2015-12-01 In this report, the Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for 14 simulated high level waste glasses fabricated by the Pacific Northwest National Laboratory. The results of these analyses will be used as part of efforts to revise or extend the validation regions of the current Hanford Waste Treatment and Immobilization Plant glass property models to cover a broader span of waste compositions. The measured chemical composition data are reported and compared with the targeted values for each component for each glass. All of the measured sums of oxides for the study glasses fell within the interval of 96.9 to 100.8 wt %, indicating recovery of all components. Comparisons of the targeted and measured chemical compositions showed that the measured values for the glasses met the targeted concentrations within 10% for those components present at more than 5 wt %. The PCT results were normalized to both the targeted and measured compositions of the study glasses. Several of the glasses exhibited increases in normalized concentrations (NCi) after the canister centerline cooled (CCC) heat treatment. Five of the glasses, after the CCC heat treatment, had NCB values that exceeded that of the Environmental Assessment (EA) benchmark glass. These results can be combined with additional characterization, including X-ray diffraction, to determine the cause of the higher release rates. 5. Excited-atom production by electron and ion bombardment of alkali halides International Nuclear Information System (INIS) Walkup, R.E.; Avouris, P.; Ghosh, A.P. 1987-01-01 We present experimental results on the production of excited atoms by electron and ion bombardment of alkali halides. For the case of electron bombardment, Doppler shift measurements show that the electronically excited atoms have a thermal velocity distribution in equilibrium with the surface temperature. Measurements of the absolute yield of excited atoms, the distribution of population among the excited states, and the systematic dependence on incident electron current and sample temperature support a model in which the excited atoms are produced by gas-phase collisions between desorbed ground-state atoms and secondary electrons. In contrast, for the case of ion bombardment, the excited atoms are directly sputtered from the surface, with velocity distributions characteristic of a collision cascade, and with typical energies of --10 eV 6. Atomic iodine production in a gas flow by decomposing methyl iodide in a dc glow discharge International Nuclear Information System (INIS) Mikheyev, P A; Shepelenko, A A; Voronov, A I; Kupryaev, Nikolai V 2002-01-01 The production of atomic iodine for an oxygen - iodine laser is studied by decomposing methyl iodide in a dc glow discharge in a vortex gas flow. The concentration of iodine atoms in discharge products was measured from the atomic iodine absorption of the radiation of a single-frequency tunable diode laser at a wavelength of 1.315 μm. Atomic iodine concentrations sufficient for the operation of an oxygen - iodine laser were obtained. The concentration of atomic iodine amounted to 3.6 x 10 15 cm -3 for a pressure of the carrying argon gas of 15 Torr. The discharge stabilisation by a vortex gas flow allowed the glow discharge to be sustained in a strongly electronegative halogen-containing gas mixture for pressures up to 20 Torr. (active media) 7. Time-resolved production and detection of reactive atoms International Nuclear Information System (INIS) Grossman, L.W.; Hurst, G.S. 1977-09-01 Cesium iodide in the presence of a buffer gas was dissociated with a pulsed ultraviolet laser, which will be referred to as the source laser. This created a population of atoms at a well defined time and in a compact, well defined volume. A second pulsed laser, with a beam that completely surrounded that of the first, photoionized the cesium after a known time delay. This laser will be referred to as the detector laser. It was determined that for short time delays, all of the cesium atoms were easily ionized. When focused, the source laser generated an extremely intense fluence. By accounting for the beam intensity profile it was shown that all of the molecules in the central portion of the beam can be dissociated and detected. Besides proving the feasibility of single-molecule detection, this enabled a determination of the absolute photodissociation cross section as a function of wavelength. Initial studies of the time decay of the cesium signal at low argon pressures indicated a non-exponential decay. This was consistent with a diffusion mechanism transporting cesium atoms out of the laser beam. Therefore, it was desired to conduct further experiments using a tightly focused source beam, passing along the axis of the detector beam. The theoretical behavior of this simple geometry accounting for diffusion and reaction is easily calculated. A diffusion coefficient can then be extracted by data fitting. If reactive decay is due to impurities constituting a fixed percentage of the buffer gas, then two-body reaction rates will scale linearly with pressure and three-body reaction rates will scale quadratically. Also, the diffusion coefficient will scale inversely with pressure. At low pressures it is conceivable that decay due to diffusion would be sufficiently rapid that all other processes can be neglected. Extraction of a diffusion coefficient would then be quite direct. Finally, study of the reaction of cesium and oxygen was undertaken 8. HANFORD GROUNDWATER REMEDIATION Energy Technology Data Exchange (ETDEWEB) CHARBONEAU, B; THOMPSON, M; WILDE, R.; FORD, B.; GERBER, M.S. 2006-02-01 By 1990 nearly 50 years of producing plutonium put approximately 1.70E + 12 liters (450 billion gallons) of liquid wastes into the soil of the 1,518-square kilometer (586-square mile) Hanford Site in southeast Washington State. The liquid releases consisted of chemicals used in laboratory experiments, manufacturing and rinsing uranium fuel, dissolving that fuel after irradiation in Hanford's nuclear reactors, and in liquefying plutonium scraps needed to feed other plutonium-processing operations. Chemicals were also added to the water used to cool Hanford's reactors to prevent corrosion in the reactor tubes. In addition, water and acid rinses were used to clean plutonium deposits from piping in Hanford's large radiochemical facilities. All of these chemicals became contaminated with radionuclides. As Hanford raced to help win World War II, and then raced to produce materials for the Cold War, these radioactive liquid wastes were released to the Site's sandy soils. Early scientific experiments seemed to show that the most highly radioactive components of these liquids would bind to the soil just below the surface of the land, thus posing no threat to groundwater. Other experiments predicted that the water containing most radionuclides would take hundreds of years to seep into groundwater, decaying (or losing) most of its radioactivity before reaching the groundwater or subsequently flowing into the Columbia River, although it was known that some contaminants like tritium would move quickly. Evidence today, however, shows that many contaminants have reached the Site's groundwater and the Columbia River, with more on its way. Over 259 square kilometers (100 square miles) of groundwater at Hanford have contaminant levels above drinking-water standards. Also key to successfully cleaning up the Site is providing information resources and public-involvement opportunities to Hanford's stakeholders. This large, passionate, diverse, and 9. Technical basis for internal dosimetry at Hanford International Nuclear Information System (INIS) Sula, M.J.; Carbaugh, E.H.; Bihl, D.E. 1991-07-01 The Hanford Internal Dosimetry Program, administered by Pacific Northwest Laboratory for the US Department of Energy, provides routine bioassay monitoring for employees who are potentially exposed to radionuclides in the workplace. This report presents the technical basis for routine bioassay monitoring and the assessment of internal dose at Hanford. The radionuclides of concern include tritium, corrosion products ( 58 Co, 60 Co, 54 Mn, and 59 Fe), strontium, cesium, iodine, europium, uranium, plutonium, and americium,. Sections on each of these radionuclides discuss the sources and characteristics; dosimetry; bioassay measurements and monitoring; dose measurement, assessment, and mitigation and bioassay follow-up treatment. 78 refs., 35 figs., 115 tabs 10. Technical basis for internal dosimetry at Hanford Energy Technology Data Exchange (ETDEWEB) Sula, M.J.; Carbaugh, E.H.; Bihl, D.E. 1991-07-01 The Hanford Internal Dosimetry Program, administered by Pacific Northwest Laboratory for the US Department of Energy, provides routine bioassay monitoring for employees who are potentially exposed to radionuclides in the workplace. This report presents the technical basis for routine bioassay monitoring and the assessment of internal dose at Hanford. The radionuclides of concern include tritium, corrosion products ({sup 58}Co, {sup 60}Co, {sup 54}Mn, and {sup 59}Fe), strontium, cesium, iodine, europium, uranium, plutonium, and americium,. Sections on each of these radionuclides discuss the sources and characteristics; dosimetry; bioassay measurements and monitoring; dose measurement, assessment, and mitigation and bioassay follow-up treatment. 78 refs., 35 figs., 115 tabs. 11. Technical basis for internal dosimetry at Hanford Energy Technology Data Exchange (ETDEWEB) Sula, M.J.; Carbaugh, E.H.; Bihl, D.E. 1989-04-01 The Hanford Internal Dosimetry Program, administered by Pacific Northwest Laboratory for the US Department of Energy, provides routine bioassay monitoring for employees who are potentially exposed to radionuclides in the workplace. This report presents the technical basis for routine bioassay monitoring and the assessment of internal dose at Hanford. The radionuclides of concern include tritium, corrosion products (/sup 58/Co, /sup 60/Co, /sup 54/Mn, and /sup 59/Fe), strontium, cesium, iodine, europium, uranium, plutonium, and americium. Sections on each of these radionuclides discuss the sources and characteristics; dosimetry; bioassay measurements and monitoring; dose measurement, assessment, and mitigation; and bioassay follow-up treatment. 64 refs., 42 figs., 118 tabs. 12. Technical basis for internal dosimetry at Hanford International Nuclear Information System (INIS) Sula, M.J.; Carbaugh, E.H.; Bihl, D.E. 1989-04-01 The Hanford Internal Dosimetry Program, administered by Pacific Northwest Laboratory for the US Department of Energy, provides routine bioassay monitoring for employees who are potentially exposed to radionuclides in the workplace. This report presents the technical basis for routine bioassay monitoring and the assessment of internal dose at Hanford. The radionuclides of concern include tritium, corrosion products ( 58 Co, 60 Co, 54 Mn, and 59 Fe), strontium, cesium, iodine, europium, uranium, plutonium, and americium. Sections on each of these radionuclides discuss the sources and characteristics; dosimetry; bioassay measurements and monitoring; dose measurement, assessment, and mitigation; and bioassay follow-up treatment. 64 refs., 42 figs., 118 tabs 13. Westinghouse Hanford Company package testing capabilities International Nuclear Information System (INIS) Hummer, J.H.; Mercado, M.S. 1993-07-01 The Department of Energy's Hanford Site is a 1,450-km 2 (560-mi 2 ) installation located in southeastern Washington State. Established in 1943 as a plutonium production facility, Hanford's role has evolved into one of environmental restoration and remediation. Many of these environmental restoration and remediation activities involve transportation of radioactive/hazardous materials. Packagings used for the transportation of radioactive/hazardous materials must be capable of meeting certain normal transport and hypothetical accident performance criteria. Evaluations of performance to these criteria typically involve a combination of analysis and testing. Required tests may include the free drop, puncture, penetration, compression, thermal, heat, cold, vibration, water spray, water immersion, reduced pressure, and increased pressure tests. The purpose of this paper is to outline the Hanford capabilities for performing each of these tests 14. Hanford study: a review of its limitations and controversial conclusions International Nuclear Information System (INIS) Gilbert, E.S. 1984-10-01 The Hanford data set has attracted attention primarily because of analyses conducted by Mancuso, Stewart, and Kneale (MSK). These investigators claim that the Hanford data provide evidence that our current estimates of cancer mortality resulting from radiation exposure are too low, and advocate replacing estimates based on populations exposed at relatively high doses (such as the Japanese atom bomb survivors) with estimates based on the Hanford data. In this paper, it is shown that the only evidence of association of radiation exposure and mortality provided by the Hanford data is a small excess of multiple myeloma, and that this data set is not adequate for reliable risk estimation. It is demonstrated that confidence limits for risk estimates are very wide, and that the data are not adequate to differentiate among models. The more recent MSK analyses, which claim to provide adequate models and risk estimates, are critiqued. 18 references, 1 table 15. 1993 report on Hanford Site land disposal restrictions for mixed wastes International Nuclear Information System (INIS) Black, D. 1993-04-01 Since the early 1940s, the contractors at the Hanford Site have been involved in the production and purification of nuclear defense materials. These production activities have resulted in the generation of large quantities of liquid and solid radioactive mixed waste (RMW). This waste is subject to regulation under authority of both the Resource Conservation and Recovery Act of 1976 2 (RCRA) and Atomic Energy Act 3 . This report covers mixed waste only. Hazardous waste that is not contaminated with radionuclides is not addressed in this report. The Washington State Department of Ecology, US Environmental Protection Agency, and US Department of Energy have entered into an agreement, the Hanford Federal Facility Agreement and Consent Order 1 (commonly referred to as the Tri-Party Agreement) to bring the Hanford Site operations into compliance with dangerous waste regulations. The Tri-Party Agreement required development of the original land disposal restrictions (LDR) plan and its annual updates to comply with LDR requirements for RMW. This report is the third update of the plan first issued in 1990. The Tri-Party Agreement requires, and the baseline plan and annual update reports provide, the information that follows: Waste characterization information; storage data; treatment information; waste reduction information; schedule; and progress 16. Radioactive waste management at the Hanford Reservation International Nuclear Information System (INIS) Anon. 1979-01-01 During some 30 years of plutonium production, the Hanford Reservation has accumulated large quantities of low- and high-level radioactive wastes. The high-level wastes have been stored in underground tanks, and the low-level wastes have been percolated into the soil. In recent years some programs for solidification and separation of the high-level wastes have been initiated. The Hanford waste-management system was studied by a panel of the Committee on Radioactive Waste Management of the National Academy of Sciences. The panel concluded that Hanford waste-management practices were adequate at present and for the immediate future but recommended increased research and development programs related to long-term isolation of the wastes. The panel also considered some alternatives for on-site disposal of the wastes. The Hanford Reservation was originally established for the production of plutonium for military purposes. During more than 30 years of operation, large volumes of high- and low-level radioactive wastes have been accumulated and contained at the site. The Management of these wastes has been the subject of controversy and criticism. To obtain a true technical evaluation of the Hanford waste situation, the Energy Research and Development Administration (now part of the Department of Energy) issued a contract to the National Academy of Sciences and the National Research Councilto conduct an independent review and evaluation of the Hanford waste-management practices and plans. A panel of the Committee on Radioactive Waste Management (CRWM) of the National Academy of Sciences conducted this study between the summer of 1976 and the summer of 1977. This article is a summary of the final report of that panel 17. Hanford Works monthly report, October 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-11-20 This is a progress report of the production reactors on the Hanford Reservation for the month of October 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 18. Hanford Works monthly report, December 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-01-22 This is a progress report of the production reactors on the Hanford Reservation for the month of December 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 19. Hanford Works monthly report, May 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-06-20 This is a progress report of the production reactors on the Hanford Reservation for the month of May 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 20. Hanford Works monthly report, July 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-08-18 This is a progress report of the production reactors on the Hanford Reservation for the month of July 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 1. Hanford Works monthly report, March 1952 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1952-04-18 This is a progress report of the production reactors on the Hanford Reservation for the month of April 1952. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 2. Hanford Works monthly report, April 1952 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1952-05-20 This is a progress report of the production reactors on the Hanford Reservation for the month of April 1952. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 3. Hanford Works monthly report, July 1952 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1952-08-15 This is a progress report of the production reactors on the Hanford Reservation for the month of July 1952. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 4. Hanford Works monthly report, January 1952 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1952-02-21 This is a progress report of the production reactors on the Hanford Reservation for the month of January 1952. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 5. Hanford Works monthly report, September 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-10-20 This is a progress report of the production reactors on the Hanford Reservation for the month of September 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 6. Hanford Works monthly report, July 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-08-24 This is a progress report of the production reactors on the Hanford Reservation for the month of July 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 7. Hanford Works monthly report, March 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-04-20 This is a progress report of the production reactors on the Hanford Reservation for the month of March 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 8. Hanford Works monthly report, June 1951 Energy Technology Data Exchange (ETDEWEB) 1951-07-20 This is a progress report of the production on the Hanford Reservation for the month of June 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 9. Hanford works monthly report, September 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-10-19 This is a progress report of the production reactors on the Hanford Reservation for the month of September 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 10. Hanford Works monthly report, May 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-06-21 This is a progress report of the production reactors on the Hanford Reservation for the month of May 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 11. Hanford Works monthly report, June 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-07-20 This is a progress report of the production reactors on the Hanford Reservation for the month of June 1950. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 12. Hanford Works monthly report, November 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-12-21 This is a progress report of the production reactors on the Hanford Reservation for the month of November 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 13. Hanford Works monthly report, August 1951 Energy Technology Data Exchange (ETDEWEB) 1951-09-24 This is a progress report of the production reactors on the Hanford Reservation for the month of August 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 14. Hanford Works monthly report, August 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-09-18 This is a progress report of the production reactors on the Hanford Reservation for the month of August 1950. This report takes each division (e.g. manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 15. Hanford Works monthly report, November 1950 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1950-12-20 This is a progress report of the production reactors on the Hanford Reservation for the month of November 1950. This report takes each division (e.g. manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 16. Hanford Works monthly report, February 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-03-20 This is a progress report of the production on the Hanford Reservation for the month of February 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 17. Hanford Works monthly report, December 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1952-01-22 This is a progress report of the production reactors on the Hanford Reservation for the month of December 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 18. Hanford Works monthly report, January 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-02-16 This is a progress report of the production reactors on the Hanford Reservation for the month of January 1951. This report takes each division (e.g. manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 19. Hanford Works monthly report, April 1951 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1951-05-21 This is a progress report of the production reactors on the Hanford Reservation for the month of April 1951. This report takes each division (e.g., manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. 20. Hanford Works monthly report, March 1949 Energy Technology Data Exchange (ETDEWEB) Prout, G.R. 1949-04-19 This is a progress report of the production reactors on the Hanford Reservation for the month of March 1949. This report takes each division (e.g. manufacturing, medical, accounting, occupational safety, security, reactor operations, etc.) of the site and summarizes its accomplishments and employee relations for that month. (MB) 1. Evolution of containment facilities for spectroscopic analysis at Rockwell Hanford Operations International Nuclear Information System (INIS) Hiller, J.M. 1984-01-01 The analysis of radioactive material requires much thought concerning getting the job done while still maintaining a safe working environment. A Rockwell Hanford Operations, several stages of evolution in instrumentation for spectroscopic elemental analysis have evolved, reflecting different philosophies respect to shielding and contamination control. Atomic absorption and inductively coupled plasma emission spectroscopic systems have been used for analyzing samples in support of a fission product recovery plant, nuclear waste processing and characterization programs, and U and Pu separation plants. Design thoughts, criticisms, and lessons learned in 20 years of containment for spectroscopic analysis are presented 2. Evolution of containment facilities for spectroscopic analysis at Rockwell Hanford Operations International Nuclear Information System (INIS) Hiller, J.M. 1984-01-01 The analysis of radioactive material requires much thought concerning getting the job done while still maintaining a safe working environment. At Rockwell Hanford Operations, we have gone through several stages of evolution in instrumentation for spectroscopic elemental analysis, reflecting different philosophies with respect to shielding and contamination control. Atomic absorption and inductively coupled plasma emission spectroscopic systems have been used for analyzing samples in support of a fission product recovery plant, nuclear waste processing and characterization programs, and U and Pu separation plants. Design thoughts, criticisms, and lessons learned in 20 years of containment for spectroscopic analysis are presented. 3 refs., 6 figs., 2 tabs 3. FLUOR HANFORD (FH) MAKES CLEANUP A REALITY IN NEARLY 11 YEARS AT HANFORD Energy Technology Data Exchange (ETDEWEB) GERBER, M.S. 2007-05-24 For nearly 11 years, Fluor Hanford has been busy cleaning up the legacy of nuclear weapons production at one of the Department of Energy's (DOE'S) major sites in the United States. As prime nuclear waste cleanup contractor at the vast Hanford Site in southeastern Washington state, Fluor Hanford has changed the face of cleanup. Fluor beginning on October 1, 1996, Hanford Site cleanup was primarily a ''paper exercise.'' The Tri-Party Agreement, officially called the Hanford Federal Facility Agreement and Consent Order - the edict governing cleanup among the DOE, U.S. Environmental Protection Agency (EPA) and Washington state - was just seven years old. Milestones mandated in the agreement up until then had required mainly waste characterization, reporting, and planning, with actual waste remediation activities off in the future. Real work, accessing waste ''in the field'' - or more literally in huge underground tanks, decaying spent fuel POO{approx}{approx}S, groundwater, hundreds of contaminated facilities, solid waste burial grounds, and liquid waste disposal sites -began in earnest under Fluor Hanford. The fruits of labors initiated, completed and/or underway by Fluor Hanford can today be seen across the site. Spent nuclear fuel is buttoned up in secure, dry containers stored away from regional water resources, reactive plutonium scraps are packaged in approved containers, transuranic (TRU) solid waste is being retrieved from burial trenches and shipped offsite for permanent disposal, contaminated facilities are being demolished, contaminated groundwater is being pumped out of aquifers at record rates, and many other inventive solutions are being applied to Hanford's most intransigent nuclear wastes. (TRU) waste contains more than 100 nanocuries per gram, and contains isotopes higher than uranium on the Periodic Table of the Elements. (A nanocurie is one-billionth of a curie.) At the same time, Fluor Hanford 4. Close-out report Fitzner-Eberhardt Arid Lands Ecology Reserve remedial action, Hanford, Washington International Nuclear Information System (INIS) 1996-04-01 The Fitzner-Eberhardt Arid Lands Ecology (ALE) Reserve consists of 312 km 2 (120 mi 2 ) of shrub-steppe land on the western edge of the Hanford Site. It is located south of Highway 240 and east of the point where the Yakima River borders the site. The land was set aside as a natural research area in 1967 by the Atomic Energy Commission. Historically tribal land, the area was homesteaded by pioneers before it was taken by the federal government in 1943 as a security buffer to protect the Hanford Site defense production facilities. One antiaircraft artillery battery (latter converted to a Nike missile site) was located on this land; plutonium production plants or storage facilities were never built there. A more complete account can be found in the Preliminary Assessment Screening (PAS) Report for the Arid Lands Ecology Reserve, Hanford. With the recent change in mission at the Hanford Site from plutonium production to environmental cleanup, much attention has been given to releasing clean tracts of land for other uses. The ALE Reserve is one such tract of land. The existing areas of contamination in the ALE Reserve were considered to be small. In March 1993, the U.S. Department of Energy (DOE), the Environmental Protection Agency (EPA), and the Washington State Department of Ecology (Ecology) signed and Agreement in Principle in which they agreed to expedite cleanup of the ALE Reserve. Cleanup activities and a draft close-out report were to be completed by October 1994. Additionally, DOE proposed to mitigate hazards that may pose a physical threat to wildlife or humans 5. Review of Hanford international activities International Nuclear Information System (INIS) Panther, D.G. 1993-01-01 Hanford initiated a review of international activities to collect, review, and summarize information on international environmental restoration and waste management initiatives considered for use at Hanford. This effort focused on Hanford activities and accomplishments, especially international technical exchanges and/or the implementation of foreign-developed technologies 6. Binary encounter electron production in ion-atom collisions International Nuclear Information System (INIS) Grabbe, S.; Bhalla, C.P.; Shingal, R. 1993-01-01 The binary encounter electrons are produced by hard collisions between the target electrons and the energetic projectiles. Richard et al. found the measured double differential cross section for BEe production at zero degree laboratory scattering angle, in collisions of F q+ with H 2 and He targets, to increase as the charge state of the projectile was decreased. The binary encounter electron production has recently been a subject of detailed investigations. We have calculated the differential elastic scattering cross sections of electrons from several ions incorporating the exchange contribution of the continuum and the bound orbitals in addition to the static potential. The double differential binary encounter electron production cross sections are presented using the impulse approximation 7. Hanford groundwater scenario studies International Nuclear Information System (INIS) Arnett, R.C.; Gephart, R.E.; Deju, R.A.; Cole, C.R.; Ahlstrom, S.W. 1977-05-01 This report documents the results of two Hanford groundwater scenario studies. The first study examines the hydrologic impact of increased groundwater recharge resulting from agricultural development in the Cold Creek Valley located west of the Hanford Reservation. The second study involves recovering liquid radioactive waste which has leaked into the groundwater flow system from a hypothetical buried tank containing high-level radioactive waste. The predictive and control capacity of the onsite Hanford modeling technology is used to evaluate both scenarios. The results of the first study indicate that Cold Creek Valley irrigationis unlikely to cause significant changes in the water table underlying the high-level waste areas or in the movement of radionuclides already in the groundwater. The hypothetical tank leak study showed that an active response (in this case waste recovery) can be modeled and is a possible alternative to passive monitoring of radionuclide movement in the unlikely event that high-level waste is introduced into the groundwater 8. Hanford Area 2000 Population International Nuclear Information System (INIS) Elliott, Douglas B.; Scott, Michael J.; Antonio, Ernest J.; Rhoads, Kathleen 2004-01-01 This report was prepared for the U.S. Department of Energy (DOE) Richland Operations Office, Surface Environmental Surveillance Project, to provide demographic data required for ongoing environmental assessments and safety analyses at the DOE Hanford Site near Richland, Washington. This document includes 2000 Census estimates for the resident population within an 80-kilometer (50-mile) radius of the Hanford Site. Population distributions are reported relative to five reference points centered on meteorological stations within major operating areas of the Hanford Site - the 100 F, 100 K, 200, 300, and 400 Areas. These data are presented in both graphical and tabular format, and are provided for total populations residing within 80 km (50 mi) of the reference points, as well as for Native American, Hispanic and Latino, total minority, and low-income populations 9. Hanford Waste Vitrification Plant International Nuclear Information System (INIS) Larson, D.E.; Allen, C.R.; Kruger, O.L.; Weber, E.T. 1991-10-01 The Hanford Waste Vitrification Plant (HWVP) is being designed to immobilize pretreated Hanford high-level waste and transuranic waste in borosilicate glass contained in stainless steel canisters. Testing is being conducted in the HWVP Technology Development Project to ensure that adapted technologies are applicable to the candidate Hanford wastes and to generate information for waste form qualification. Empirical modeling is being conducted to define a glass composition range consistent with process and waste form qualification requirements. Laboratory studies are conducted to determine process stream properties, characterize the redox chemistry of the melter feed as a basis for controlling melt foaming and evaluate zeolite sorption materials for process waste treatment. Pilot-scale tests have been performed with simulated melter feed to access filtration for solids removal from process wastes, evaluate vitrification process performance and assess offgas equipment performance. Process equipment construction materials are being selected based on literature review, corrosion testing, and performance in pilot-scale testing. 3 figs., 6 tabs 10. Synthetic Strategies toward Natural Products Containing Contiguous Stereogenic Quaternary Carbon Atoms. Science.gov (United States) Büschleb, Martin; Dorich, Stéphane; Hanessian, Stephen; Tao, Daniel; Schenthal, Kyle B; Overman, Larry E 2016-03-18 Strategies for the total synthesis of complex natural products that contain two or more contiguous stereogenic quaternary carbon atoms in their intricate structures are reviewed with 12 representative examples. Emphasis has been put on methods to create quaternary carbon stereocenters, including syntheses of the same natural product by different groups, thereby showcasing the diversity of thought and individual creativity. A compendium of selected natural products containing two or more contiguous stereogenic quaternary carbon atoms and key reactions in their total or partial syntheses is provided in the Supporting Information. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim. 11. Quality of worklife and productivity in drilling in Atomic Minerals Division International Nuclear Information System (INIS) Singh, Rajendra. 1992-01-01 The existing trend of drilling output in Atomic Minerals Division has been briefly discussed and human resources and its bearing on productivity in our drilling efforts emphasised. Remedial measures concerning the human resources development including creation of cells at regional and zonal levels have been suggested to give effective shape to all aspects of quality of worklife and productivity. (author) 12. Coherent anti-Stokes Raman scattering (CARS) detection or hot atom reaction product internal energy distributions International Nuclear Information System (INIS) Quick, C.R. Jr.; Moore, D.S. 1983-01-01 Coherent anti-Stokes Raman spectroscopy (CARS) is being utilized to investigate the rovibrational energy distributions produced by reactive and nonreactive collisions of translationally hot atoms with simple molecules. Translationally hot H atoms are produced by ArF laser photolysis of HBr. Using CARS we have monitored, in a state-specific and time-resolved manner, rotational excitation of HBr (v = 0), vibrational excitation of HBr and H 2 , rovibrational excitation of H 2 produced by the reaction H + HBr → H 2 + Br, and Br atom production by photolysis of HBr 13. Resource book: Decommissioning of contaminated facilities at Hanford International Nuclear Information System (INIS) 1991-09-01 In 1942 Hanford was commissioned as a site for the production of weapons-grade plutonium. The years since have seen the construction and operation of several generations of plutonium-producing reactors, plants for the chemical processing of irradiated fuel elements, plutonium and uranium processing and fabrication plants, and other facilities. There has also been a diversification of the Hanford site with the building of new laboratories, a fission product encapsulation plant, improved high-level waste management facilities, the Fast Flux test facility, commercial power reactors and commercial solid waste disposal facilities. Obsolescence and changing requirements will result in the deactivation or retirement of buildings, waste storage tanks, waste burial grounds and liquid waste disposal sites which have become contaminated with varying levels of radionuclides. This manual was established as a written repository of information pertinent to decommissioning planning and operations at Hanford. The Resource Book contains, in several volumes, descriptive information of the Hanford Site and general discussions of several classes of contaminated facilities found at Hanford. Supplementing these discussions are appendices containing data sheets on individual contaminated facilities and sites at Hanford. Twelve appendices are provided, corresponding to the twelve classes into which the contaminated facilities at Hanford have been organized. Within each appendix are individual data sheets containing administrative, geographical, physical, radiological, functional and decommissioning information on each facility within the class. 68 refs., 54 figs., 18 tabs 14. Resource book: Decommissioning of contaminated facilities at Hanford International Nuclear Information System (INIS) 1991-09-01 In 1942 Hanford was commissioned as a site for the production of weapons-grade plutonium. The years since have seen the construction and operation of several generations of plutonium-producing reactors, plants for the chemical processing of irradiated fuel elements, plutonium and uranium processing and fabrication plants, and other facilities. There has also been a diversification of the Hanford site with the building of new laboratories, a fission product encapsulation plant, improved high-level waste management facilities, the Fast Flux test facility, commercial power reactors and commercial solid waste disposal facilities. Obsolescence and changing requirements will result in the deactivation or retirement of buildings, waste storage tanks, waste burial grounds and liquid waste disposal sites which have become contaminated with varying levels of radionuclides. This manual was established as a written repository of information pertinent to decommissioning planning and operations at Hanford. The Resource Book contains, in several volumes, descriptive information of the Hanford Site and general discussions of several classes of contaminated facilities found at Hanford. Supplementing these discussions are appendices containing data sheets on individual contaminated facilities and sites at Hanford. Twelve appendices are provided, corresponding to the twelve classes into which the contaminated facilities at Hanford have been organized. Within each appendix are individual data sheets containing administrative, geographical, physical, radiological, functional and decommissioning information on each facility within the class. 49 refs., 44 figs., 14 tabs 15. Nuclear isotope measurement in the Hanford environment International Nuclear Information System (INIS) Wacker, J.F.; Stoffel, J.J.; Kelley, J.M. 1995-01-01 The Pacific Northwest Laboratory (PNL) is located at the federal government's Hanford Site in southeastern Washington State, which was built during World War II as part of the secret Manhattan Project to develop the atomic bomb. Monitoring of the Site itself and surrounding environs for Hanford-related radionuclides has been a routine part of the operations since 1944. One of the most sensitive analytical methods used is thermal ionization mass spectrometry (TIMS) with triple-sector mass spectrometers. Normal geometry instruments have an abundance sensitivity of 10 -9 for uranium while the authors' newest Triple-Sector Isotope Mass Spectrometer (TRISM), utilizing a new ion-optical design developed at PNL, has an abundance sensitivity of 10 -11 . In favorable cases, sensitivity is such that complete isotopic analyses are obtained on total samples in the femtogram range; and minor isotopes in the attogram range are measured 16. An overview of the Hanford controversy International Nuclear Information System (INIS) Stewart, A.M.; Kneale, G.W. 1991-01-01 In 1964, the Atomic Energy Commission agreed to sponsor 'a study of the lifetime health and mortality experiences of all employees of AEC contractors.' The commission put in charge of this study a physician (Thomas Mancuso) who had recently shown how the U.S. Social Security system could be used to identify the dates and causes of death of all insured workers. As director of the AEC project, Mancuso was at liberty to include any or all the postwar offshoots of the Manhattan Project. His master plan included workers from Oak Ridge, Los Alamos, and Hanford, but it soon became apparent that his attempts to link radiation exposures to subsequent events were proving more successful at Hanford than elsewhere. The authors of this paper, who participated in the study, review the controversy surrounding its eventual publication.22 references 17. Hanford Engineer Works technical manual Energy Technology Data Exchange (ETDEWEB) 1944-05-01 The uranium metal, as discharged from the piles in the 100 Areas, contains the alpha emitting product, plutonium, in concentration in the neighborhood of 150--250 grams per metric ton, along with similar amounts of beta and gamma fission elements. It is the purpose of the Separations Plant to effect the separation of this product from the uranium metal and fission elements, and to prepare a concentrated, relatively pure solution of plutonium nitrate as the final product of the Hanford Plant. This section of the manual discusses the chemistry of the separations process, describes the buildings and equipment provided for carrying out the various steps in the operation, and presents the detailed operating procedures used. There are included, in many instances, references to other documents presenting a more detailed view of a specific point in the process. 18. Hazardous chemical and radioactive wastes at Hanford International Nuclear Information System (INIS) Keller, J.F.; Stewart, T.L. 1991-07-01 The Hanford Site was established in 1944 to produce plutonium for defense. During the past four decades, a number of reactors, processing facilities, and waste management facilities have been built at Hanford for plutonium production. Generally, Hanford's 100 Area was dedicated to reactor operation; the 200 Area to fuel reprocessing, plutonium recovery, and waste management; and the 300 Area to fuel fabrication and research and development. Wastes generated from these operations included highly radioactive liquid wastes, which were discharged to single- and double-shell tanks; solid wastes, including both transuranic (TRU) and low-level wastes, which were buried or discharged to caissons; and waste water containing low- to intermediate-level radioactivity, which was discharged to the soil column via near-surface liquid disposal units such as cribs, ponds, and retention basins. Virtually all of the wastes contained hazardous chemical as well as radioactive constituents. This paper will focus on the hazardous chemical components of the radioactive mixed waste generated by plutonium production at Hanford. The processes, chemicals used, methods of disposition, fate in the environment, and actions being taken to clean up this legacy are described by location 19. Hazardous chemical and radioactive wastes at Hanford International Nuclear Information System (INIS) Keller, J.F.; Stewart, T.L. 1993-01-01 The Hanford Site was established in 1944 to produce plutonium for defense. During the past four decades, a number of reactors, processing facilities, and waste management facilities were built at Hanford for plutonium production. Generally, Hanford's 100 Area was dedicated to reactor operation; the 200 Areas to fuel reprocessing, plutonium recovery, and waste management; and the 300 Area to fuel fabrication and research and development. Wastes generated from these operations included highly radioactive liquid wastes, which were discharged to single- and double-shell tanks; solid wastes, including both transuranic and low-level wastes, which were buried or discharged to caissons; and waste water containing low- to intermediate-level radioactivity, which was discharged to the soil column via near-surface liquid disposal units such as cribs, ponds, and retention basins. Virtually all of the wastes contained hazardous chemicals as well as radioactive constituents. This paper focuses on the hazardous chemical components of the radioactive mixed waste generated by plutonium production at Hanford. The processes, chemicals used, methods of disposition, fate in the environment, and actions being taken to clean up this legacy are described by location 20. Nonlinear effect of pion production in collisions of atomic nuclei International Nuclear Information System (INIS) Grin', Yu.T. 1982-01-01 The phenomenon of pion production in relativistic nucleon-nucleus and nucleus-nucleus interactions is investigated. The present experimental data are analyzed. It is shown that average multiplicity of pions in the (p, C), (C, C) collision reactions with the momentum p=4.2 GeV/cA and (p, Ar), (Ar, KCl) with the momentum p=2.3 GeV/cA non-linearly depends on the nucleon number. The calculated values of average multiplicity of negative pions per one nucleon of nucleus-pro ectile, probability of pion production and number of nucleon interactions for the investigated reactions are presented as a table. A comparative analysis of average multiplicities of pions per nucleon-participant in the nucleon-nucleus and nucleus-nucleus reactions at the p=2.3 GeV/cA momentum for argon and at the p=4.2 GeV/cA for carbon reveals that decrease of multiplicity by 30-35% is observed in nucleus-nucleus collision. Non-linearity is associated with decrease of effective interaction of each incident nucleon in the collision of nuclei as compared with the number of nucleon interactions in the ''elementary'' nucleon-nucleus reaction. Knock-out of nucleons from the colliding nuclei is the most probable reason for the decrease of the number of interactions 1. DOE wants Hanford change International Nuclear Information System (INIS) Anon. 1994-01-01 Nine months ago, Energy Secretary Hazel O'Leary promised local officials running the agency's huge Hanford, Washington, weapon complex more control in directing its projected 57-billion waste cleanup. Earlier this month, she returned to the site for a follow-on open-quotes summit,close quotes this time ordering teamwork with contractors, regulators and local activities 2. Production and transport chemistry of atomic fluorine in remote plasma source and cylindrical reaction chamber International Nuclear Information System (INIS) Gangoli, S P; Johnson, A D; Fridman, A A; Pearce, R V; Gutsol, A F; Dolgopolsky, A 2007-01-01 Increasingly, NF 3 -based plasmas are being used in semiconductor manufacturing to clean chemical vapour deposition (CVD) chambers. With advantages such as faster clean times, substantially lower emissions of gases having high global warming potentials, and reduced chamber damage, NF 3 plasmas are now favoured over fluorocarbon-based processes. Typically, a remote plasma source (RPS) is used to dissociate the NF 3 gas and produce atomic fluorine that etches the CVD residues from the chamber surfaces. However, it is important to efficiently transport F atoms from the plasma source into the process chamber. The current work is aimed at understanding and improving the key processes involved in the production and transport of atomic fluorine atoms. A zero-dimensional model of NF 3 dissociation and F production chemistry in the RPS is developed based on various known and derived plasma parameters. Additionally, a model describing the transport of atomic fluorine is proposed that includes both physical (diffusion, adsorption and desorption) and chemical processes (surface and three-body volume recombination). The kinetic model provides an understanding of the impact of chamber geometry, gas flow rates, pressure and temperature on fluorine recombination. The plasma-kinetic model is validated by comparing model predictions (percentage F atom density) with experimental results (etch rates) 3. Laser diagnostics of atomic hydrogen and oxygen production in rf and microwave plasma discharges International Nuclear Information System (INIS) Preppernau, B.L. 1993-01-01 The research for this thesis involved the application of two-photon allowed laser-induced fluorescence (TALIF) to the study of atomic hydrogen and oxygen production in industrial scale radio-frequency and microwave plasma discharge apparatus. Absolute atomic hydrogen concentration profiles were measured in a Gaseous Electronics Conference Reference Cell installed at Wright-Patterson AFB, Ohio operating with a simple H 2 discharge. Two-dimensional atomic hydrogen concentration profiles were also measured in an ASTEX HPMM microwave plasma diamond deposition reactor during actual diamond growth. In addition absolute atomic oxygen concentrations were measured in the ASTEX system. Particular attention as paid to refining the concentration calibration technique and in determining a correction to account for the collisional quenching of excited state fluorescence in high pressure gases 4. Proposal for the International Atomic Energy Agency Training Course International Nuclear Information System (INIS) McCarthy, T.L. 1994-06-01 The Hanford Site has hosted similar activities, including both Hanford Summits I and II. The Hanford Summits were two-day televised events to discuss the commitment of the current Presidential administration to the environmental restoration of the Hanford Site. Public involvement and strategic issues established from Hanford Summit I include: Regulatory issues, training and education, economic development and partnership, and technology transfer. Hanford Summit II provided a summary of how Secretary of Energy O'Leary is proceeding on the above strategic issues. The DOE and Westinghouse School for Environmental Excellence frequently offers a six-week course for environmental professionals and workers. Approximately thirty to forty individuals attend the training course, which provides training in environmental regulation compliance. The Hanford Site has hosted two previous International Atomic Energy Agency training courses. The courses lasted two weeks and had approximately eight to ten participants. Nuclear Material Management and Neutron Monitoring were the courses hosted by the Hanford Site 5. Hanford Site Groundwater Monitoring for Fiscal Year 2002 Energy Technology Data Exchange (ETDEWEB) Hartman, Mary J.; Morasch, Launa F.; Webber, William D. 2003-02-28 This report presents the results of groundwater and vadose zone monitoring and remediation for fiscal year 2002 on the U.S. Department of Energy's Hanford Site in Washington State. This report is written to meet the requirements in CERCLA, RCRA, the Atomic Energy Act of 1954, and Washington State Administrative Code. 6. Update on worker mortality data at Hanford International Nuclear Information System (INIS) Gilbert, E.S. 1979-01-01 The subject of this paper is a study of the effects on mortality of occupational exposure to ionizing radiation at the Hanford plant. The Hanford plant, which is located in southeastern Washington State, was established in the early forties as an installation for plutonium production. Many workers employed by the various contractors hold jobs involving some exposure to radiation. Yearly records of this exposure, obtained from dosimeter readings, as well as occupational data, are maintained for all employees. Mortality data are obtained by having the Social Security Administration periodically search their records for deaths of persons identified in the personnel rosters of Hanford contractors. Published analyses of worker mortality at Hanford have included workers initially employed before 1965 and mortality up to April 1, 1974. In this paper, the mortality data are updated to include deaths up to May 1, 1977, workers employed 1965 and later, and the most recent exposure data. In addition to updating results of earlier analyses, this paper provides a discussion of the problems involved in analyzing and interpreting occupational exposure and mortality data. For a more detailed discussion of these problems the reader is referred to the papers noted above 7. Enantioselective H-atom transfer reaction: a strategy to synthesize formaldehyde aldol products. Science.gov (United States) Sibi, Mukund P; Patil, Kalyani 2005-04-14 [reaction: see text] Enantioselective radical alkylation of Baylis-Hillman adducts furnished aldol products in good yield and selectivity. The results illustrate that the selectivity in the hydrogen atom transfer is dependent on the size of the ester substituent, with smaller substituents providing better enantioselectivity. 8. Hanford well custodians. Revision 1 International Nuclear Information System (INIS) Schatz, A.L.; Underwood, D.J. 1995-01-01 The Hanford Site Groundwater Protection Management Program recognized the need to integrate monitoring well activities in a centralized manner. A key factor to Hanford Site well integration was the need to clearly identify a responsible party for each of the wells. WHC was asked to identify all wells on site, the program(s) using each well, and the program ultimately responsible for the well. This report lists the custodian and user(s) for each Hanford well and supplies a comprehensive list of all decommissioned and orphaned wells on the Hanford Site. This is the first update to the original report released in December 1993 9. Hanford process review International Nuclear Information System (INIS) 1991-12-01 This report is a summary of past incidents at the US Department of Energy's (DOE) Hanford Site. The purpose of the report is to provide the major, significant, nuclear-safety-related incidents which incurred at the Hanford Site in a single document for ease of historical research. It should be noted that the last major accident occurred in 1980. This document is a summary of reports released and available to the public in the DOE Headquarters and Richland public reading rooms. This document provides no new information that has not previously been reported. This report is not intended to cover all instances of radioactivity release or contamination, which are already the subject of other major reviews, several of which are referenced in Section 1.3 10. Mixed waste management at the Hanford Site International Nuclear Information System (INIS) Roberts, R.J.; Jasen, W.G. 1991-01-01 Various types of waste have been generated during the 50-year history of the Hanford Site. Regulatory changes in the last 20 years have provided the emphasis for better management of these wastes. Interpretations of the Atomic Energy Act (AEA) and the Resource Conservation and Recovery Act (RCRA) have led to the definition of a group of wastes called radioactive mixed wastes (RMW). As a result of the radioactive and hazardous properties of these wastes, special projects have been initiated for the management of RMW. This paper addresses the management of solid RMW. The management of bulk liquid RMW will not be described. 7 refs., 4 figs 11. Hanford Tank Cleanup Update International Nuclear Information System (INIS) Berriochoa, M.V. 2011-01-01 Access to Hanford's single-shell radioactive waste storage tank C-107 was significantly improved when workers completed the cut of a 55-inch diameter hole in the top of the tank. The core and its associated cutting equipment were removed from the tank and encased in a plastic sleeve to prevent any potential spread of contamination. The larger tank opening allows use of a new more efficient robotic arm to complete tank retrieval. 12. Considerations on post-production obligations in terms of Atomic Energy Law International Nuclear Information System (INIS) Rebentisch, M. 1992-01-01 The article describes and evaluates the laws concerning the decommissioning and dismantling of nuclear power plants and offers suggestions for possible new regulations. The contribution examines fundamental legal aspects, the instrumentalization of post-production obligations in terms of atomic energy laws, especially the question as to how to bring the Atomic Energy Law into accord with the Federal Emission Control Law within the realm of decommissioning laws, licences for safe confinement and dismantling of a plant, and in addition questions on making financial provisions for decommissioning. (orig./HSCH) [de 13. Improved production of Br atoms near zero speed by photodissociating laser aligned Br2 molecules. Science.gov (United States) Deng, L Z; Yin, J P 2014-10-28 We theoretically investigated the improvement on the production rate of the decelerated bromine (Br) atoms near zero speed by photodissociating laser aligned Br2 precursors. Adiabatic alignment of Br2 precursors exposed to long laser pulses with duration on the order of nanoseconds was investigated by solving the time-dependent Schrödinger equation. The dynamical fragmentation of adiabatically aligned Br2 precursors was simulated and velocity distribution of the Br atoms produced was analyzed. Our study shows that the larger the degree of the precursor alignment, ⟨cos(2) θ⟩, the higher the production rate of the decelerated Br atoms near zero speed. For Br2 molecules with an initial rotational temperature of ~1 K, a ⟨cos(2) θ⟩ value of ~0.88 can result in an improvement factor of over ~20 on the production rate of the decelerated Br atoms near zero speed, requiring a laser intensity of only ~1 × 10(12) W/cm(2) for alignment. 14. Production of antihydrogen at reduced magnetic field for anti-atom trapping CERN Document Server Andresen, G.B.; Boston, A.; Bowe, P.D.; Cesar, C.L.; Chapman, S.; Charlton, M.; Chartier, M.; Deutsch, A.; Fajans, J.; Fujiwara, M.C.; Funakoshi, R.; Gill, D.R.; Gomberoff, K.; Hangst, J.S.; Hayano, R.S.; Hydomako, R.; Jenkins, M.J.; Jorgensen, L.V.; Kurchaninov, L.; Madsen, N.; Nolan, P.; Olchanski, K.; Olin, A.; Page, R.D.; Povilus, A.; Robicheaux, F.; Sarid, E.; Silveira, D.M.; Storey, J.W.; Thompson, R.I.; van der Werf, D.P.; Wurtele, J.S.; Yamazaki, Y. 2008-01-01 We have demonstrated production of antihydrogen in a 1,$T solenoidal magnetic field. This field strength is significantly smaller than that used in the first generation experiments ATHENA (3$,$T) and ATRAP (5$,$T). The motivation for using a smaller magnetic field is to facilitate trapping of antihydrogen atoms in a neutral atom trap surrounding the production region. We report the results of measurements with the ALPHA (Antihydrogen Laser PHysics Apparatus) device, which can capture and cool antiprotons at 3$,$T, and then mix the antiprotons with positrons at 1$,T. We infer antihydrogen production from the time structure of antiproton annihilations during mixing, using mixing with heated positrons as the null experiment, as demonstrated in ATHENA. Implications for antihydrogen trapping are discussed. 15. Atomic risk insurance. Risk policy, safety production and expertise in Germany and the USA 1945 - 1986 International Nuclear Information System (INIS) Wehner, Christoph 2017-01-01 The book covers the following chapters: (I) Between threat and promise: Political change and the corporate perception, the burden of the atomic bomb, promise of nuclear energy risk criticism in the pre-ecological phase, nuclear risk as investment restraint; (II) Risk policy at the insurability limit: hazard knowledge, safety production and insurance expertise in the German nuclear policy (1955-1962); (III) Risk policy beyond the catastrophe, insurability interpretation, concepts and conflicts (1957-1968); (IV) Scandalization of risk policy: safety production, confidence and expertise in the nuclear controversial debate (1969 - 1979); (V) Nuclear risk policy and the challenge of the ''risk society'' (1975-1986); (VI) From safety production to hazard probe: atomic energy And the change of insurance. 16. Atomic oxygen production scaling in a nanosecond-pulsed externally grounded dielectric barrier plasma jet Science.gov (United States) Sands, Brian; Schmidt, Jacob; Ganguly, Biswa; Scofield, James 2014-10-01 Atomic oxygen production is studied in a capillary dielectric barrier plasma jet that is externally grounded and driven with a 20-ns risetime positive unipolar pulsed voltage at pulse repetition rates up to 25 kHz. The power coupled to the discharge can be easily increased by increasing the pulse repetition rate. At a critical turnover frequency, determined by the net energy density coupled to the discharge, the plasma chemistry abruptly changes. This is indicated by increased plasma conductance and a transition in reactive oxygen species production from an ozone-dominated production regime below the turnover frequency to atomic-oxygen-dominated production at higher pulse rates. Here, we characterize atomic oxygen production scaling using spatially- and temporally-resolved two-photon absorption laser-induced-fluorescence (TALIF). Quantitative results are obtained via calibration with xenon using a similar laser excitation and collection system. These results are compared with quantitative ozone and discharge power measurements using a helium gas flow with oxygen admixtures up to 3%. 17. Removing Phosphate from Hanford High-Phosphate Tank Wastes: FY 2010 Results Energy Technology Data Exchange (ETDEWEB) Lumetta, Gregg J.; Braley, Jenifer C.; Edwards, Matthew K.; Qafoku, Odeta; Felmy, Andrew R.; Carter, Jennifer C.; MacFarlan, Paul J. 2010-09-22 The U.S. Department of Energy (DOE) is responsible for environmental remediation at the Hanford Site in Washington State, a former nuclear weapons production site. Retrieving, processing, immobilizing, and disposing of the 2.2 × 105 m3 of radioactive wastes stored in the Hanford underground storage tanks dominates the overall environmental remediation effort at Hanford. The cornerstone of the tank waste remediation effort is the Hanford Tank Waste Treatment and Immobilization Plant (WTP). As currently designed, the capability of the WTP to treat and immobilize the Hanford tank wastes in the expected lifetime of the plant is questionable. For this reason, DOE has been pursuing supplemental treatment options for selected wastes. If implemented, these supplemental treatments will route certain waste components to processing and disposition pathways outside of WTP and thus will accelerate the overall Hanford tank waste remediation mission. 18. Rate Constants and H-Atom Product Yields for the Reactions of O(1D) Atoms with Ethane and Acetylene from 50 to 296 K. Science.gov (United States) Nunez-Reyes, Dianailys; Hickson, Kevin M 2018-05-01 The gas phase reactions of atomic oxygen in its first excited state with ethane and acetylene have been investigated in a continuous supersonic flow reactor over the temperature range 50 K to 296 K. O(1D) atoms were produced by pulsed laser photolysis of ozone at 266 nm. Two different types of experiments, kinetics measurements and H-atom product yield determinations, were performed by detecting O(1D) atoms and H(2S) atoms respectively by vacuum ultraviolet laser induced fluorescence. The measured rate constants are in agreement with previous work at room temperature and little or no temperature dependence was observed as the temperature is decreased to 50 K. H-atoms yields were found to be independent of temperature for the reaction of O(1D) with ethane. These product yields are discussed in the context of earlier dynamics measurements at higher temperature. Due to the influence of secondary reactions, no H-atom yields could be obtained for the reaction of O(1D) with acetylene. 19. Catalog of borehole lithologic logs from the 600 Area, Hanford Site International Nuclear Information System (INIS) Fecht, K.R.; Lillie, J.T. 1982-03-01 Rockwell Hanford Operations (Rockwell) geoscientists are studying the Hanford Site subsurface environment to assure safe management operations, disposal, and storage of radioactive waste. As part of this effort, geoscientists have collected geotechnical data from about 3000 boreholes drilled on the Hanford Site since the early 1900s. These boreholes have been used for subsurface geologic, hydrologic, and engineering investigation, water supply, ground-water monitoring, and natural gas production. This report is a catalog of all obtainable (about 800) lithologic logs from boreholes in a portion of the Hanford Site known as the 600 Area 20. Mortality studies of Hanford workers International Nuclear Information System (INIS) Gilbert, E.S. 1986-04-01 Radiation exposures at Hanford have been deliberately limited as a protection to the worker. This means that if current estimates of radiation risks, which have been determined by national and international groups, are correct, it's highly unlikely that noticeable radiation-induced health effects will be identified among Hanford workers. 1 fig., 4 tabs 1. Development of the process for production of UO2 powder by atomization of uranyl nitrate International Nuclear Information System (INIS) Oliveira Lainetti, P.E. de. 1991-01-01 A method of direct conversion of uranyl nitrate hexahydrate (UNH) solution to ceramic grade uranium dioxide powders by thermal denitration in a furnace that combines atomization nozzle and a gas stirred bed is described. The main purpose of this work is to show that this alternative process is technically viable, specially if the recovery of the scrap generated in the nuclear fuel pellet production is required, without further generation of new liquid wastes. The steps for the development of the denitration unit as well as the characteristics of the final powders are described. Powder production experiments have been carried out for different atomization gas pressures and furnace upper section temperatures. Determination of impurity content, specific surface area, particle size and pore size distribution, density, U content, and O/U rate of uranium dioxide powders have been done; phase identification and morphology studies have also been performed. Sintered pellets have been studied by hydrostatic density determination and microstructure analyses. (author) 2. Characterization plan for Hanford spent nuclear fuel International Nuclear Information System (INIS) Abrefah, J.; Thornton, T.A.; Thomas, L.E.; Berting, F.M.; Marschman, S.C. 1994-12-01 Reprocessing of spent nuclear fuel (SNF) at the Hanford Site Plutonium-Uranium Extraction Plant (PUREX) was terminated in 1972. Since that time a significant quantity of N Reactor and Single-Pass Reactor SNF has been stored in the 100 Area K-East (KE) and K-West (KW) reactor basins. Approximately 80% of all US Department of Energy (DOE)-owned SNF resides at Hanford, the largest portion of which is in the water-filled KE and KW reactor basins. The basins were not designed for long-term storage of the SNF and it has become a priority to move the SNF to a more suitable location. As part of the project plan, SNF inventories will be chemically and physically characterized to provide information that will be used to resolve safety and technical issues for development of an environmentally benign and efficient extended interim storage and final disposition strategy for this defense production-reactor SNF 3. Urgent problems of radioecology concerned with the problems of the Atomic Energy production International Nuclear Information System (INIS) Aleksakhin, R.M.; Polikarpov, G.G. 1982-11-01 Fundamentals tasks of contemporary radioecology concerning migration of natural and artificial radionuclides and the effect of ionizing radiation on natural biogeocenosis are expounded which arose from the developing production and uses of atomic energy. The authors discuss the problems of ecological control over radiation affection of ecosystems and present the classification of natural areas according to their ecological condition. The authors also stress the urgency of studies of migration in the biosphere of radionuclides of the complete nuclear fuel turnover [fr 4. K-Vacancy Production in the Collision of Highly Charged Relativistic Ions With Heavy Atoms OpenAIRE KHABIBULLAEV, P. K. 2014-01-01 A general expression for the cross section of the inelastic collision of relativistic highly charged ion with heavy (relativistic) atoms is obtained using the generalized eikonal approximation. In the ultrarelativistic limit, the obtained formula coincides with a known exact one. As an application of the obtained result, probability and cross section of the K-vacany production in the U92+ - U91+ collision are calculated. 5. Trends in actinide processing at Hanford International Nuclear Information System (INIS) Harmon, H.D. 1993-09-01 In 1989, the mission at the Hanford Site began a dramatic and sometimes painful transition. The days of production--as we used to know it--are over. Our mission officially has become waste management and environmental cleanup. This mission change didn't eliminate many jobs--in fact, budgets have grown dramatically to support the new mission. Most all of the same skilled crafts, engineers, and scientists are still required for the new mission. This change has not eliminated the need for actinide processing, but it has certainly changed the focus that our actinide chemists and process engineers have. The focus used to be on such things as increasing capacity, improving separations efficiency, and product purity. Minimizing waste had become a more important theme in recent years and it is still a very important concept in the waste management and environmental cleanup arena. However, at Hanford, a new set of words dominates the actinide process scene as we work to deal with actinides that still reside in a variety of forms at the Hanford Site. These words are repackage, stabilize, remove, store and dispose. Some key activities in each of these areas are described in this report 6. Nonlinear effects in defect production by atomic and molecular ion implantation International Nuclear Information System (INIS) David, C.; Dholakia, Manan; Chandra, Sharat; Nair, K. G. M.; Panigrahi, B. K.; Amirthapandian, S.; Amarendra, G.; Varghese Anto, C.; Santhana Raman, P.; Kennedy, John 2015-01-01 This report deals with studies concerning vacancy related defects created in silicon due to implantation of 200 keV per atom aluminium and its molecular ions up to a plurality of 4. The depth profiles of vacancy defects in samples in their as implanted condition are carried out by Doppler broadening spectroscopy using low energy positron beams. In contrast to studies in the literature reporting a progressive increase in damage with plurality, implantation of aluminium atomic and molecular ions up to Al 3 , resulted in production of similar concentration of vacancy defects. However, a drastic increase in vacancy defects is observed due to Al 4 implantation. The observed behavioural trend with respect to plurality has even translated to the number of vacancies locked in vacancy clusters, as determined through gold labelling experiments. The impact of aluminium atomic and molecular ions simulated using MD showed a monotonic increase in production of vacancy defects for cluster sizes up to 4. The trend in damage production with plurality has been explained on the basis of a defect evolution scheme in which for medium defect concentrations, there is a saturation of the as-implanted damage and an increase for higher defect concentrations 7. Hanford tanks initiative plan International Nuclear Information System (INIS) McKinney, K.E. 1997-01-01 Abstract: The Hanford Tanks Initiative (HTI) is a five-year project resulting from the technical and financial partnership of the U.S. Department of Energy's Office of Waste Management (EM-30) and Office of Science and Technology Development (EM-50). The HTI project accelerates activities to gain key technical, cost performance, and regulatory information on two high-level waste tanks. The HTI will provide a basis for design and regulatory decisions affecting the remainder of the Tank Waste Remediation System's tank waste retrieval Program 8. Sympathetic cooling in a rubidium cesium mixture: Production of ultracold cesium atoms International Nuclear Information System (INIS) Haas, M. 2007-01-01 This thesis presents experiments for the production of ultracold rubidium cesium mixture in a magnetic trap. The long-termed aim of the experiment is the study of the interaction of few cesium atoms with a Bose-Einstein condensate of rubidium atoms. Especially by controlled variation of the cesium atom number the transition in the description of the interaction by concepts of the one-particle physics to the description by concepts of the many-particle physics shall be studied. The rubidium atoms are trapped in a magneto-optical trap (MOT) and from there reloaded into a magnetic trap. In this the rubidium atoms are stored in the state vertical stroke f=2,m f =2 right angle of the electronic ground state and evaporatively cooled by means of microwave-induced transitions into the state vertical stroke f=1,m f =1] (microwave cooling). The cesium atoms are also trppaed in a MOT and into the same magnetic trap reloaded, in which they are stored in the state vertical stroke f=4,m f =4 right angle of the electronic ground state together with rubidium. Because of the different hyperfine splitting only rubidium is evaporatively cooled, while cesium is cooled jointly sympathetically - i.e. by theramal contact via elastic collisions with rubidium atoms. The first two chapters contain a description of interatomic interactions in ultracold gases as well as a short summary of theoretical concepts in the description of Bose-Einstein condensates. The chapters 3 and 4 contain a short presentation of the methods applied in the experiment for the production of ultracold gases as well as the experimental arrangement; especially in the framework of this thesis a new coil system has been designed, which offers in view of future experiments additionally optical access for an optical trap. Additionally the fourth chapter contains an extensive description of the experimental cycle, which is applied in order to store rubidium and cesium atoms together into the magnetic trap. The last chapter 9. Measurement of product rotational alignment in associative-ionization collisions between polarized Na(3p) atoms International Nuclear Information System (INIS) Wang, M.; de Vries, M.S.; Weiner, J. 1986-01-01 We have studied the effect of reactant Na(3p) polarization on the rotational angular momentum alignment of product Na 2 + ions arising from associative-ionization (AI) collisions. Our results show that sensitivity of the AI rate constant to initial atomic polarization persists even when all hyperfine states are populated with broadband (3 cm -1 ) pulsed laser excitation of Na( 2 P/sub 3/2/) and that the spatial distribution of product rotational angular momentum vectors is anisotropic. We discuss a qualitative description of the collision process consistent with our measurements which indicates that sigma-orbital symmetry is preferred to π-orbital symmetry as the colliding partners approach 10. Hanford/Tomsk reciprocal site visit: Plutonium agreement compliance talks International Nuclear Information System (INIS) Libby, R.A.; Sorenson, R.; Six, D.; Schiegel, S.C. 1994-11-01 The objective of the visit to Hanford Site was to: demonstrate equipment, technology, and methods for calculating Pu production, measuring integrated reactor power, and storing and safeguarding PuO 2 ; demonstrate the shutdown of Hanford production reactors; and foster openness and transparency of Hanford operations. The first day's visit was an introduction to Hanford and a review of the history of the reactors. The second day consisted of discussions on the production reactors, reprocessing operations, and PuO 2 storage. The group divided on the third day to tour facilities. Group A toured the N reactor, K-West reactor, K-West Basins, B reactor, and participated in a demonstration and discussion of reactor modeling computer codes. Group B toured the Hanford Pu Storage Facility, 200-East Area, N-cell (oxide loadout station), the Automated Storage Facility, and the Nondestructive Assay Measurement System. Group discussions were held during the last day of the visit, which included scheduling of a US visit to Russia 11. Overview of the spent nuclear fuel project at Hanford International Nuclear Information System (INIS) Daily, J.L. 1995-02-01 The Spent Nuclear Fuel Project's mission at Hanford is to open-quotes Provide safe, economic and environmentally sound management of Hanford spent nuclear fuel in a manner which stages it to final disposition.close quotes The inventory of spent nuclear fuel (SNF) at the Hanford Site covers a wide variety of fuel types (production reactor to space reactor) in many facilities (reactor fuel basins to hot cells) at locations all over the Site. The 2,129 metric tons of Hanford SNF represents about 80% of the total US Department of Energy (DOE) inventory. About 98.5% of the Hanford SNF is 2,100 metric tons of metallic uranium production reactor fuel currently stored in the 1950s vintage K Basins in the 100 Area. This fuel has been slowly corroding, generating sludge and contaminating the basin water. This condition, coupled with aging facilities with seismic vulnerabilities, has been identified by several groups, including stakeholders, as being one of the most urgent safety and environmental concerns at the Hanford Site. As a direct result of these concerns, the Spent Nuclear Fuel Project was recently formed to address spent fuel issues at Hanford. The Project has developed the K Basins Path Forward to remove fuel from the basins and place it in dry interim storage. Alternatives that addressed the requirements were developed and analyzed. The result is a two-phased approach allowing the early removal of fuel from the K Basins followed by its stabilization and interim storage consistent with the national program 12. Molecular Ions in Ion Upflows and their Effects on Hot Atomic Oxygen Production Science.gov (United States) Foss, V.; Yau, A. W.; Shizgal, B. 2017-12-01 We present new direct ion composition observations of molecular ions in auroral ion upflows from the CASSIOPE Enhanced Polar Outflow Probe (e-POP). These observed molecular ions are N2+, NO+, and possibly O2+, and are found to occur at all e-POP altitudes starting at about 400 km, during auroral substorms and the different phases of magnetic storms, sometimes with upflow velocities exceeding a few hundred meters per second and abundances of 5-10%. The dissociative recombination of both O2+ and NO+ was previously proposed as an important source of hot oxygen atoms in the topside thermosphere [Hickey et al., 1995]. We investigate the possible effect of the observed molecular ions on the production of hot oxygen atoms in the storm and substorm-time auroral thermosphere. We present numerical solutions of the Boltzmann equation for the steady-state oxygen energy distribution function, taking into account both the production of the hot atoms and their subsequent collisional relaxation. Our result suggests the formation of a hot oxygen population with a characteristic temperature on the order of 0.3 eV and constituting 1-5% of the oxygen density near the exobase. We discuss the implication of this result in the context of magnetosphere-ionosphere-thermosphere coupling. 13. Deactivation completed at historic Hanford Fuels Laboratory Energy Technology Data Exchange (ETDEWEB) Gerber, M.S. 1994-03-01 This report discusses deactivation work which was completed as of March 31, 1994 at the 308 Fuels Development Laboratory (FDL) at the Hanford Site near Richland, Washington. The decision to deactivate the structure, formerly known as the Plutonium Fabrication Pilot Plant (PFPP), was driven by a 1980s Department of Energy (DOE) decision that plutonium fuels should not be fabricated in areas near the Sites boundaries, as well as by changing facility structural requirements. Inventory transfer has been followed by the cleanout and stabilization of plutonium oxide (PuO{sub 2}) and enriched uranium oxide (UO{sub 2}) residues and powders in the facilitys equipment and duct work. The Hanford Site, located in southeastern Washington state, was one of Americas primary arsenals of nuclear defense production for nearly 50 years beginning in World War II. Approximately 53 metric tons of weapons grade plutonium, over half of the national supply and about one quarter of the worlds supply, were produced at Hanford between 1944 and 1989. Today, many Site buildings are undergoing deactivation, a precursor phase to decontamination and decommissioning (D&D). The primary difference between the two activities is that equipment and structural items are not removed or torn down in deactivation. However, utilities are disconnected, and special nuclear materials (SNM) as well as hazardous and pyrophoric substances are removed from structures undergoing this process. 14. Deactivation completed at historic Hanford Fuels Laboratory International Nuclear Information System (INIS) Gerber, M.S. 1994-03-01 This report discusses deactivation work which was completed as of March 31, 1994 at the 308 Fuels Development Laboratory (FDL) at the Hanford Site near Richland, Washington. The decision to deactivate the structure, formerly known as the Plutonium Fabrication Pilot Plant (PFPP), was driven by a 1980s Department of Energy (DOE) decision that plutonium fuels should not be fabricated in areas near the Site's boundaries, as well as by changing facility structural requirements. Inventory transfer has been followed by the cleanout and stabilization of plutonium oxide (PuO 2 ) and enriched uranium oxide (UO 2 ) residues and powders in the facility's equipment and duct work. The Hanford Site, located in southeastern Washington state, was one of America's primary arsenals of nuclear defense production for nearly 50 years beginning in World War II. Approximately 53 metric tons of weapons grade plutonium, over half of the national supply and about one quarter of the world's supply, were produced at Hanford between 1944 and 1989. Today, many Site buildings are undergoing deactivation, a precursor phase to decontamination and decommissioning (D ampersand D). The primary difference between the two activities is that equipment and structural items are not removed or torn down in deactivation. However, utilities are disconnected, and special nuclear materials (SNM) as well as hazardous and pyrophoric substances are removed from structures undergoing this process 15. Hanford inventory program user's manual International Nuclear Information System (INIS) Hinkelman, K.C. 1994-01-01 Provides users with instructions and information about accessing and operating the Hanford Inventory Program (HIP) system. The Hanford Inventory Program is an integrated control system that provides a single source for the management and control of equipment, parts, and material warehoused by Westinghouse Hanford Company in various site-wide locations. The inventory is comprised of spare parts and equipment, shop stock, special tools, essential materials, and convenience storage items. The HIP replaced the following systems; ACA, ASP, PICS, FSP, WSR, STP, and RBO. In addition, HIP manages the catalog maintenance function for the General Supplies inventory stocked in the 1164 building and managed by WIMS 16. Hanford Waste Vitrification Plant technical manual Energy Technology Data Exchange (ETDEWEB) Larson, D.E. [ed.; Watrous, R.A.; Kruger, O.L. [and others 1996-03-01 A key element of the Hanford waste management strategy is the construction of a new facility, the Hanford Waste Vitrification Plant (HWVP), to vitrify existing and future liquid high-level waste produced by defense activities at the Hanford Site. The HWVP mission is to vitrify pretreated waste in borosilicate glass, cast the glass into stainless steel canisters, and store the canisters at the Hanford Site until they are shipped to a federal geological repository. The HWVP Technical Manual (Manual) documents the technical bases of the current HWVP process and provides a physical description of the related equipment and the plant. The immediate purpose of the document is to provide the technical bases for preparation of project baseline documents that will be used to direct the Title 1 and Title 2 design by the A/E, Fluor. The content of the Manual is organized in the following manner. Chapter 1.0 contains the background and context within which the HWVP was designed. Chapter 2.0 describes the site, plant, equipment and supporting services and provides the context for application of the process information in the Manual. Chapter 3.0 provides plant feed and product requirements, which are primary process bases for plant operation. Chapter 4.0 summarizes the technology for each plant process. Chapter 5.0 describes the engineering principles for designing major types of HWVP equipment. Chapter 6.0 describes the general safety aspects of the plant and process to assist in safe and prudent facility operation. Chapter 7.0 includes a description of the waste form qualification program and data. Chapter 8.0 indicates the current status of quality assurance requirements for the Manual. The Appendices provide data that are too extensive to be placed in the main text, such as extensive tables and sets of figures. The Manual is a revision of the 1987 version. 17. Hanford Waste Vitrification Plant technical manual International Nuclear Information System (INIS) Larson, D.E.; Watrous, R.A.; Kruger, O.L. 1996-03-01 A key element of the Hanford waste management strategy is the construction of a new facility, the Hanford Waste Vitrification Plant (HWVP), to vitrify existing and future liquid high-level waste produced by defense activities at the Hanford Site. The HWVP mission is to vitrify pretreated waste in borosilicate glass, cast the glass into stainless steel canisters, and store the canisters at the Hanford Site until they are shipped to a federal geological repository. The HWVP Technical Manual (Manual) documents the technical bases of the current HWVP process and provides a physical description of the related equipment and the plant. The immediate purpose of the document is to provide the technical bases for preparation of project baseline documents that will be used to direct the Title 1 and Title 2 design by the A/E, Fluor. The content of the Manual is organized in the following manner. Chapter 1.0 contains the background and context within which the HWVP was designed. Chapter 2.0 describes the site, plant, equipment and supporting services and provides the context for application of the process information in the Manual. Chapter 3.0 provides plant feed and product requirements, which are primary process bases for plant operation. Chapter 4.0 summarizes the technology for each plant process. Chapter 5.0 describes the engineering principles for designing major types of HWVP equipment. Chapter 6.0 describes the general safety aspects of the plant and process to assist in safe and prudent facility operation. Chapter 7.0 includes a description of the waste form qualification program and data. Chapter 8.0 indicates the current status of quality assurance requirements for the Manual. The Appendices provide data that are too extensive to be placed in the main text, such as extensive tables and sets of figures. The Manual is a revision of the 1987 version 18. Hanford Site air operating permit application Energy Technology Data Exchange (ETDEWEB) NONE 1995-05-01 The Clean Air Act Amendments of 1990, which amended the Federal Clean Air Act of 1977, required that the US Environmental Protection Agency develop a national Air Operating Permit Program, which in turn would require each state to develop an Air Operating Permit Program to identify all sources of regulated pollutants. Regulated pollutants include criteria pollutants (oxides of nitrogen, sulfur oxides, total suspended particulates, carbon monoxide, particulate matter greater than 10 micron, lead) plus 189 other Hazardous Air Pollutants. The Hanford Site, owned by the US Government and operated by the US Department of Energy, Richland Operations Office, is located in southcentral Washington State and covers 560 square miles of semi-arid shrub and grasslands located just north of the confluence of the Snake and Yakima Rivers with the Columbia River. This land, with restricted public access, provides a buffer for the smaller areas historically used for the production of nuclear materials, waste storage, and waste disposal. About 6 percent of the land area has been disturbed and is actively used. The Hanford Site Air Operating Permit Application consists of more than 1,100 sources and in excess of 300 emission points. Before January 1995, the maintenance and operations contractor and the environmental restoration contractor for the US Department of Energy completed an air emission inventory on the Hanford Site. The inventory has been entered into a database so that the sources and emission points can be tracked and updated information readily can be retrieved. The Hanford Site Air Operating Permit Application contains information current as of April 19, 1995. 19. Hanford Site air operating permit application International Nuclear Information System (INIS) 1995-05-01 The Clean Air Act Amendments of 1990, which amended the Federal Clean Air Act of 1977, required that the US Environmental Protection Agency develop a national Air Operating Permit Program, which in turn would require each state to develop an Air Operating Permit Program to identify all sources of ''regulated'' pollutants. Regulated pollutants include ''criteria'' pollutants (oxides of nitrogen, sulfur oxides, total suspended particulates, carbon monoxide, particulate matter greater than 10 micron, lead) plus 189 other ''Hazardous'' Air Pollutants. The Hanford Site, owned by the US Government and operated by the US Department of Energy, Richland Operations Office, is located in southcentral Washington State and covers 560 square miles of semi-arid shrub and grasslands located just north of the confluence of the Snake and Yakima Rivers with the Columbia River. This land, with restricted public access, provides a buffer for the smaller areas historically used for the production of nuclear materials, waste storage, and waste disposal. About 6 percent of the land area has been disturbed and is actively used. The Hanford Site Air Operating Permit Application consists of more than 1,100 sources and in excess of 300 emission points. Before January 1995, the maintenance and operations contractor and the environmental restoration contractor for the US Department of Energy completed an air emission inventory on the Hanford Site. The inventory has been entered into a database so that the sources and emission points can be tracked and updated information readily can be retrieved. The Hanford Site Air Operating Permit Application contains information current as of April 19, 1995 20. Bottom-up production of meta-atoms for optical magnetism in visible and NIR light Science.gov (United States) Barois, Philippe; Ponsinet, Virginie; Baron, Alexandre; Richetti, Philippe 2018-02-01 Many unusual optical properties of metamaterials arise from the magnetic response of engineered structures of sub-wavelength size (meta-atoms) exposed to light. The top-down approach whereby engineered nanostructure of well-defined morphology are engraved on a surface proved to be successful for the generation of strong optical magnetism. It faces however the limitations of high cost and small active area in visible light where nanometre resolution is needed. The bottom-up approach whereby the fabrication metamaterials of large volume or large area results from the combination of nanochemitry and self-assembly techniques may constitute a cost-effective alternative. This approach nevertheless requires the large-scale production of functional building-blocks (meta-atoms) bearing a strong magnetic optical response. We propose in this paper a few tracks that lead to the large scale synthesis of magnetic metamaterials operating in visible or near IR light. 1. ITP Hanford Type 40 pin electrical connector failure analysis International Nuclear Information System (INIS) Imrich, K.J. 1993-01-01 Corrosion products observed on the ITP Hanford Type 40 pin electrical connectors would be expected to adversely affect the power and control signals supplied to process equipment in the filter cell by the connectors. Corrosion products were consistent with those found on similar pins in DWPF. The recommendations based on the findings in this investigation are as follows: (1) Replace male and female rhodium plated pins with gold plated pins. (2) Replace the galvanized carbon steel spring on the male connector with a stainless steel spring. (3) Install protective caps over Hanford connectors when jumpers are removed 2. The method of atomic absorption spectrophotometry for determining of cadmium in fruit and vegetable products International Nuclear Information System (INIS) Brzozowska, B. 1977-01-01 The method of atomic absorption with the technique of standard addition was used for determination of cadmium in the following tinned products: green peas, cut bean pods, sorel, stewed black currants, greengage plums, orange juice. The products were dry mineralized. Each mineralizate was divided into three portions, known amounts of cadmium were added to two portions and all portions were supplemented to a defined volume. Determinations were performed using a Pye Unicam SP 90 A spectrophotometer and they served as a base for plotting a curve in the system: absorbance - concentration of added metal. The curve was extrapolated to zero absorbance for reading directly the content of the metal in the product. This content was in the range from 10 to 80 μg/kg at variance coefficient 5-15% and the recovery was 80-130%. (author) 3. Hanford Site Environmental Report 1993 Energy Technology Data Exchange (ETDEWEB) Dirkes, R.L.; Hanf, R.W.; Woodruff, R.K. [eds. 1994-06-01 The Hanford Site Environmental Report is prepared annually to summarize environmental data and information, describe environmental management performance, and demonstrate the status of compliance with environmental regulations. The report also highlights major environmental programs and efforts. The report is written to meet reporting requirements and Guidelines of the U.S. Department of Energy (DOE) an to meet the needs of the public. This summary has been written with a minimum of technical terminology. Individual sections of the report are designed to (a) describe the Hanford Site and its mission, (b) summarize the status in 1993 of compliance with environmental regulations, (c) describe the environmental programs at the Hanford Site, (d) discuss estimated radionuclide exposure to the public from 1993 Hanford activities, (e) present information on effluent monitoring and environmental surveillance, including ground-water protection and monitoring, (f) discuss activities to ensure quality. More detailed information can be found in the body of the report, the appendixes, and the cited references. 4. Hanford Site Environmental Report 1993 International Nuclear Information System (INIS) Dirkes, R.L.; Hanf, R.W.; Woodruff, R.K. 1994-06-01 The Hanford Site Environmental Report is prepared annually to summarize environmental data and information, describe environmental management performance, and demonstrate the status of compliance with environmental regulations. The report also highlights major environmental programs and efforts. The report is written to meet reporting requirements and Guidelines of the U.S. Department of Energy (DOE) an to meet the needs of the public. This summary has been written with a minimum of technical terminology. Individual sections of the report are designed to (a) describe the Hanford Site and its mission, (b) summarize the status in 1993 of compliance with environmental regulations, (c) describe the environmental programs at the Hanford Site, (d) discuss estimated radionuclide exposure to the public from 1993 Hanford activities, (e) present information on effluent monitoring and environmental surveillance, including ground-water protection and monitoring, (f) discuss activities to ensure quality. More detailed information can be found in the body of the report, the appendixes, and the cited references 5. Hanford Site Environmental Report 1999 International Nuclear Information System (INIS) Poston, TM; Hanf, RW; Dirkes, RL 2000-01-01 This Hanford Site environmental report is prepared annually to summarize environmental data and information, to describe environmental management performance, to demonstrate the status of compliance with environmental regulations, and to highlight major environmental programs and efforts. The report is written to meet requirements and guidelines of the U.S. Department of Energy (DOE) and to meet the needs of the public. This summary has been written with a minimum of technical terminology. Individual sections of the report are designed to: (1) describe the Hanford Site and its mission; (2) summarize the status of compliance with environmental regulations; (3) describe the environmental programs at the Hanford Site; (4) discuss the estimated radionuclide exposure to the public from 1999 Hanford Site activities; (5) present the effluent monitoring, environmental surveillance, groundwater protection and monitoring information; and (6) discuss the activities to ensure quality 6. Hanford Facility RCRA permit handbook Energy Technology Data Exchange (ETDEWEB) NONE 1996-03-01 Purpose of this Hanford Facility (HF) RCRA Permit Handbook is to provide, in one document, information to be used for clarification of permit conditions and guidance for implementing the HF RCRA Permit. 7. Hanford Surplus Facilities Program plan International Nuclear Information System (INIS) Hughes, M.C.; Wahlen, R.K.; Winship, R.A. 1989-09-01 The Hanford Surplus Facilities Program is responsible for the safe and cost-effective surveillance, maintenance, and decommissioning of surplus facilities at the Hanford Site. The management of these facilities requires a surveillance and maintenance program to keep them in a safe condition and development of a plan for ultimate disposition. Criteria used to evaluate each factor relative to decommissioning are based on the guidelines presented by the US Department of Energy-Richland Operations Office, Defense Facilities Decommissioning Program Office, and are consistent with the Westinghouse Hanford Company commitment to decommission the Hanford Site retired facilities in the safest and most cost-effective way achievable. This document outlines the plan for managing these facilities to the end of disposition 8. Mortality studies of Hanford workers International Nuclear Information System (INIS) Gilbert, E.S. 1986-03-01 The relationships of cancer mortality with radiation exposure as influenced by age, sex, follow-up time length of employment, and job category are discussed in relation to workers at the Hanford facilities 9. Hanford Waste Management Plan, 1987 International Nuclear Information System (INIS) 1987-01-01 The purpose of the Hanford Waste Management Plan (HWMP) is to provide an integrated plan for the safe storage, interim management, and disposal of existing waste sites and current and future waste streams at the Hanford Site. The emphasis of this plan is, however, on the disposal of Hanford Site waste. The plans presented in the HWMP are consistent with the preferred alternative which is based on consideration of comments received from the public and agencies on the draft Hanford Defense Waste Environmental Impact Statement (HDW-EIS). Low-level waste was not included in the draft HDW-EIS whereas it is included in this plan. The preferred alternative includes disposal of double-shell tank waste, retrievably stored and newly generated TRU waste, one pre-1970 TRU solid waste site near the Columbia River and encapsulated cesium and strontium waste 10. Hanford site waste tank characterization International Nuclear Information System (INIS) De Lorenzo, D.S.; Simpson, B.C. 1994-08-01 This paper describes the on-going work in the characterization of the Hanford-Site high-level waste tanks. The waste in these tanks was produced as part of the nuclear weapons materials processing mission that occupied the Hanford Site for the first 40 years of its existence. Detailed and defensible characterization of the tank wastes is required to guide retrieval, pretreatment, and disposal technology development, to address waste stability and reactivity concerns, and to satisfy the compliance criteria for the various regulatory agencies overseeing activities at the Hanford Site. The resulting Tank Characterization Reports fulfill these needs, as well as satisfy the tank waste characterization milestones in the Hanford Federal Facility Agreement and Consent Order 11. Hanford Site 1998 Environmental Report Energy Technology Data Exchange (ETDEWEB) RL Dirkes; RW Hanf; TM Poston 1999-09-21 This Hanford Site environmental report is prepared annually to summarize environmental data and information, to describe environmental management performance, to demonstrate the status of compliance with environmental regulations, and to highlight major environmental programs and efforts. The report is written to meet requirements and guidelines of the U.S. Department of Energy (DOE) and to meet the needs of the public. This summary has been written with a minimum of technical terminology. Individual sections of the report are designed to: describe the Hanford Site and its mission; summarize the status of compliance with environmental regulations; describe the environmental programs at the Hanford Site; discuss the estimated radionuclide exposure to the public from 1998 Hanford Site activities; present the effluent monitoring, environmental surveillance, and groundwater protection and monitoring information; and discuss the activities to ensure quality. 12. Hanford Site Environmental Report 1999 Energy Technology Data Exchange (ETDEWEB) TM Poston; RW Hanf; RL Dirkes 2000-09-28 This Hanford Site environmental report is prepared annually to summarize environmental data and information, to describe environmental management performance, to demonstrate the status of compliance with environmental regulations, and to highlight major environmental programs and efforts. The report is written to meet requirements and guidelines of the U.S. Department of Energy (DOE) and to meet the needs of the public. This summary has been written with a minimum of technical terminology. Individual sections of the report are designed to: (1) describe the Hanford Site and its mission; (2) summarize the status of compliance with environmental regulations; (3) describe the environmental programs at the Hanford Site; (4) discuss the estimated radionuclide exposure to the public from 1999 Hanford Site activities; (5) present the effluent monitoring, environmental surveillance, groundwater protection and monitoring information; and (6) discuss the activities to ensure quality. 13. UPDATE HANFORD SITE D and D PROGRAMS ACCELERATE EXPAND International Nuclear Information System (INIS) GERBER, M.S. 2004-01-01 A large, new decontamination and decommissioning organization targeted toward rapid, focused work on aging and highly contaminated structures was formed at the DOE's Hanford Site in southeast Washington state in autumn 2003. Managed by prime contractor Fluor Hanford, the new organization has made significant progress during its first six months. Under the direction of Mike Lackey, who recently joined Fluor from the Portland General Electric Trojan Plant, the Fluor Hanford DandD organization is tackling the Plutonium Finishing Plant (PFP) complex and the Fast Flux Test Facility (FFTF), and is nearly finished demolishing the 233-S Plutonium Concentration Facility. In addition, the DandD organization is progressing through the development and public comment phases of its required environmental permitting, planning work and procurement services to DandD three other Hanford facilities: 224-T and 224-B Plutonium Concentration Facilities, and the U Plant radiochemical processing facility. It is also planning and beginning to DandD the spent fuel handling areas of the Site's 100-K Reactor Area. The 586-square mile Hanford Site, the oldest plutonium production center in the world, served as the ''workhorse'' of the American nuclear defense arsenal from 1944 through 1989. Hanford produced the special nuclear material for the plutonium cores of the Trinity (test) and Nagasaki explosions, and then went on to produce more than half of the weapons plutonium ever manufactured by the United States, and about one-fourth of that manufactured worldwide. As a result, Hanford, the top-secret ''Paul Bunyan'' in the desert, is one of the most contaminated areas in the world. Its cleanup agreement with state and federal regulators, known as the ''Tri-Party Agreement,'' celebrates its 15th anniversary this spring, at a time when operations dealing with unstable plutonium leftovers, corroded spent fuel, and liquids wastes in single-shelled tanks conclude. As these crucial jobs are coming to 14. The hydrogen atom-deuterium molecule reaction: Experimental determination of product quantum state distributions International Nuclear Information System (INIS) Rinnen, K. 1989-01-01 The H + H 2 atom exchange reaction (and its isotopic analogs) is the simplest neutral bimolecular chemical reaction because of the small number of electrons in the system and the lightness of the nuclei. The H 3 potential energy surface (PES) is the most accurately known reactive surface (LSTH surface); there have been both quasiclassical trajectory (QCT) and quantal calculations performed on it. This is one of the few systems for which theory is ahead of experiment, and many theoretical predictions await experimental comparison. The H + D 2 → HD + D reaction is studied using thermal D 2 (∼298 K) and translationally hot hydrogen atoms. Photolysis of HI at 266 nm generates H atoms with center-of-mass collision energies of 1.3 and 0.55 eV, both of which are above the classical reaction barrier of 0.42 eV. The rovibrational population distribution of the molecular product is measured by (2+1) resonance-enhanced multiphoton ionization (REMPI). A major effort has been directed toward calibrating the (2+1) REMPI detection procedure, to determine quantitatively the relationship between ion signals and relative quantum state populations for HD. An effusive, high-temperature nozzle has been constructed to populate thermally the high rovibrational levels observed in the reaction. The results are compared to theoretical calculations of the E,F 1 Σ g + - X 1 Σ g + two-photon transition moments. For the H + D 2 reaction, the populations of all energetically accessible HD product levels are measured. Specifically, the following levels are observed: HD(v = 0, J = 0-15), HD(v = 1, J = 0-12), and HD(v = 2, J = 0-8). Of the available energy, 73% is partitioned into product translation, 18% into HD rotation, and 9% into HD vibration 15. Hanford Environmental Dose Reconstruction Project International Nuclear Information System (INIS) Cannon, S.D.; Finch, S.M. 1992-10-01 The objective of the Hanford Environmental Dose Reconstruction (HEDR) Project is to estimate the radiation doses that individuals and populations could have received from nuclear operations at Hanford since 1944. The independent Technical Steering Panel (TSP) provides technical direction. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed from release to impact on humans (dose estimates):Source Terms, Environmental Transport, Environmental Monitoring Data, Demography, Food Consumption, and Agriculture, and Environmental Pathways and Dose Estimates 16. HANFORD WASTE MINEROLOGY REFERENCE REPORT International Nuclear Information System (INIS) Disselkamp, R.S. 2010-01-01 This report lists the observed mineral phases present in the Hanford tanks. This task was accomplished by performing a review of numerous reports using experimental techniques including, but not limited to: x-ray diffraction, polarized light microscopy, scanning electron microscopy, transmission electron microscopy, energy dispersive spectroscopy, electron energy loss spectroscopy, and particle size distribution analyses. This report contains tables that can be used as a quick reference to identify the crystal phases present observed in Hanford waste. 17. HANFORD WASTE MINERALOGY REFERENCE REPORT Energy Technology Data Exchange (ETDEWEB) DISSELKAMP RS 2010-06-29 This report lists the observed mineral phases present in the Hanford tanks. This task was accomplished by performing a review of numerous reports that used experimental techniques including, but not limited to: x-ray diffraction, polarized light microscopy, scanning electron microscopy, transmission electron microscopy, energy dispersive spectroscopy, electron energy loss spectroscopy, and particle size distribution analyses. This report contains tables that can be used as a quick reference to identify the crystal phases observed in Hanford waste. 18. HANFORD WASTE MINEROLOGY REFERENCE REPORT Energy Technology Data Exchange (ETDEWEB) DISSELKAMP RS 2010-06-18 This report lists the observed mineral phase phases present in the Hanford tanks. This task was accomplished by performing a review of numerous reports using experimental techniques including, but not limited to: x-ray diffraction, polarized light microscopy, scanning electron microscopy, transmission electron microscopy, energy dispersive spectroscopy, electron energy loss spectroscopy, and particle size distribution analyses. This report contains tables that can be used as a quick reference to identify the crystal phases present observed in Hanford waste. 19. Hanford Waste Mineralogy Reference Report International Nuclear Information System (INIS) Disselkamp, R.S. 2010-01-01 This report lists the observed mineral phases present in the Hanford tanks. This task was accomplished by performing a review of numerous reports that used experimental techniques including, but not limited to: x-ray diffraction, polarized light microscopy, scanning electron microscopy, transmission electron microscopy, energy dispersive spectroscopy, electron energy loss spectroscopy, and particle size distribution analyses. This report contains tables that can be used as a quick reference to identify the crystal phases observed in Hanford waste. 20. Hanford internal dosimetry program manual International Nuclear Information System (INIS) Carbaugh, E.H.; Sula, M.J.; Bihl, D.E.; Aldridge, T.L. 1989-10-01 This document describes the Hanford Internal Dosimetry program. Program Services include administrating the bioassay monitoring program, evaluating and documenting assessments of internal exposure and dose, ensuring that analytical laboratories conform to requirements, selecting and applying appropriate models and procedures for evaluating internal radionuclide deposition and the resulting dose, and technically guiding and supporting Hanford contractors in matters regarding internal dosimetry. 13 refs., 16 figs., 42 tabs 1. Production and quality control of concrete for the Rajasthan Atomic Power Station - [Part 2 International Nuclear Information System (INIS) Singh Roy, P.K.; Sukhtankar, K.D.; Prasad, K. 1975-01-01 The following aspects of the production and quality control of concrete and concrete materials used in the construction of twin-reactor Rajasthan Atomic Power Station are discussed : (1) relationship between strength of cubes and cylinders made of concrete used for the prestressed dome (2) temperature control during pouring of concrete (3) thermal conductivity of heavy concrete (4) various types of grouting procedures used for different structures forming part of reactors (5) quality control of normal and heavy concrete and (6) leakage through form ties. Typical concrete mixes used for grouts are also given. (M.G.B.) 2. Dynamics of atoms-ions transformation processes in the radioactive ion production systems ISOL International Nuclear Information System (INIS) Jardin, Pascal 2013-01-01 The aims of this work were 1)to study the effect of diffusion, effusion and ionization processes in the atom-ion transformation, 2)to better understand the temporal behaviour of ISOL devices and to apply it to the developments of the ISOL production systems. These aims were partially reached: the results obtained with 'ECS ECR' of SPIRAL 1 and SPIRAL 2 and their confrontation have allowed to analytically described their temporal behaviour and to reveal under which conditions it is possible to consider the processes of diffusion, effusion and ionization as separable processes and consequently to consider them as consecutive. (O.M.) [fr 3. Mechanisms of defect production and atomic mixing in high energy displacement cascades: A molecular dynamics study International Nuclear Information System (INIS) Diaz de la Rubia, T.; Guinan, M.W. 1991-01-01 We have performed molecular dynamics computer simulation studies of displacement cascades in Cu at low temperature. For 25 keV recoils we observe the splitting of a cascade into subcascades and show that cascades in Cu may lead to the formation of vacancy and interstitial dislocation loops. We discuss a new mechanism of defect production based on the observation of interstitial prismatic dislocation loop punching from cascades at 10 K. We also show that below the subcascade threshold, atomic mixing in the cascade is recoil-energy dependent and obtain a mixing efficiency that scales as the square root of the primary recoil energy. 44 refs., 12 figs 4. Efficiencies for production of atomic nitrogen and oxygen by relativistic proton impact in air Science.gov (United States) Porter, H. S.; Jackman, C. H.; Green, A. E. S. 1976-01-01 Relativistic electron and proton impact cross sections are obtained and represented by analytic forms which span the energy range from threshold to 1 GeV. For ionization processes, the Massey-Mohr continuum generalized oscillator strength surface is parameterized. Parameters are determined by simultaneous fitting to (1) empirical data, (2) the Bethe sum rule, and (3) doubly differential cross sections for ionization. Branching ratios for dissociation and predissociation from important states of N2 and O2 are determined. The efficiency for the production of atomic nitrogen and oxygen by protons with kinetic energy less than 1 GeV is determined using these branching ratio and cross section assignments. 5. Hanford Central Waste Complex: Radioactive mixed waste storage facility dangerous waste permit application International Nuclear Information System (INIS) 1991-10-01 The Hanford Site is owned by the US Government and operated by the US Department of Energy Field Office, Richland. The Hanford Site manages and produces dangerous waste and mixed waste (containing both radioactive and dangerous components). The dangerous waste is regulated in accordance with the Resource Conversation and Recovery Act of 1976 and the State of Washington Hazardous Waste Management Act of 1976. The radioactive component of mixed waste is interpreted by the US Department of Energy to be regulated under the Atomic Energy Act of 1954; the nonradioactive dangerous component of mixed waste is interpreted to be regulated under the Resource Conservation and Recovery Act of 1976 and Washington Administrative Code 173--303. Westinghouse Hanford Company is a major contractor to the US Department of Energy Field Office, Richland and serves as co-operator of the Hanford Central Waste Complex. The Hanford Central Waste Complex is an existing and planned series of treatment, storage, and/or disposal units that will centralize the management of solid waste operations at a single location on the Hanford facility. The Hanford Central Waste Complex units include the Radioactive Mixed Waste Storage Facility, the unit addressed by this permit application, and the Waste Receiving and Processing Facility. The Waste Receiving and Processing Facility is covered in a separate permit application submittal 6. The projects for heavy water production of the Argentine National Atomic Energy Commission International Nuclear Information System (INIS) Garcia Bourg, J.M.; Garcia, E.E. 1982-01-01 The bases and scope of the projects for heavy water production that are being currently developed by the Argentine National Atomic Energy Commission (CNEA) are described. As an introduction, the following points are presented: a) the fundamentals of heavy water utilization in a nuclear reactor, with a mention of its properties and uses, b) a review of the physicochemical bases of the principal methods for heavy water production: chemical exchange (monothermal and bithermal processes), distillation and electrolysis, with tables summarizing the fundamental characteristics of the first two ones, and an evaluation of the different production methods from the viewpoint of their application in an industrial scale; and c) a synthetic information, in the form of tables, about the world's heavy water production. The subject of heavy water production in Argentina is treated in the principal section, describing the scope, location, main characteristics and chemical processes corresponding to the projects being developed by CNEA, which currently are the installation of an Industrial Plant in Arroyito (Province of Neuquen), purchased on a turnkey basis and using the NH 3 /H 2 isotopic exchange method; the installation of an Experimental Plant in Atucha (Province of Buenos Aires), for the development of the domestic technology of heavy-water production by the SH 2 /H 2 O isotopic exchange method, and the development of the engineering of an industrial plant (''Module 80''), based on the Experimental Plant's technology. (M.E.L.) [es 7. Influence of Glass Property Restrictions on Hanford HLW Glass Volume International Nuclear Information System (INIS) Kim, Dong-Sang; Vienna, John D. 2001-01-01 A systematic evaluation of Hanford High-Level Waste (HLW) loading in alkali-alumino-borosilicate glasses was performed. The waste feed compositions used were obtained from current tank waste composition estimates, Hanford's baseline retrieval sequence, and pretreatment processes. The waste feeds were sorted into groups of like composition by cluster analysis. Glass composition optimization was performed on each cluster to meet property and composition constraints while maximizing waste loading. Glass properties were estimated using property models developed for Hanford HLW glasses. The impacts of many constraints on the volume of HLW glass to be produced at Hanford were evaluated. The liquidus temperature, melting temperature, chromium concentration, formation of multiple phases on cooling, and product consistency test response requirements for the glass were varied one- or many-at-a-time and the resultant glass volume was calculated. This study shows clearly that the allowance of crystalline phases in the glass melter can significantly decrease the volume of HLW glass to be produced at Hanford. 8. Consequence ranking of radionuclides in Hanford tank waste International Nuclear Information System (INIS) Schmittroth, F.A.; De Lorenzo, T.H. 1995-09-01 Radionuclides in the Hanford tank waste are ranked relative to their consequences for the Low-Level Tank Waste program. The ranking identifies key radionuclides where further study is merited. In addition to potential consequences for intrude and drinking-water scenarios supporting low-level waste activities, a ranking based on shielding criteria is provided. The radionuclide production inventories are based on a new and independent ORIGEN2 calculation representing the operation of all Hanford single-pass reactors and the N Reactor 9. Pretreatment of Hanford purex plant first-cycle waste International Nuclear Information System (INIS) Gibson, M.W.; Gerboth, D.M.; Peters, B.B. 1987-01-01 A process has been developed to pretreat neutralized, first-cycle high-level waste from the fuels reprocessing facility (PUREX Plant) at the Hanford Site. The process separates solids from the supernate liquid, which contains soluble salts. The solids, including most of the fission products and transuranic elements, may then be vitrified for disposal, while the low-level supernate stream may be processed into a less expensive grout waste form. The process also includes ion exchange treatment of the separated supernate stream to remove radiocesium. A flow sheet based on these operations was completed to support a planned demonstration of the process in the Hanford site B Plant canyon facility 10. Westinghouse Hanford Company risk management strategy for retired surplus facilities International Nuclear Information System (INIS) Taylor, W.E.; Coles, G.A.; Shultz, M.V.; Egge, R.G. 1993-09-01 This paper describes an approach that facilitates management of personnel safety and environmental release risk from retired, surplus Westinghouse Hanford Company-managed facilities during the predemolition time frame. These facilities are located in the 100 and 200 Areas of the 1,450-km 2 (570-mi 2 ) Hanford Site in Richland, Washington. The production reactors are located in the 100 Area and the chemical separation facilities are located in the 200 Area. This paper also includes a description of the risk evaluation process, shows applicable results, and includes a description of comparison costs for different risk reduction options 11. Women and the Hanford Site Science.gov (United States) Gerber, Michele 2014-03-01 When we study the technical and scientific history of the Manhattan Project, women's history is sometimes left out. At Hanford, a Site whose past is rich with hard science and heavy construction, it is doubly easy to leave out women's history. After all, at the World War II Hanford Engineer Works - the earliest name for the Hanford Site - only nine percent of the employees were women. None of them were involved in construction, and only one woman was actually involved in the physics and operations of a major facility - Dr. Leona Woods Marshall. She was a physicist present at the startup of B-Reactor, the world's first full-scale nuclear reactor - now a National Historic Landmark. Because her presence was so unique, a special bathroom had to be built for her in B-Reactor. At World War II Hanford, only two women were listed among the nearly 200 members of the top supervisory staff of the prime contractor, and only one regularly attended the staff meetings of the Site commander, Colonel Franklin Matthias. Overall, women comprised less than one percent of the managerial and supervisory staff of the Hanford Engineer Works, most of them were in nursing or on the Recreation Office staff. Almost all of the professional women at Hanford were nurses, and most of the other women of the Hanford Engineer Works were secretaries, clerks, food-service workers, laboratory technicians, messengers, barracks workers, and other support service employees. The one World War II recruiting film made to attract women workers to the Site, that has survived in Site archives, is entitled A Day in the Life of a Typical Hanford Girl.'' These historical facts are not mentioned to criticize the past - for it is never wise to apply the standards of one era to another. The Hanford Engineer Works was a 1940s organization, and it functioned by the standards of the 1940s. Just as we cannot criticize the use of asbestos in constructing Hanford (although we may wish they hadn't used so much of it), we 12. Hanford Waste Vitrification Plant Dangerous Waste Permit Application International Nuclear Information System (INIS) 1991-10-01 The Hanford Facility currently stores mixed waste, resulting from various processing operations, in underground storage tanks. The Hanford Waste Vitrification Plant will be constructed and operated to process the high-activity fraction of mixed waste stored in these underground tanks. The Hanford Waste Vitrification Plant will solidify pretreated tank waste into a glass product that will be packaged for disposal in a national repository. This Vitrification Plant Dangerous Waste Permit Application, Revision 2, consists of both a Part A and a Part B permit application. An explanation of the Part A revisions, including Revision 4 submitted with this application, is provided at the beginning of the Part A section. The Part B consists of 15 chapters addressing the organization and content of the Part B Checklist prepared by the Washington State Department of Ecology (Ecology 1987) 13. Integrated environmental monitoring program at the Hanford Site International Nuclear Information System (INIS) Jaquish, R.E. 1990-08-01 The US Department of Energy's Hanford Site, north of Richland, Washington, has a mission of defense production, waste management, environmental restoration, advanced reactor design, and research development. Environmental programs at Hanford are conducted by Pacific Northwest Laboratory (PNL) and the Westinghouse Hanford Company (WHC). The WHC environmental programs include the compliance and surveillance activities associated with site operations and waste management. The PNL environmental programs address the site-wide and the of-site areas. They include the environmental surveillance and the associated support activities, such as dose calculations, and also the monitoring of environmental conditions to comply with federal and state environmental regulations on wildlife and cultural resources. These are called ''independent environmental programs'' in that they are conducted completely separate from site operations. The Environmental Surveillance and Oversight Program consists of the following projects: surface environmental surveillance; ground-water surveillance; wildlife resources monitoring; cultural resources; dose overview; radiation standards and calibrations; meteorological and climatological services; emergency preparedness 14. Strategic plan for Hanford site information management International Nuclear Information System (INIS) 1994-09-01 The Hanford Site missions are to clean up the Site, to provide scientific knowledge and technology to meet global needs, and to partner in the economic diversification of the region. To achieve these long-term missions and increase confidence in the quality of the Site's decision making process, a dramatically different information management culture is required, consistent with US Department of Energy (DOE) mandates on increased safety, productivity, and openness at its sites. This plan presents a vision and six strategies that will move the Site toward an information management culture that will support the Site missions and address the mandates of DOE 15. Advances in the production of isotopes and radiopharmaceuticals at the Atomic Energy Corporation of South Africa International Nuclear Information System (INIS) Louw, P.A.; De Villiers, W.Y.Z.; Jarvis, N.V. 1997-01-01 The Atomic Energy Corporation of South Africa Ltd (AEC) owns and operates the 20 MW research reactor, SAFARI-1. Utilisation of the reactor has in recent years changed from research and materials testing to the production of isotopes. The most important breakthrough achieved in recent years is the production of high quality fission 99Mo. This has been produced routinely since April 1993 and supplied to clients across the world. A capability for the reliable production of 1000 Ci of 99Mo per week (calibrated for six days after production) has been proven. The AEC has also established facilities to produce its own 99mTc generators together with a most of radiopharmaceutical kits for diagnostic nuclear medicine purposes. The production of 153 Sm and 131 I (tellurium oxide route) has been operational for many years. Applications include therapeutic radiopharmaceuticals such as 153 Sm-EDTMP for bone cancer pain palliation, 13' I-Lipiodol for liver cancer and 131 I capsules for thyroid treatment. Facilities for the production of other isotopes such as 131 I (from fission), 32 P and 35 S are in various stages of completion. Extensive analytical methods and equipment have been developed and are routinely used to certify the quality of exported isotopes. Irradiation and encapsulation of 192 Ir is also performed routinely at the AEC. Modern facilities allow for the production of isotopes such as 131 Ba and 140 La on an ad hoc basis. Quality assurance procedures based on ISO9000 were developed for all aspects of the production of the various isotopes. Documentation, such as Drug Master Files, required by authorities in various countries has also been submitted and accepted 16. Hanford Site Transuranic (TRU) Waste Certification Plan International Nuclear Information System (INIS) GREAGER, T.M. 1999-01-01 The Hanford Site Transuranic Waste Certification Plan establishes the programmatic framework and criteria with in which the Hanford Site ensures that contract-handled TRU wastes can be certified as compliant with the WIPP WAC and TRUPACT-II SARP 17. Hanford Site Transuranic (TRU) Waste Certification Plan Energy Technology Data Exchange (ETDEWEB) GREAGER, T.M. 1999-09-09 The Hanford Site Transuranic Waste Certification Plan establishes the programmatic framework and criteria within which the Hanford Site ensures that contract-handled TRU wastes can be certified as compliant with the WIPP WAC and TRUPACT-II SARP. 18. Architecture synthesis basis for the Hanford Cleanup system: First issue International Nuclear Information System (INIS) Holmes, J.J. 1994-06-01 This document describes a set of candidate alternatives proposed to accomplish the Hanford Cleanup system functions defined in a previous work. Development of alternatives is part of a sequence of system engineering activities which lead to definition of all the products which, when completed, accomplish the cleanup mission. The alternative set is developed to functional level four or higher depending on need 19. Hanford Waste Vitrification Plant Quality Assurance Program description for high-level waste form development and qualification International Nuclear Information System (INIS) 1993-08-01 The Hanford Waste Vitrification Plant Project has been established to convert the high-level radioactive waste associated with nuclear defense production at the Hanford Site into a waste form suitable for disposal in a deep geologic repository. The Hanford Waste Vitrification Plant will mix processed radioactive waste with borosilicate material, then heat the mixture to its melting point (vitrification) to forin a glass-like substance that traps the radionuclides in the glass matrix upon cooling. The Hanford Waste Vitrification Plant Quality Assurance Program has been established to support the mission of the Hanford Waste Vitrification Plant. This Quality Assurance Program Description has been written to document the Hanford Waste Vitrification Plant Quality Assurance Program 20. Hanford Site peak gust wind speeds International Nuclear Information System (INIS) Ramsdell, J.V. 1998-01-01 Peak gust wind data collected at the Hanford Site since 1945 are analyzed to estimate maximum wind speeds for use in structural design. The results are compared with design wind speeds proposed for the Hanford Site. These comparisons indicate that design wind speeds contained in a January 1998 advisory changing DOE-STD-1020-94 are excessive for the Hanford Site and that the design wind speeds in effect prior to the changes are still appropriate for the Hanford Site 1. Natural product-like virtual libraries: recursive atom-based enumeration. Science.gov (United States) Yu, Melvin J 2011-03-28 A new molecular enumerator is described that allows chemically and architecturally diverse sets of natural product-like and drug-like structures to be generated from a core structure as simple as a single carbon atom or as complex as a polycyclic ring system. Integrated with a rudimentary machine-learning algorithm, the enumerator has the ability to assemble biased virtual libraries enriched in compounds predicted to meet target criteria. The ability to dynamically generate relatively small focused libraries in a recursive manner could reduce the computational time and infrastructure necessary to construct and manage extremely large static libraries. Depending on enumeration conditions, natural product-like structures can be produced with a wide range of heterocyclic and alicyclic ring assemblies. Because natural products represent a proven source of validated structures for identifying and designing new drug candidates, mimicking the structural and topological diversity found in nature with a dynamic set of virtual natural product-like compounds may facilitate the creation of new ideas for novel, biologically relevant lead structures in areas of uncharted chemical space. 2. Off-Gas Analysis During the Vitrification of Hanford Radioactive Waste Samples International Nuclear Information System (INIS) Ha, B.C.; Ferrara, D.M.; Crawford, C.L.; Choi, A.S.; Bibler, N.E. 1998-01-01 This paper describes the off-gas analysis of samples collected during the radioactive vitrification experiments. Production and characterization of the Hanford waste-containing LAW and HAW glasses are presented in related reports from this conference 3. Hanford Sitewide Groundwater Remediation Strategy International Nuclear Information System (INIS) Knepp, A.J.; Isaacs, J.D. 1997-09-01 This document fulfills the requirements of the Hanford Federal Facility Agreement and Consent Order, Milestone M-13-81, to develop a concise statement of strategy that describe show the Hanford Site groundwater remediation will be accomplished. The strategy addresses objectives and goals, prioritization of activities, and technical approaches for groundwater cleanup. The strategy establishes that the overall goal of groundwater remediation on the Hanford Site is to restore groundwater to its beneficial uses in terms of protecting human health and the environment, and its use as a natural resource. The Hanford Future Site Uses Working Group established two categories for groundwater commensurate with various proposed landuses: (1) restricted use or access to groundwater in the Central Plateau and in a buffer zone surrounding it and (2) unrestricted use or access to groundwater for all other areas. In recognition of the Hanford Future Site Uses Working Group and public values, the strategy establishes that the sitewide approach to groundwater cleanup is to remediate the major plumes found in the reactor areas that enter the Columbia River and to contain the spread and reduce the mass of the major plumes found in the Central Plateau 4. The Hanford Site focus, 1994 International Nuclear Information System (INIS) Peterson, J.M. 1994-03-01 This report describes what the Hanford Site will look like in the next two years. We offer thumbnail sketches of Hanford Site programs and the needs we are meeting through our efforts. We describe our goals, some recent accomplishments, the work we will do in fiscal year (FY) 1994, the major activities the FY 1995 budget request covers, and the economic picture in the next few years. The Hanford Site budget shows the type of work being planned. US Department of Energy (DOE) sites like the Hanford Site use documents called Activity Data Sheets to meet this need. These are building blocks that are included in the budget. Each Activity Data Sheet is a concise (usually 4 or 5 pages) summary of a piece of work funded by the DOE's Environmental Restoration and Waste Management budget. Each sheet describes a waste management or environmental restoration need over a 5-year period; related regulatory requirements and agreements; and the cost, milestones, and steps proposed to meet the need. The Hanford Site is complex and has a huge budget, and its Activity Data Sheets run to literally thousands of pages. This report summarizes the Activity Data Sheets in a less detailed and much more reader-friendly fashion 5. Interim Hanford Waste Management Plan International Nuclear Information System (INIS) 1985-09-01 The September 1985 Interim Hanford Waste Management Plan (HWMP) is the third revision of this document. In the future, the HWMP will be updated on an annual basis or as major changes in disposal planning at Hanford Site require. The most significant changes in the program since the last release of this document in December 1984 include: (1) Based on studies done in support of the Hanford Defense Waste Environmental Impact Statement (HDW-EIS), the size of the protective barriers covering contaminated soil sites, solid waste burial sites, and single-shell tanks has been increased to provide a barrier that extends 30 m beyond the waste zone. (2) As a result of extensive laboratory development and plant testing, removal of transuranic (TRU) elements from PUREX cladding removal waste (CRW) has been initiated in PUREX. (3) The level of capital support in years beyond those for which specific budget projections have been prepared (i.e., fiscal year 1992 and later) has been increased to maintain Hanford Site capability to support potential future missions, such as the extension of N Reactor/PUREX operations. The costs for disposal of Hanford Site defense wastes are identified in four major areas in the HWMP: waste storage and surveillance, technology development, disposal operations, and capital expenditures 6. Differential turbidity at Hanford International Nuclear Information System (INIS) Laulainen, N.S.; Kleckner, E.W.; Michalsky, J.J.; Stokes, G.M. 1980-01-01 Experiments continued in FY 1979 to examine differential turbidity effects on insolation as measured at the earth's surface. These experiments are primarily intended to provide means for interpreting insolation-data assessment studies. These data are also valuable for inferring aerosol radiative or optical effects, which is an important consideration in evaluating inadvertent climate modification and visibility degradation as a result of aerosols. The experiments are characterized by frequent, nearly simultaneous observations at the Rattlesnake Mountain Observatory (RMO) and the Hanford Meteorological Station (HMS) and take advantage of the nearly 1-km altitude difference between these two observing sites. This study indicated that nearly simultaneous measurements of the direct solar beam from stationary sites that are separated in altitude can be used to monitor the incremental optical depth arising from aerosols in the intervening layer. Once appropriate calbiration procedures have been established for the MASP unit, the direct solar data can be used to document on a routine basis aerosol variations in the first kilometer between HMS and RMO 7. Hanford gas dispersion analysis International Nuclear Information System (INIS) Fujita, R.K.; Travis, J.R. 1994-01-01 An analysis was performed to verify the design of a waste gas exhauster for use in support of rotary core sampling activities at the Westinghouse Hanford Waste Tank Farm. The exhauster was designed to remove waste gases from waste storage tanks during the rotary core drilling process of the solid materials in the tank. Some of the waste gases potentially are very hazardous and must be monitored during the exhauster's operation. If the toxic gas concentrations in specific areas near the exhauster exceed minimum Threshold Limit Values (TLVs), personnel must be excluded from the area. The exhauster stack height is of interest because an increase in stack height will alter the gas concentrations at the critical locations. The exhaust stack is currently ∼4.6 m (15 ft) high. An equipment operator will be located within a 6.1 m (20 ft) radius of the exhaust stack, and his/her head will be at an elevation 3.7 m (12 ft) above ground level (AGL). Therefore, the maximum exhaust gas concentrations at this location must be below the TLV for the toxic gases. Also, the gas concentrations must be within the TLV at a 61 m (200 ft) radius from the stack. If the calculated gas concentrations are above the TLV, where the operator is working below the stack at the 61 m (200 ft) radius location, the stack height may need to be increased 8. 1976 Hanford americium accident International Nuclear Information System (INIS) Heid, K.R.; Breitenstein, B.D.; Palmer, H.E.; McMurray, B.J.; Wald, N. 1979-01-01 This report presents the 2.5-year medical course of a 64-year-old Hanford nuclear chemical operator who was involved in an accident in an americium recovery facility in August 1976. He was heavily externally contaminated with americium, sustained a substantial internal deposition of this isotope, and was burned with concentrated nitric acid and injured by flying debris about the face and neck. The medical care given the patient, including the decontamination efforts and clinical laboratory studies, are discussed. In-vivo measurements were used to estimate the dose rates and the accumulated doses to body organs. Urinary and fecal excreta were collected and analyzed for americium content. Interpretation of these data was complicated by the fact that the intake resulted both from inhalation and from solubilization of the americium embedded in facial tissues. A total of 1100 μCi was excreted in urine and feces during the first 2 years following the accident. The long-term use of diethylenetriaminepentate (DTPA), used principally as the zinc salt, is discussed including the method, route of administration, and effectiveness. To date, the patient has apparently experienced no complications attributable to this extensive course of therapy, even though he has been given approximately 560 grams of DTPA. 4 figures, 1 table 9. Basic mechanisms of atomic displacement production in cubic silicon carbide: A molecular dynamics study International Nuclear Information System (INIS) Malerba, L.; Perlado, J.M. 2002-01-01 Studying the effects of radiation in silicon carbide (SiC) is important for its possible use in both nuclear and electronic technology. One of the most important parameters to describe radiation damage in a material is the threshold displacement energy (TDE). In this paper, the computational technique known as molecular dynamics (MD) is used to determine the TDE's along different crystallographic directions for Si and C atoms in SiC, also allowing for irradiation temperature effects, and to study in detail the mechanisms of atomic displacement production in this material. For this purpose, the widely tested Tersoff potential, implemented in a MD code optimized to study the interaction of high-energy ions with crystals, is used to describe the interatomic forces in SiC. It is found that it is difficult to define a single threshold for this material. Instead, the introduction of two thresholds, upper and lower, becomes necessary. These two thresholds delimit an uncertainty band, within which the displacement may or may not be produced, because the Frenkel pairs generated in such a transferred-kinetic-energy range are metastable. The Arrhenius law expressing the lifetime of one of these metastable defects has also been deduced from the simulation. Finally, on the basis of the results of the simulation, possible values for the recombination distance and the average threshold energy (E d,Si and E d,C ) in SiC are proposed and discussed 10. Atomic physics and synchrotron radiation: The production and accumulation of highly charged ions International Nuclear Information System (INIS) Johnson, B.M.; Meron, M.; Agagu, A.; Jones, K.W. 1986-01-01 Synchrotron radiation can be used to produce highly-charged ions, and to study photoexcitation and photoionization for ions of virtually any element in the periodic table. To date, with few exceptions, atomic physics studies have been limited to rare gases and a few metal vapors, and to photoexcitation energies in the VUV region of the electromagnetic spectrum. These limitations can now be overcome using photons produced by high-brightness synchrotron storage rings, such as the x-ray ring at the National Synchrotron Light Source (NSLS) at Brookhaven. Furthermore, calculations indicate that irradiation of an ion trap with an intense energetic photon beam will result in a viable source of highly-charged ions that can be given the name PHOBIS: the PHOton Beam Ion Source. Promising results, which encourage the wider systematic use of synchrotron radiation in atomic physics research, have been obtained in recent experiments on VUV photoemission and the production and storage of multiply-charged ions. 26 refs., 4 figs., 1 tab 11. Atomistic simulation of damage production by atomic and molecular ion irradiation in GaN International Nuclear Information System (INIS) Ullah, M. W.; Kuronen, A.; Nordlund, K.; Djurabekova, F.; Karaseov, P. A.; Titov, A. I. 2012-01-01 We have studied defect production during single atomic and molecular ion irradiation having an energy of 50 eV/amu in GaN by molecular dynamics simulations. Enhanced defect recombination is found in GaN, in accordance with experimental data. Instantaneous damage shows non-linearity with different molecular projectile and increasing molecular mass. Number of instantaneous defects produced by the PF 4 molecule close to target surface is four times higher than that for PF 2 molecule and three times higher than that calculated as a sum of the damage produced by one P and four F ion irradiation (P+4×F). We explain this non-linearity by energy spike due to molecular effects. On the contrary, final damage created by PF 4 and PF 2 shows a linear pattern when the sample cools down. Total numbers of defects produced by Ag and PF 4 having similar atomic masses are comparable. However, defect-depth distributions produced by these species are quite different, also indicating molecular effect. 12. Production of excited nitrogen atoms and ions by electron impact on nitrogen molecules International Nuclear Information System (INIS) Rall, D.L.A.; Anderson, L.W.; Lin, C.C.; Sharpton, F.A. 1984-01-01 Emission lines of the N atoms and N ions are produced by electron-beam dissociative excitation of N 2 molecules. The ns→3 /sub p/ (n=5 to 9), np→3s (n=3 to 7), nd→3 /sub p/ (n=4 to 8), nf→3d (n=4,5) transitions of N and the 3 /sub p/ →3s, 3d→3p, 4s→3p, 4p→3d, 4f→3d transitions of N + have been observed and optical emission cross sections at various incident electron energies have been measured. The energy dependence of the cross sections of the N emission lines is similar to that of the N + lines at high incident electron energies, but the low-energy behaviors are quite different. These features are explained by the mechanisms involved in the production of the excited N atoms and N + ions. Absolute optical emission cross sections for the N and N + lines are presented 13. Dynamics of production of iodine atoms by dissociation of iodides in a pulsed self-sustained discharge International Nuclear Information System (INIS) Vagin, Nikolai P; Kochetov, Igor' V; Napartovich, A P; Yuryshev, Nikolai N 2013-01-01 Absorption at the laser transition has been used for the first time to assess the evolution of concentration of iodine atoms in a pulsed self-sustained discharge in mixtures of iodides with a buffer gas such as molecular nitrogen and helium. Dynamics of the iodine atom production is studied by the method of absorption spectroscopy. The dissociation of C n F 2n+1 I and CnH 2n+1 I (n = 1, 2) iodides is investigated. The energy required to produce atomic iodine is evaluated. The experimental data obtained for CF 3 I are compared with the results of numerical simulations, their reasonable agreement being demonstrated. (active media) 14. The production and certification of a plutonium equal-atom reference material: NBL CRM 128 International Nuclear Information System (INIS) Crawford, D.W. 1990-07-01 This report describes the design, production, and certification of the New Brunswick Laboratory plutonium equal-atom certified reference material (CRM), NBL CRM 128. The primary use of this CRM is for the determination of bias corrections encountered in the operation of a mass spectrometer. This reference material is available to the US Department of Energy contractor-operated and government-operated laboratories, as well as to the international nuclear safeguards community. The absolute, or unbiased, certified value for the CRM's Pu-242/Pu-239 ratio is 1.00063 ± 0.00026 (95% confidence interval) as of October 1, 1984. This value was obtained through the quantitative blending of high-purity, chemically and isotopically characterized separated isotopes, as well as through intercomparisons of CRM samples with calibration mixtures using thermal ionization mass spectrometry. 32 tabs 15. Selective Production of Renewable para-Xylene by Tungsten Carbide Catalyzed Atom-Economic Cascade Reactions. Science.gov (United States) Dai, Tao; Li, Changzhi; Li, Lin; Zhao, Zongbao Kent; Zhang, Bo; Cong, Yu; Wang, Aiqin 2018-02-12 Tungsten carbide was employed as the catalyst in an atom-economic and renewable synthesis of para-xylene with excellent selectivity and yield from 4-methyl-3-cyclohexene-1-carbonylaldehyde (4-MCHCA). This intermediate is the product of the Diels-Alder reaction between the two readily available bio-based building blocks acrolein and isoprene. Our results suggest that 4-MCHCA undergoes a novel dehydroaromatization-hydrodeoxygenation cascade process by intramolecular hydrogen transfer that does not involve an external hydrogen source, and that the hydrodeoxygenation occurs through the direct dissociation of the C=O bond on the W 2 C surface. Notably, this process is readily applicable to the synthesis of various (multi)methylated arenes from bio-based building blocks, thus potentially providing a petroleum-independent solution to valuable aromatic compounds. © 2018 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim. 16. Hanford Site sustainable development initiatives International Nuclear Information System (INIS) Sullivan, C.T. 1994-05-01 Since the days of the Manhattan Project of World War II, the economic well being of the Tri-Cities (Pasco, Kennewick, and Richland) of Washington State has been tied to the US Department of Energy missions at the nearby Hanford Site. As missions at the Site changed, so did the economic vitality of the region. The Hanford Site is now poised to complete its final mission, that of environmental restoration. When restoration is completed, the Site may be closed and the effect on the local economy will be devastating if action is not taken now. To that end, economic diversification and transition are being planned. To facilitate the process, the Hanford Site will become a sustainable development demonstration project 17. FLUOR HANFORD SAFETY MANAGEMENT PROGRAMS Energy Technology Data Exchange (ETDEWEB) GARVIN, L. J.; JENSEN, M. A. 2004-04-13 This document summarizes safety management programs used within the scope of the ''Project Hanford Management Contract''. The document has been developed to meet the format and content requirements of DOE-STD-3009-94, ''Preparation Guide for US. Department of Energy Nonreactor Nuclear Facility Documented Safety Analyses''. This document provides summary descriptions of Fluor Hanford safety management programs, which Fluor Hanford nuclear facilities may reference and incorporate into their safety basis when producing facility- or activity-specific documented safety analyses (DSA). Facility- or activity-specific DSAs will identify any variances to the safety management programs described in this document and any specific attributes of these safety management programs that are important for controlling potentially hazardous conditions. In addition, facility- or activity-specific DSAs may identify unique additions to the safety management programs that are needed to control potentially hazardous conditions. 18. HANFORD SITE RIVER CORRIDOR CLEANUP International Nuclear Information System (INIS) BAZZELL, K.D. 2006-01-01 In 2005, the US Department of Energy (DOE) launched the third generation of closure contracts, including the River Corridor Closure (RCC) Contract at Hanford. Over the past decade, significant progress has been made on cleaning up the river shore that bordes Hanford. However, the most important cleanup challenges lie ahead. In March 2005, DOE awarded the Hanford River Corridor Closure Contract to Washington Closure Hanford (WCH), a limited liability company owned by Washington Group International, Bechtel National and CH2M HILL. It is a single-purpose company whose goal is to safely and efficiently accelerate cleanup in the 544 km 2 Hanford river corridor and reduce or eliminate future obligations to DOE for maintaining long-term stewardship over the site. The RCC Contract is a cost-plus-incentive-fee closure contract, which incentivizes the contractor to reduce cost and accelerate the schedule. At1.9 billion and seven years, WCH has accelerated cleaning up Hanford's river corridor significantly compared to the $3.2 billion and 10 years originally estimated by the US Army Corps of Engineers. Predictable funding is one of the key features of the new contract, with funding set by contract at$183 million in fiscal year (FY) 2006 and peaking at $387 million in FY2012. Another feature of the contract allows for Washington Closure to perform up to 40% of the value of the contract and subcontract the balance. One of the major challenges in the next few years will be to identify and qualify sufficient subcontractors to meet the goal 19. Atomic Layer Deposition to Enable the Production, Optimization and Protection of Spaceflight Hardware Data.gov (United States) National Aeronautics and Space Administration — Atomic Layer Deposition (ALD) a cost effective nano-manufacturing technique allows for the conformal coating of substrates with atomic control in a benign... 20. Atomic Layer Deposition to Enable the Production, Optimization and Protection of Spaceflight Hardware Project Data.gov (United States) National Aeronautics and Space Administration — Atomic Layer Deposition (ALD) a cost effective nano-manufacturing technique allows for the conformal coating of substrates with atomic control in a benign... 1. Hanford Environmental Dose Reconstruction Project International Nuclear Information System (INIS) Finch, S.M.; McMakin, A.H. 1991-04-01 The objective of the Hanford Environmental Dose Reconstruction Project is to estimate the radiation doses that populations could have received from nuclear operations at Hanford since 1944. The project is being managed and conducted by the Pacific Northwest Laboratory (PNL) under the direction of an independent Technical Steering Panel (TSP). The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed, from released to impact on humans (dose estimates): source terms; environmental transport; environmental monitoring data; demographics, agriculture, food habits; and, environmental pathways and dose estimates 2. Hanford Environmental Dose Reconstruction Project International Nuclear Information System (INIS) Finch, S.M.; McMakin, A.H. 1992-06-01 The objective of the Hanford Environmental Dose Reconstruction Project is to estimate the radiation doses that individuals and populations could have received from nuclear operations at Hanford since 1944. The project is being managed and conducted by the Battelle Pacific Northwest Laboratories under contract with the Centers for Disease Control. The independent Technical Steering Panel (TSP) provides technical direction. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed, from release to impact on humans (dose estimates): source terms; environmental transport; environmental monitoring data; demography, food consumption, and agriculture; environmental pathways and dose estimates 3. Disposal of Hanford defense waste International Nuclear Information System (INIS) Holten, R.A.; Burnham, J.B.; Nelson, I.C. 1986-01-01 An Environmental Impact Statement (EIS) on the disposal of Hanford Defense Waste is scheduled to be released near the end of March, 1986. This EIS will evaluate the impacts of alternatives for disposal of high-level, tank, and transuranic wastes which are now stored at the Department of Energy's Hanford Site or will be produced there in the future. In addition to releasing the EIS, the Department of Energy is conducting an extensive public participation process aimed at providing information to the public and receiving comments on the EIS 4. Hanford isotope project strategic business analysis yttrium-90 (Y-90) Energy Technology Data Exchange (ETDEWEB) NONE 1995-10-01 The purpose of this analysis is to address the short-term direction for the Hanford yttrium-90 (Y-90) project. Hanford is the sole DOE producer of Y-90, and is the largest repository for its source in this country. The production of Y-90 is part of the DOE Isotope Production and Distribution (IP and D) mission. The Y-90 is milked from strontium-90 (Sr-90), a byproduct of the previous Hanford missions. The use of Sr-90 to produce Y-90 could help reduce the amount of waste material processed and the related costs incurred by the clean-up mission, while providing medical and economic benefits. The cost of producing Y-90 is being subsidized by DOE-IP and D due to its use for research, and resultant low production level. It is possible that the sales of Y-90 could produce full cost recovery within two to three years, at two curies per week. Preliminary projections place the demand at between 20,000 and 50,000 curies per year within the next ten years, assuming FDA approval of one or more of the current therapies now in clinical trials. This level of production would incentivize private firms to commercialize the operation, and allow the government to recover some of its sunk costs. There are a number of potential barriers to the success of the Y-90 project, outside the control of the Hanford Site. The key issues include: efficacy, Food and Drug Administration (FDA) approval and medical community acceptance. There are at least three other sources for Y-90 available to the US users, but they appear to have limited resources to produce the isotope. Several companies have communicated interest in entering into agreements with Hanford for the processing and distribution of Y-90, including some of the major pharmaceutical firms in this country. 5. Hanford isotope project strategic business analysis yttrium-90 (Y-90) International Nuclear Information System (INIS) 1995-10-01 The purpose of this analysis is to address the short-term direction for the Hanford yttrium-90 (Y-90) project. Hanford is the sole DOE producer of Y-90, and is the largest repository for its source in this country. The production of Y-90 is part of the DOE Isotope Production and Distribution (IP and D) mission. The Y-90 is ''milked'' from strontium-90 (Sr-90), a byproduct of the previous Hanford missions. The use of Sr-90 to produce Y-90 could help reduce the amount of waste material processed and the related costs incurred by the clean-up mission, while providing medical and economic benefits. The cost of producing Y-90 is being subsidized by DOE-IP and D due to its use for research, and resultant low production level. It is possible that the sales of Y-90 could produce full cost recovery within two to three years, at two curies per week. Preliminary projections place the demand at between 20,000 and 50,000 curies per year within the next ten years, assuming FDA approval of one or more of the current therapies now in clinical trials. This level of production would incentivize private firms to commercialize the operation, and allow the government to recover some of its sunk costs. There are a number of potential barriers to the success of the Y-90 project, outside the control of the Hanford Site. The key issues include: efficacy, Food and Drug Administration (FDA) approval and medical community acceptance. There are at least three other sources for Y-90 available to the US users, but they appear to have limited resources to produce the isotope. Several companies have communicated interest in entering into agreements with Hanford for the processing and distribution of Y-90, including some of the major pharmaceutical firms in this country 6. Waste management (Truck and rail shipments to Hanford) International Nuclear Information System (INIS) O'Donnell, J.P.; Culbertson, R.C. 1988-01-01 As part of the physical decommissioning of the Shippingport Atomic Power Station, Shippingport, PA, a large volume of Low Specific Activity (LSA) radioactive waste was accumulated. The waste, which consisted primarily of radioactive reactor plant components, piping, contaminated asbestos, tanks, building rubble, sludge and ion exchange resins was packaged and prepared for shipment. The waste was transported by truck and rail from Shippingport, PA, to the Department of Energy burial ground at Hanford, Washington, a journey of 2,329 miles. This presentation will discuss the successful management of over 2,600 packages weighing in excess of 3,600 tons of radioactive waste from the cradle-to-the-grave, that is from the time it was generated during the decommissioning process until its final burial at the Hanford, Washington burial site. 1 tab 7. Optical measurements of atomic oxygen concentration, temperature and nitric oxide production rate in flames Science.gov (United States) Myhr, Franklin Henry An optical method for measuring nitric oxide (NO) production rates in flames was developed and characterized in a series of steady, one-dimensional, atmospheric-pressure laminar flames of 0.700 Hsb2/0.199 Nsb2/0.101 COsb2 or 0.700 CHsb4/0.300 Nsb2 (by moles) with dry air, with equivalence ratios from 0.79 to 1.27. Oxygen atom concentration, (O), was measured by two-photon laser-induced fluorescence (LIF), temperature was measured by ultraviolet Rayleigh scattering, and nitrogen concentration was calculated from supplied reactant flows; together this information was used to calculate the NO production rate through the thermal (Zel'dovich) mechanism. Measurements by two other techniques were compared with results from the above method. In the first comparison, gas sampling was used to measure axial NO concentration profiles, the slopes of which were multiplied by velocity to obtain total NO production rates. In the second comparison, LIF measurements of hydroxyl radical (OH) were used with equilibrium water concentrations and a partial equilibrium assumption to find (O). Nitric oxide production rates from all three methods agreed reasonably well. Photolytic interference was observed during (O) LIF measurements in all of the flames; this is the major difficulty in applying the optical technique. Photolysis of molecular oxygen in lean flames has been well documented before, but the degree of interference observed in the rich flames suggests that some other molecule is also dissociating; the candidates are OH, CO, COsb2 and Hsb2O. An extrapolative technique for removing the effects of photolysis from (O) LIF measurements worked well in all flames where NO production was significant. Using the optical method to measure NO production rates in turbulent flames will involve a tradeoff among spatial resolution, systematic photolysis error, and random shot noise. With the conventional laser system used in this work, a single pulse with a resolution of 700 mum measured NO 8. Hanford Site technical baseline database. Revision 1 International Nuclear Information System (INIS) Porter, P.E. 1995-01-01 This report lists the Hanford specific files (Table 1) that make up the Hanford Site Technical Baseline Database. Table 2 includes the delta files that delineate the differences between this revision and revision 0 of the Hanford Site Technical Baseline Database. This information is being managed and maintained on the Hanford RDD-100 System, which uses the capabilities of RDD-100, a systems engineering software system of Ascent Logic Corporation (ALC). This revision of the Hanford Site Technical Baseline Database uses RDD-100 version 3.0.2.2 (see Table 3). Directories reflect those controlled by the Hanford RDD-100 System Administrator. Table 4 provides information regarding the platform. A cassette tape containing the Hanford Site Technical Baseline Database is available 9. Accelerated clean-up at the Hanford Site International Nuclear Information System (INIS) Frain, J.M.; Johnson, W.L. 1994-01-01 The Hanford Site began operations in 1943 as one of the sites for plutonium production associated with the Manhattan Project. It has been used, in part, for nuclear reactor operation, reprocessing of spent fuel, and management of radioactive waste. The Hanford Site covers approximately 1,434 km 2 (560 mi 2 2) in southeastern Washington State. The subject of this paper, the 618-9 Burial Ground, is located on the Hanford Site approximately 1.6 km (1 mi) west of the Columbia River, and a few miles north of Richland, Washington. Throughout Hanford Site history, prior to legislation regarding disposal of chemical waste products, some chemical waste byproducts were disposed ,ia burial in trenches. One such trench was the 618-9 Burial Ground. This burial ground was suspected to contain approximately 19,000 L (5,000 gal) of uranium-contaminated organic solvent, disposed in standard 55-gal (208-L) metal drums. The waste was produced from research and development activities related to fuel reprocessing 10. Advancements in Ti Alloy Powder Production by Close-Coupled Gas Atomization Energy Technology Data Exchange (ETDEWEB) Heidloff, Andy; Rieken, Joel; Anderson, Iver; Byrd, David 2011-04-01 As the technology for titanium metal injection molding (Ti-MIM) becomes more readily available, efficient Ti alloy fine powder production methods are required. An update on a novel close-coupled gas atomization system has been given. Unique features of the melting apparatus are shown to have measurable effects on the efficiency and ability to fully melt within the induction skull melting system (ISM). The means to initiate the melt flow were also found to be dependent on melt apparatus. Starting oxygen contents of atomization feedstock are suggested based on oxygen pick up during the atomization and MIM processes and compared to a new ASTM specification. Forming of titanium by metal injection molding (Ti-MIM) has been extensively studied with regards to binders, particle shape, and size distribution and suitable de-binding methods have been discovered. As a result, the visibility of Ti-MIM has steadily increased as reviews of technology, acceptability, and availability have been released. In addition, new ASTM specification ASTM F2885-11 for Ti-MIM for biomedical implants was released in early 2011. As the general acceptance of Ti-MIM as a viable fabrication route increases, demand for economical production of high quality Ti alloy powder for the preparation of Ti-MIM feedstock correspondingly increases. The production of spherical powders from the liquid state has required extensive pre-processing into different shapes thereby increasing costs. This has prompted examination of Ti-MIM with non-spherical particle shape. These particles are produced by the hydride/de-hydride process and are equi-axed but fragmented and angular which is less than ideal. Current prices for MIM quality titanium powder range from$40-$220/kg. While it is ideal for the MIM process to utilize spherical powders within the size range of 0.5-20 {mu}m, titanium's high affinity for oxygen to date has prohibited the use of this powder size range. In order to meet oxygen requirements the top 11. Products and mechanism of the reaction of Cl atoms with unsaturated alcohols Science.gov (United States) Rodríguez, Ana; Rodríguez, Diana; Soto, Amparo; Bravo, Iván; Diaz-de-Mera, Yolanda; Notario, Alberto; Aranda, Alfonso 2012-04-01 The products of the chlorine atom initiated oxidation of different unsaturated alcohols were determined at atmospheric pressure and ambient temperature, in a 400 L teflon reaction chamber using GC-FID and GC-MS for the analysis. The major products detected (with molar yields in brackets) are: chloroacetaldehyde (50 ± 8%) and acrolein (27 ± 2%) from allyl alcohol; acetaldehyde (77 ± 11%), chloroacetaldehyde (75 ± 18%), and methyl vinyl ketone (17 ± 2%) from 3-buten-2-ol; acetone (55 ± 4%) and chloroacetaldehyde (59 ± 8%) from 2-methyl-3-buten-2-ol; chloroacetone (18 ± 1%) and methacrolein (8 ± 1%) from 2-methyl-2-propen-1-ol; acetaldehyde (20 ± 1%), crotonaldehyde (6 ± 3%), 3-choloro-4-hydroxy-2-butanone (2 ± 2%) and 2-chloro-propanal (4 ± 5%) from crotyl alcohol; and acetone (24 ± 3%) from 3-methyl-2-buten-1-ol. The experimental data suggests that addition of Cl to the double bond of the unsaturated alcohol is the dominant reaction pathway compared to the H-abstraction channel. 12. Generalized position-momentum uncertainty products: Inclusion of moments with negative order and application to atoms International Nuclear Information System (INIS) Angulo, J. C. 2011-01-01 Rigorous and universal relationships among radial expectation values of any D-dimensional quantum-mechanical system are obtained, using Renyi-like position-momentum inequalities in an information-theoretical framework. Although the results are expressed in terms of four moments (two in position space and two in the momentum one), especially interesting are the cases that provide expressions of uncertainty in terms of products a > 1/a b > 1/b , widely considered in the literature, including the famous Heisenberg relationship 2 > 2 >≥D 2 /4. Improved bounds for these products have recently been provided, but are always restricted to positive orders a,b>0. The interesting part of this work are the inequalities for negative orders. A study of these relationships is carried out for atomic systems in their ground state. Some results are given in terms of relevant physical quantities, including the kinetic and electron-nucleus attraction energies, the diamagnetic susceptibility, and the height of the peak of the Compton profile, among others. 13. Forward electron production in heavy ion-atom and ion-solid collisions International Nuclear Information System (INIS) Sellin, I.A. 1984-01-01 A sharp cusp in the velocity spectrum of electrons, ejected in ion-atom and ion-solid collisions, is observed when the ejected electron velocity vector v/sub e/ matches that of the emergent ion vector v/sub p/ in both speed and direction. In ion-atom collisions, the electrons originate from capture to low-lying, projectile-centered continuum states (ECC) for fast bare or nearly bare projectiles, and from loss to those low-lying continuum states (ELC) when loosely bound projectile electrons are available. Most investigators now agree that ECC cusps are strongly skewed toward lower velocities, and exhibit full widths half maxima roughly proportional to v/sub p/ (neglecting target-shell effects, which are sometimes strong). A close examination of recent ELC data shows that ELC cusps are instead nearly symmetric, with widths nearly independent on v/sub p/ in the velocity range 6 to 18 a.u., a result only recently predicted by theory. Convoy electron cusps produced in heavy ion-solid collisions at MeV/u energies exhibit approximately velocity-independent widths very similar to ELC cusp widths. While the shape of the convoy peaks is approximately independent of projectile Z, velocity, and of target material, it is found that the yields in polycrystalline targets exhibit a strong dependence on projectile Z and velocity. While attempts have been made to link convoy electron production to binary ECC or ELC processes, sometimes at the last layer, or alternatively to a solid-state wake-riding model, our measured dependences of cusp shape and yield on projectile charge state and energy are inconsistent with the predictions of available theories. 10 references, 8 figures, 1 table 14. Hanford Site environmental management specification International Nuclear Information System (INIS) Grygiel, M.L. 1998-01-01 The US Department of Energy, Richland Operations Office (RL) uses this Hanford Site Environmental Management Specification (Specification) to document top-level mission requirements and planning assumptions for the prime contractors involved in Hanford Site cleanup and infrastructure activities under the responsibility of the US Department of Energy, Office of Environmental Management. This Specification describes at a top level the activities, facilities, and infrastructure necessary to accomplish the cleanup of the Hanford Site and assigns this scope to Site contractors and their respective projects. This Specification also references the key National Environmental Policy Act of 1969 (NEPA), Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA), and safety documentation necessary to accurately describe the cleanup at a summary level. The information contained in this document reflects RL's application of values, priorities, and critical success factors expressed by those involved with and affected by the Hanford Site project. The prime contractors and their projects develop complete baselines and work plans to implement this Specification. These lower-level documents and the data that support them, together with this Specification, represent the full set of requirements applicable to the contractors and their projects. Figure 1-1 shows the relationship of this Specification to the other basic Site documents. Similarly, the documents, orders, and laws referenced in this specification represent only the most salient sources of requirements. Current and contractual reference data contain a complete set of source documents 15. Hanford Site Waste Management Plan International Nuclear Information System (INIS) 1988-12-01 The Hanford Site Waste Management Plan (HWMP) was prepared in accordance with the outline and format described in the US Department of Energy Orders. The HWMP presents the actions, schedules, and projected costs associated with the management and disposal of Hanford defense wastes, both radioactive and hazardous. The HWMP addresses the Waste Management Program. It does not include the Environmental Restoration Program, itself divided into the Environmental Restoration Remedial Action Program and the Decontamination and Decommissioning Program. The executive summary provides the basis for the plans, schedules, and costs within the scope of the Waste Management Program at Hanford. It summarizes fiscal year (FY) 1988 including the principal issues and the degree to which planned activities were accomplished. It further provides a forecast of FY 1989 including significant milestones. Section 1 provides general information for the Hanford Site including the organization and administration associated with the Waste Management Program and a description of the Site focusing on waste management operations. Section 2 and Section 3 describe radioactive and mixed waste management operations and hazardous waste management, respectively. Each section includes descriptions of the waste management systems and facilities, the characteristics of the wastes managed, and a discussion of the future direction of operations 16. Differential turbidity measurements at Hanford International Nuclear Information System (INIS) Laulainen, N.S.; Bates, J.A.; Kleckner, E.W.; Michalsky, J.J.; Schrotke, P.M.; Thorp, J.M. 1978-01-01 An experiment to exmine differential turbidity effects on measured insolation between the Rattlesnake Observatory and the Hanford Meteorological Station was conducted during summer 1977. Several types of solar radiation instruments were used, including pyranometers, multiwavelength sunphotometers, and an active cavity radiometer. Preliminary results show dramatic temporal variability of aerosol loading at HMS and significant insolation and turbidity differences between the Observatory and HMS 17. Hanford Site environmental management specification Energy Technology Data Exchange (ETDEWEB) Grygiel, M.L. 1998-06-10 The US Department of Energy, Richland Operations Office (RL) uses this Hanford Site Environmental Management Specification (Specification) to document top-level mission requirements and planning assumptions for the prime contractors involved in Hanford Site cleanup and infrastructure activities under the responsibility of the US Department of Energy, Office of Environmental Management. This Specification describes at a top level the activities, facilities, and infrastructure necessary to accomplish the cleanup of the Hanford Site and assigns this scope to Site contractors and their respective projects. This Specification also references the key National Environmental Policy Act of 1969 (NEPA), Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA), and safety documentation necessary to accurately describe the cleanup at a summary level. The information contained in this document reflects RLs application of values, priorities, and critical success factors expressed by those involved with and affected by the Hanford Site project. The prime contractors and their projects develop complete baselines and work plans to implement this Specification. These lower-level documents and the data that support them, together with this Specification, represent the full set of requirements applicable to the contractors and their projects. Figure 1-1 shows the relationship of this Specification to the other basic Site documents. Similarly, the documents, orders, and laws referenced in this specification represent only the most salient sources of requirements. Current and contractual reference data contain a complete set of source documents. 18. Mortality of Hanford radiation workers International Nuclear Information System (INIS) Gilbert, E.S. 1979-01-01 The effects of occupational exposure to low level ionizing radiation at the Hanford plant in southeastern Washington were investigated. Death rates were related to exposure status. To provide perspective, the rates were also compared with the death rates of the US population 19. Hanford site operator changes management International Nuclear Information System (INIS) Anon. 1994-01-01 This article is a brief discussion of management changes at the Westinghouse Hanford Corporation. A. LeMar Trego has relieved Thomas Anderson as president of WHC. This was in response to recent shortcomings in Westinghouse's management of the environmental restoration and their failure to receive a$10M performance bonus 20. Hanford Patrol Academy Demolition Sites Closure Plan International Nuclear Information System (INIS) 1992-11-01 From 1975 to 1991 the Hanford Patrol Academy Demolition Sites (HPADS) were used for demolition events. These demolition events were a form of thermal treatment for spent or abandoned chemical waste. Because the HPADS will no longer be used for this thermal activity, the sites will be closed. Closure will be conducted pursuant to the requirements of the Washington State Department of Ecology (Ecology) Dangerous Waste Regulations, Washington Administrative Code (WAC) 173-303-610 and 40 CFR 270.1. Closure also will satisfy closure requirements of WAC 173-303-680 and for the thermal treatment closure requirements of 40 CFR 265.381. This closure plan presents a description of the HPADS, the history of the waste treated, and the approach that will be followed to close the HPADS. Because dangerous waste does not include the source, special nuclear, and by-product material components of mixed waste, radionuclides are not within the scope of WAC 173-303 or of this closure plan. The information on radionuclides is provided only for general knowledge where appropriate. Only dangerous constituents derived from HPADS operations will be addressed in this closure plan in accordance with WAC 173-303-610(2)(b)(i). The HPADS are actually two distinct soil closure areas within the Hanford Patrol Academy training area 1. HIGH ALUMINUM HLW GLASSES FOR HANFORD'S WTP International Nuclear Information System (INIS) Kruger, A.A.; Joseph, I.; Bowman, B.W.; Gan, H.; Kot, W.; Matlack, K.S.; Pegg, I.L 2009-01-01 The world's largest radioactive waste vitrification facility is now under construction at the United State Department of Energy's (DOE's) Hanford site. The Hanford Tank Waste Treatment and Immobilization Plant (WTP) is designed to treat nearly 53 million gallons of mixed hazardous and radioactive waste now residing in 177 underground storage tanks. This multi-decade processing campaign will be one of the most complex ever undertaken because of the wide chemical and physical variability of the waste compositions generated during the cold war era that are stored at Hanford. The DOE Office of River Protection (ORP) has initiated a program to improve the long-term operating efficiency of the WTP vitrification plants with the objective of reducing the overall cost of tank waste treatment and disposal and shortening the duration of plant operations. Due to the size, complexity and duration of the WTP mission, the lifecycle operating and waste disposal costs are substantial. As a result, gains in High Level Waste (HLW) and Low Activity Waste (LAW) waste loadings, as well as increases in glass production rate, which can reduce mission duration and glass volumes for disposal, can yield substantial overall cost savings. EnergySolutions and its long-term research partner, the Vitreous State Laboratory (VSL) of the Catholic University of America, have been involved in a multi-year ORP program directed at optimizing various aspects of the HLW and LAW vitrification flow sheets. A number of Hanford HLW streams contain high concentrations of aluminum, which is challenging with respect to both waste loading and processing rate. Therefore, a key focus area of the ORP vitrification process optimization program at EnergySolutions and VSL has been development of HLW glass compositions that can accommodate high Al 2 O 3 concentrations while maintaining high processing rates in the Joule Heated Ceramic Melters (JHCMs) used for waste vitrification at the WTP. This paper, reviews the 2. Production and quality control of concrete for the Rajasthan Atomic Power Station - [Part 1 International Nuclear Information System (INIS) Singha Roy, P.K.; Sukhtankar, K.D.; Prasad, K. 1975-01-01 The production and quality control of concrete and concrete materials for the construction of the twin-reactor Rajasthan Atomic Power Station with its 400 MW net capacity posed many challenges since many of the requirements for the properties of concrete were new and were being laid down for the first time in India. Some of the conditions for the concrete included leak-tightness against gas pressure, total absence of shrinkage in the containment even when the ambient temperature during concreting was as high as 45degC, placing concrete at a temperature as low as 8degC, the use of non-shrink and high strength grout, absolute impermeability against water, high density for radiation shielding, controlled modulus of elasticity for large machine foundations, high strength with high slump for the prestressed concrete dome, etc. Though the total quantity of concrete was not very much compared with a large river valley or steel plant project, (e.g., about 1.2 X 10 6 m 3 for a 2-million tonne steel plant) it was quite significant, being about 70,000 m 3 of normal density and 2,100 m 3 of high density concrete. The production of these quantities entailed intensive material study and investigation, development of new mixes with additives not tried out before in the country, and design and quality control techniques which were unique in many respects. The paper deals with the production and quality control of concrete, including grouts used in the projects, but the actual concreting and construction operations are not discussed. (author) 3. Strategic plan for Hanford Site Environmental Restoration Information Management International Nuclear Information System (INIS) Cowley, P.J.; Beck, J.E.; Gephart, R.E. 1994-06-01 This strategic plan addresses information management for the Environmental Restoration (ER) Program at the Hanford Site. This Program leads the cleanup of the Hanford Site's soil, groundwater, buried waste, and the decontamination and decommissioning of facilities. The vision that drives this strategic plan is to ensure that quality information is available to the people who need it, when they need it, at a convenient location, in a usable form, and at an acceptable cost. Although investments are being made in managing the vast amounts of information, which include data, records and documents associated with the Hanford Site's production history and new cleanup mission, it is widely recognized that efforts to date have not accomplished the vision. Effective information management involves more than the compilation of massive amounts of electronic and non-electronic information. It also involves integrating information management into business processes that support user's needs and decisionmaking. Only then can information management complement and enable environmental restoration priorities and practices, help identify environmental restoration requirements, and enable communication within the Environmental Restoration Program and between the Program and its stakeholders. Successfully accomplishing the Hanford Site mission requires an integrated approach to information management that crosses organizational boundaries, streamlines existing systems, and builds new systems that support the needs of the future. This plan outlines that approach 4. Hanford Environmental Dose Reconstruction Project independent direction and oversight International Nuclear Information System (INIS) Blazek, M.L.; Power, M. 1991-01-01 Hanford was selected in 1942 as one of the sites for the Manhattan Project. It produced plutonium for one of the world's first nuclear weapons. The US Department of Energy (DOE) and its predecessors continued to make plutonium for nuclear weapons at Hanford for more than four decades. In the early days of Hanford operations, radioactive materials routinely were released to the environment by many processes. The DOE disclosed documents about these releases in 1986. In 1987, Washington, Oregon, and regional Indian tribes gathered an independent panel of experts. This group recommended dose reconstruction and health effects feasibility studies. Later that year, DOE hired Battelle Pacific Northwest Laboratory (PNL) to reconstruct potential public radiation doses from Hanford's past releases of radioactive material. The DOE agreed with the states and tribes that project direction would come from an independent technical steering panel (TSP). This approach was critical to gain public credibility for the project and the science. The TSP directs the project and makes policy. That is now clear - but, it was hard-earned. Conducting science in an open public process is new, challenging, and clearly worthwhile. The panel's product is good science that is believed and accepted by the public - our client 5. Radionuclide releases to the Columbia River from Hanford Operations, 1944--1971 International Nuclear Information System (INIS) Heeb, C.M.; Bates, D.J. 1994-05-01 The purpose of the Hanford Environmental Dose Reconstruction (HEDR) Project is to estimate the radiation dose that individuals could have received as a result of radionuclide emissions since 1944 from the Hanford Site. One source of radionuclide releases to the Columbia River was from production reactor operations. This report provides a quantitative estimate of the amount of radioactivity released each month (1944--1971) to the Columbia River from eleven radionuclides as well as from gross beta activity 6. Production mechanism of atomic nitrogen in atmospheric pressure pulsed corona discharge measured using two-photon absorption laser-induced fluorescence International Nuclear Information System (INIS) Teramoto, Yoshiyuki; Ono, Ryo; Oda, Tetsuji 2012-01-01 To study the production mechanism of atomic nitrogen, the temporal profile and spatial distribution of atomic nitrogen are measured in atmospheric pressure pulsed positive corona discharge using two-photon absorption laser-induced fluorescence. The absolute atomic nitrogen density in the streamer filaments is estimated from decay rate of atomic nitrogen in N 2 discharge. The results indicate that the absolute atomic nitrogen density is approximately constant against discharge energy. When the discharge voltage is 21.5 kV, production yield of atomic nitrogen produced by an N 2 discharge pulse is estimated to be 2.9 - 9.8 × 10 13 atoms and the energy efficiency of atomic nitrogen production is estimated to be about 1.8 - 6.1 × 10 16 atoms/J. The energy efficiency of atomic nitrogen production in N 2 discharge is constant against the discharge energy, while that in N 2 /O 2 discharge increases with discharge energy. In the N 2 /O 2 discharge, two-step process of N 2 dissociation plays significant role for atomic nitrogen production. 7. Investigations on the production of labelled organic compounds by recoil labelling with gamma,n-produced 11-C-atoms International Nuclear Information System (INIS) Wagenbach, U. 1981-01-01 ''Hot'' 11 C atoms are produced from 12 C(γ,n) 11 C nuclear reactions by bremsstrahlung at the 65 MeV electron linear accelerator in Giessen. The relative retention in various C-atoms of the amino acid, methionine, is determined by splitting of the terminal C-atoms of the molecule and by independent determination of the content of 11 C in the isolated and derived fragments. The terminal groups (thiomethyl or carboxyl groups) each carry approx. 25% of the total retained radioactivity, the remaining 50% being spread over the three inner carbon atoms. The activation of alkylamines, crystallised as hydrochlorides, hydrofluorides, oxalates and sulphates, leads to similar yields of direct labelling from 5 to 15%. Amines activated in the liquid state show a retention of less than 5%. The yields for labelled synthetic products are between 10 and 15% for amino acids and are often higher for crystallised amines. Amines activated in the liquid state produced greater yields of synthesis products but at the same time an increase in the product range. The labelled synthesis products can be separated faster by suitable methods such as preparative HPLC and are then available for carrier-free studies in the life sciences. (orig./EF) [de 8. Raptors of the Hanford Site and nearby areas of southcentral Washington International Nuclear Information System (INIS) Fitzner, R.E.; Rickard, W.H.; Cadwell, L.L.; Rogers, L.E. 1981-05-01 This report is concerned with the birds of prey which use the Hanford Site not only during the nesting season but throughout the year. An ecological treatment of five nesting owls (great horned, long-eared, short-eared, barn and burrowing) and five nesting hawks (marsh hawk, red-tailed hawk, Swainson's hawk, prairie falcon and American kestrel) is provided and supportive information on non-nesting species is presented. Factors which control raptor densities and population dynamics throughout all seasons of the year are discussed. Information is also provided for raptors from other areas of southcentral Washington in order to yield a comprehensive picture of how the Hanford Site fits in with regional bird of prey populations. The following were the objectives of this study: (1) to determine the numbers of birds of prey nesting on the Hanford Site, (2) to document the reproductive chronology of each nesting raptor species, (3) to provide analyses of food habits of birds of prey on the Hanford Site coupled with prey abundance data, (4) to determine the productivity of the dominant large birds of prey on the Hanford Site, (5) to determine the distribution and land use patterns of all raptors on the Hanford Site, (6) to determine the kinds and relative abundance of non-nesting raptors on the Hanford Site and adjacent areas of southcentral Washington (7) to document present land use practices on the Hanford Site and their effects on raptors, (8) to document radionuclide levels in birds of prey on the Hanford Site, and (9) to determine the role of birds of prey in radioecological monitoring 9. Raptors of the Hanford Site and nearby areas of southcentral Washington Energy Technology Data Exchange (ETDEWEB) Fitzner, R.E.; Rickard, W.H.; Cadwell, L.L.; Rogers, L.E. 1981-05-01 This report is concerned with the birds of prey which use the Hanford Site not only during the nesting season but throughout the year. An ecological treatment of five nesting owls (great horned, long-eared, short-eared, barn and burrowing) and five nesting hawks (marsh hawk, red-tailed hawk, Swainson's hawk, prairie falcon and American kestrel) is provided and supportive information on non-nesting species is presented. Factors which control raptor densities and population dynamics throughout all seasons of the year are discussed. Information is also provided for raptors from other areas of southcentral Washington in order to yield a comprehensive picture of how the Hanford Site fits in with regional bird of prey populations. The following were the objectives of this study: (1) to determine the numbers of birds of prey nesting on the Hanford Site, (2) to document the reproductive chronology of each nesting raptor species, (3) to provide analyses of food habits of birds of prey on the Hanford Site coupled with prey abundance data, (4) to determine the productivity of the dominant large birds of prey on the Hanford Site, (5) to determine the distribution and land use patterns of all raptors on the Hanford Site, (6) to determine the kinds and relative abundance of non-nesting raptors on the Hanford Site and adjacent areas of southcentral Washington (7) to document present land use practices on the Hanford Site and their effects on raptors, (8) to document radionuclide levels in birds of prey on the Hanford Site, and (9) to determine the role of birds of prey in radioecological monitoring. 10. 1997 annual report on waste generation and waste minimization progress as required by DOE Order 5400.1, Hanford Site International Nuclear Information System (INIS) Segall, P. 1998-01-01 Hanford's missions are to safely clean up and manage the site's legacy wastes, and to develop and deploy science and technology. Through these missions Hanford will contribute to economic diversification of the region. Hanford's environmental management or cleanup mission is to protect the health and safety of the public, workers, and the environment; control hazardous materials; and utilize the assets (people, infra structure, site) for other missions. Hanford's science and technology mission is to develop and deploy science and technology in the service of the nation including stewardship of the Hanford Site. Pollution Prevention is a key to the success of these missions by reducing the amount of waste to be managed and identifying/implementing cost effective waste reduction projects. Hanford's original mission, the production of nuclear materials for the nation's defense programs, lasted more than 40 years, and like most manufacturing operations, Hanford's operations generated large quantities of waste and pollution. However, the by-products from Hanford operations pose unique problems like radiation hazards, vast volumes of contaminated water and soil, and many contaminated structures including reactors, chemical plants and evaporation ponds. The cleanup activity is an immense and challenging undertaking, which includes characterization and decommissioning of 149 single shell storage tanks, treating 28 double shell tanks, safely disposing of over 2,100 metric tons of spent nuclear fuel stored on site, removing numerous structures, and dealing with significant solid waste, ground water, and land restoration issues 11. DEVELOPMENT OF HIGH-DENSITY U/AL DISPERSION PLATES FOR MO-99 PRODUCTION USING ATOMIZED URANIUM POWDER OpenAIRE RYU, HO JIN; KIM, CHANG KYU; SIM, MOONSOO; PARK, JONG MAN; LEE, JONG HYUN 2013-01-01 Uranium metal particle dispersion plates have been proposed as targets for Molybdenum-99 (Mo-99) production to improve the radioisotope production efficiency of conventional low enriched uranium targets. In this study, uranium powder was produced by centrifugal atomization, and miniature target plates containing uranium particles in an aluminum matrix with uranium densities up to 9 g-U/cm3 were fabricated. Additional heat treatment was applied to convert the uranium particles into UAlx compou... 12. ASEA-ATOM's URANUS system for production control, economic control and safeguards International Nuclear Information System (INIS) Mattson, J.C. 1983-01-01 ASEA-ATOM needs a system for production and economic control because: (1) the uranium is the valuable property of the customer; (2) short delivery times are a prime means of competition; (3) the manufacture of fuel necessitates frequent enrichment changes and much enrichment blending; (4) minimizing uranium stock reduces interest costs. A system which meets the above needs will with minor modifications meet safeguard requirements. URANUS is an integrated man/computer system with manual input and automatic data treatment and reporting of information. The man/computer interface is monitored by the automatic checking for plausibility of all input. An item is a quantity of material which is treated as a unit in production. Each item receives a unique identification number which may be used only once. If the qualitative properties of an item are changed a new item number is issued. Items are reported on forms and fed into URANUS by authorized personnel using personal passwords or are directly reported by terminal from certain operations. Pertinent information is entered for each project. The URANUS computer system consists of terminal activated on-line and batch modules which are briefly described. Output is information for planning, economic control and safeguards. Safeguards information consists of the following. For flow and blending control: delivery information; change of nominal enrichment; enrichment blending; internal transactions that change an enrichment/origin account; material status report (MSR); item inventory per MBA; general ledger. For physical inventory: material balance report (MBR); book item inventory per MBA; physical item inventory per MBA; uranium balance per origin and enrichment 13. Production and manipulation of wave packets from ultracold atoms in an optical lattice DEFF Research Database (Denmark) Pedersen, Poul Lindholm; Gajdacz, Miroslav; Winter, Nils 2013-01-01 of the system. The modulation technique also allows for a controllable transfer (deexcitation) of atoms from such wave packets to a state bound by the lattice. Thus, it acts as a beam splitter for matter waves that can selectively address different bands, enabling the preparation of atoms in localized states... 14. York University atomic scientist contributes to new breakthrough in the production of antimatter CERN Multimedia 2002-01-01 Physicists working in Europe, including Canada Research Chair in Atomic Physics at York University, Prof. Eric Hessels, have succeeded in capturing the first glimpse of the structure of antimatter. The ATRAP group of scientists at CERN have managed to examine the internal states of anti-hydrogen atoms (1/2 page). 15. Potential radiation doses from 1994 Hanford Operations Energy Technology Data Exchange (ETDEWEB) Soldat, J.K.; Antonio, E.J. 1995-06-01 This section of the 1994 Hanford Site Environmental Report summarizes the potential radiation doses to the public from releases originating at the Hanford Site. Members of the public are potentially exposed to low-levels of radiation from these effluents through a variety of pathways. The potential radiation doses to the public were calculated for the hypothetical MEI and for the general public residing within 80 km (50 mi) of the Hanford Site. 16. Potential radiation doses from 1994 Hanford Operations International Nuclear Information System (INIS) Soldat, J.K.; Antonio, E.J. 1995-01-01 This section of the 1994 Hanford Site Environmental Report summarizes the potential radiation doses to the public from releases originating at the Hanford Site. Members of the public are potentially exposed to low-levels of radiation from these effluents through a variety of pathways. The potential radiation doses to the public were calculated for the hypothetical MEI and for the general public residing within 80 km (50 mi) of the Hanford Site 17. Degradation of dome cutting minerals in Hanford waste-13100 International Nuclear Information System (INIS) Reynolds, Jacob G.; Huber, Heinz J.; Cooke, Gary A. 2013-01-01 At the Hanford Tank Farms, recent changes in retrieval technology require cutting new risers in several single-shell tanks. The Hanford Tank Farm Operator is using water jet technology with abrasive silicate minerals such as garnet or olivine to cut through the concrete and rebar dome. The abrasiveness of these minerals, which become part of the high-level waste stream, may enhance the erosion of waste processing equipment. However, garnet and olivine are not thermodynamically stable in Hanford waste, slowly degrading over time. How likely these materials are to dissolve completely in the waste before the waste is processed in the Waste Treatment and Immobilization Plant can be evaluated using theoretical analysis for olivine and collected direct experimental evidence for garnet. Based on an extensive literature study, a large number of primary silicates decompose into sodalite and cancrinite when exposed to Hanford waste. Given sufficient time, the sodalite also degrades into cancrinite. Even though cancrinite has not been directly added to any Hanford tanks during process times, it is the most common silicate observed in current Hanford waste. By analogy, olivine and garnet are expected to ultimately also decompose into cancrinite. Garnet used in a concrete cutting demonstration was immersed in a simulated supernate representing the estimated composition of the liquid retrieving waste from Hanford tank 241-C-107 at both ambient and elevated temperatures. This simulant was amended with extra NaOH to determine if adding caustic would help enhance the degradation rate of garnet. The results showed that the garnet degradation rate was highest at the highest NaOH concentration and temperature. At the end of 12 weeks, however, the garnet grains were mostly intact, even when immersed in 2 molar NaOH at 80 deg C. Cancrinite was identified as the degradation product on the surface of the garnet grains. In the case of olivine, the rate of degradation in the high-pH regimes 18. Degradation of Dome Cutting Minerals in Hanford Waste - 13100 Energy Technology Data Exchange (ETDEWEB) Reynolds, Jacob G.; Cooke, Gary A.; Huber, Heinz J. [Washington River Protection Solutions, LLC, P.O. Box 850, Richland, WA 99352 (United States) 2013-07-01 At the Hanford Tank Farms, recent changes in retrieval technology require cutting new risers in several single-shell tanks. The Hanford Tank Farm Operator is using water jet technology with abrasive silicate minerals such as garnet or olivine to cut through the concrete and rebar dome. The abrasiveness of these minerals, which become part of the high-level waste stream, may enhance the erosion of waste processing equipment. However, garnet and olivine are not thermodynamically stable in Hanford waste, slowly degrading over time. How likely these materials are to dissolve completely in the waste before the waste is processed in the Waste Treatment and Immobilization Plant can be evaluated using theoretical analysis for olivine and collected direct experimental evidence for garnet. Based on an extensive literature study, a large number of primary silicates decompose into sodalite and cancrinite when exposed to Hanford waste. Given sufficient time, the sodalite also degrades into cancrinite. Even though cancrinite has not been directly added to any Hanford tanks during process times, it is the most common silicate observed in current Hanford waste. By analogy, olivine and garnet are expected to ultimately also decompose into cancrinite. Garnet used in a concrete cutting demonstration was immersed in a simulated supernate representing the estimated composition of the liquid retrieving waste from Hanford tank 241-C-107 at both ambient and elevated temperatures. This simulant was amended with extra NaOH to determine if adding caustic would help enhance the degradation rate of garnet. The results showed that the garnet degradation rate was highest at the highest NaOH concentration and temperature. At the end of 12 weeks, however, the garnet grains were mostly intact, even when immersed in 2 molar NaOH at 80 deg. C. Cancrinite was identified as the degradation product on the surface of the garnet grains. In the case of olivine, the rate of degradation in the high 19. DOE Hanford Network Upgrades and Disaster Recovery Exercise Support the Cleanup Mission Now and into the Future International Nuclear Information System (INIS) Eckman, Todd J.; Hertzel, Ali K.; Lane, James J. 2013-01-01 In 2013, the U.S. Department of Energy's (DOE) Hanford Site, located in Washington State, funded an update to the critical network infrastructure supporting the Hanford Federal Cloud (HFC). The project, called ET-50, was the final step in a plan that was initiated five years ago called 'Hanford's IT Vision, 2015 and Beyond.' The ET-50 project upgraded Hanford's core data center switches and routers along with a majority of the distribution layer switches. The upgrades allowed HFC the network intelligence to provide Hanford with a more reliable and resilient network architecture. The culmination of the five year plan improved network intelligence and high performance computing as well as helped to provide 10 Gbps capable links between core backbone devices (10 times the previous bandwidth). These improvements allow Hanford the ability to further support bandwidth intense applications, such as video teleconferencing. The ET-50 switch upgrade, along with other upgrades implemented from the five year plan, have prepared Hanford's network for the next evolution of technology in voice, video, and data. Hand-in-hand with ET-50's major data center outage, Mission Support Alliance's (MSA) Information Management (IM) organization executed a disaster recovery (DR) exercise to perform a true integration test and capability study. The DR scope was planned within the constraints of ET-50's 14 hour datacenter outage window. This DR exercise tested Hanford's Continuity of Operations (COOP) capability and failover plans for safety and business critical Hanford Federal Cloud applications. The planned suite of services to be tested was identified prior to the outage and plans were prepared to test the services ability to failover from the primary Hanford data center to the backup data center. The services tested were: Core Network (backbone, firewall, load balancers); Voicemail; Voice over IP (VoIP); Emergency Notification; Virtual desktops; and, Select set of production applications 20. Evaluation of the Potential for Agricultural Development at the Hanford Site Energy Technology Data Exchange (ETDEWEB) Evans, Robert G.; Hattendorf, Mary J.; Kincaid, Charles T. 2000-02-25 By 2050, when cleanup of the Hanford Site is expected to be completed, large worldwide demands to increase the global production of animal and fish protein, food, and fiber are anticipated, despite advancements in crop breeding, genetic engineering, and other technologies. The most likely large areas for expanded irrigation in the Pacific Northwest are the undeveloped East High areas of the Columbia Basin Project and non-restricted areas within the Hanford Site in south-central Washington State. The area known as the Hanford Site has all the components that favor successful irrigated farming. Constraints to agricultural development of the Hanford Site are political and social, not economic or technical. Obtaining adequate water rights for any irrigated development will be a major issue. Numerous anticipated future advances in irrigation and resource conservation techniques such as precision agriculture techniques, improved irrigation systems, and irrigation system controls will greatly minimize the negative environmental impacts of agricultural activities. 1. Hanford Site Solid Waste Acceptance Criteria Energy Technology Data Exchange (ETDEWEB) 1993-11-17 This manual defines the Hanford Site radioactive, hazardous, and sanitary solid waste acceptance criteria. Criteria in the manual represent a guide for meeting state and federal regulations; DOE Orders; Hanford Site requirements; and other rules, regulations, guidelines, and standards as they apply to acceptance of radioactive and hazardous solid waste at the Hanford Site. It is not the intent of this manual to be all inclusive of the regulations; rather, it is intended that the manual provide the waste generator with only the requirements that waste must meet in order to be accepted at Hanford Site TSD facilities. 2. HANFORD SCIENCE & TECHNOLOGY NEEDS STATEMENTS 2002 Energy Technology Data Exchange (ETDEWEB) WIBLE, R.A. 2002-04-01 This document: (a) provides a comprehensive listing of the Hanford sites science and technology needs for fiscal year (FY) 2002; and (b) identifies partnering and commercialization opportunities within industry, other federal and state agencies, and the academic community. These needs were prepared by the Hanford projects (within the Project Hanford Management Contract, the Environmental Restoration Contract and the River Protection Project) and subsequently reviewed and endorsed by the Hanford Site Technology Coordination Group (STCG). The STCG reviews included participation of DOE-RL and DOE-ORP Management, site stakeholders, state and federal regulators, and Tribal Nations. These needs are reviewed and updated on an annual basis and given a broad distribution. 3. Field trip guide to the Hanford Site International Nuclear Information System (INIS) Reidel, S.P.; Lindsey, K.A.; Fecht, K.R. 1992-11-01 This report is designed to provide a guide to the key geologic and hydrologic features of the US Department of Energy's Hanford Site located in south-central Washington. The guide is divided into two parts. The first part is a general introduction to the geology of the Hanford Site and its relation to the regional framework of south-central Washington. The second part is a road log that provides directions to important geologic features on the Hanford Site and descriptions of the locality. The exposures described were chosen for their accessibility and importance to the geologic history of the Hanford Site and to understanding the geohydrology of the Site 4. Hanford Site Solid Waste Acceptance Criteria International Nuclear Information System (INIS) 1993-01-01 This manual defines the Hanford Site radioactive, hazardous, and sanitary solid waste acceptance criteria. Criteria in the manual represent a guide for meeting state and federal regulations; DOE Orders; Hanford Site requirements; and other rules, regulations, guidelines, and standards as they apply to acceptance of radioactive and hazardous solid waste at the Hanford Site. It is not the intent of this manual to be all inclusive of the regulations; rather, it is intended that the manual provide the waste generator with only the requirements that waste must meet in order to be accepted at Hanford Site TSD facilities 5. Mortality of Hanford radiation workers International Nuclear Information System (INIS) Gilbert, E.S. 1980-01-01 Mortality from all causes for white males employed at Hanford for at least two years is 75 percent of that expected on the basis of US vital statistics. Mortality from cancer is 85 percent of that expected. These results are typical of a working population. Neither death from all causes nor death from all cancer types shows a positive correlation with external radiation exposures. Myeloid leukemia, the disease that several studies have found to be associated most strongly with radiation exposure, is not correlated with external radiation exposure of Hanford workers. Two specific cancers, multiple myeloma and to a lesser extent cancer of the pancreas, were found to be positively correlated with radiation exposure. The correlations identified result entirely from a small number of deaths (3 each for multiple myeloma and cancer of the pancreas) with cumulative exposure greater than 15 rem 6. Hanford Environmental Dose Reconstruction Project International Nuclear Information System (INIS) McMakin, A.H.; Cannon, S.D.; Finch, S.M. 1992-07-01 The objective of the Hanford Environmental Dose Reconstruction (HEDR) Project is to estimate the radiation doses that individuals and populations could have received from nuclear operations at Hanford since 1944. The TSP consists of experts in environmental pathways, epidemiology, surface-water transport, ground-water transport, statistics, demography, agriculture, meteorology, nuclear engineering, radiation dosimetry, and cultural anthropology. Included are appointed technical members representing the states of Oregon, Washington, and Idaho, a representative of Native American tribes, and an individual representing the public. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed from release to impact on humans (dose estimates): Source terms, environmental transport, environmental monitoring data, demography, food consumption, and agriculture, and environmental pathways and dose estimates. Progress is discussed 7. Hanford whole body counting manual International Nuclear Information System (INIS) Palmer, H.E.; Brim, C.P.; Rieksts, G.A.; Rhoads, M.C. 1987-05-01 This document, a reprint of the Whole Body Counting Manual, was compiled to train personnel, document operation procedures, and outline quality assurance procedures. The current manual contains information on: the location, availability, and scope of services of Hanford's whole body counting facilities; the administrative aspect of the whole body counting operation; Hanford's whole body counting facilities; the step-by-step procedure involved in the different types of in vivo measurements; the detectors, preamplifiers and amplifiers, and spectroscopy equipment; the quality assurance aspect of equipment calibration and recordkeeping; data processing, record storage, results verification, report preparation, count summaries, and unit cost accounting; and the topics of minimum detectable amount and measurement accuracy and precision. 12 refs., 13 tabs 8. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 The Well subject area of the Hanford Environmental Information System (HEIS) manages data relevant to wells, boreholes and test pits constructed at the Hanford Site for soil sampling, geologic analysis and/or ground-water monitoring, and sampling for hydrochemical and radiological analysis. Data stored in the Well subject area include information relevant to the construction of the wells and boreholes, structural modifications to existing wells and boreholes, the location of wells, boreholes and test pits, and the association of wells, boreholes and test pits with organization entities such as waste sites. Data resulting from ground-water sampling performed at wells are stored in tables in the Ground-Water subject area. Geologic data collected during drilling, including particle sizing and interpretative geologic summaries, are stored in tables in the Geologic subject area. Data from soil samples taken during the drilling or excavation and sent for chemical and/or radiological analysis are stored in the Soil subject area 9. Atomic reactor thermal engineering International Nuclear Information System (INIS) Kim, Gwang Ryong 1983-02-01 This book starts the introduction of atomic reactor thermal engineering including atomic reaction, chemical reaction, nuclear reaction neutron energy and soon. It explains heat transfer, heat production in the atomic reactor, heat transfer of fuel element in atomic reactor, heat transfer and flow of cooler, thermal design of atomic reactor, design of thermodynamics of atomic reactor and various. This deals with the basic knowledge of thermal engineering for atomic reactor. 10. Hanford waste tank cone penetrometer International Nuclear Information System (INIS) Seda, R.Y. 1995-12-01 A new tool is being developed to characterize tank waste at the Hanford Reservation. This tool, known as the cone penetrometer, is capable of obtaining chemical and physical properties in situ. For the past 50 years, this tool has been used extensively in soil applications and now has been modified for usage in Hanford Underground Storage tanks. These modifications include development of new ''waste'' data models as well as hardware design changes to accommodate the hazardous and radioactive environment of the tanks. The modified cone penetrometer is scheduled to be deployed at Hanford by Fall 1996. At Hanford, the cone penetrometer will be used as an instrumented pipe which measures chemical and physical properties as it pushes through tank waste. Physical data, such as tank waste stratification and mechanical properties, is obtained through three sensors measuring tip pressure, sleeve friction and pore pressure. Chemical data, such as chemical speciation, is measured using a Raman spectroscopy sensor. The sensor package contains other instrumentation as well, including a tip and side temperature sensor, tank bottom detection and an inclinometer. Once the cone penetrometer has reached the bottom of the tank, a moisture probe will be inserted into the pipe. This probe is used to measure waste moisture content, water level, waste surface moisture and tank temperature. This paper discusses the development of this new measurement system. Data from the cone penetrometer will aid in the selection of sampling tools, waste tank retrieval process, and addressing various tank safety issues. This paper will explore various waste models as well as the challenges associated with tank environment 11. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 The purpose of the Biota subject area of the Hanford Environmental Information System (HEIS) is to manage the data collected from samples of plants and animals. This includes both samples taken from the plant or animal or samples related to the plant or animal. Related samples include animal feces and animal habitat. Data stored in the Biota subject area include data about the biota samples taken, analysis results counts from population studies, and species distribution maps 12. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 The purpose of the Soil subject area of the Hanford Environmental Information System (HEIS) is to manage the data acquired from soil samples, both geologic and surface, and sediment samples. Stored in the Soil subject area are data relevant to the soil samples, laboratory analytical results, and field measurements. The two major types of data make up the Soil subject area are data concerning the samples and data about the chemical and/or radiologic analyses of soil samples 13. Hanford Generic Interim Safety Basis Energy Technology Data Exchange (ETDEWEB) Lavender, J.C. 1994-09-09 The purpose of this document is to identify WHC programs and requirements that are an integral part of the authorization basis for nuclear facilities that are generic to all WHC-managed facilities. The purpose of these programs is to implement the DOE Orders, as WHC becomes contractually obligated to implement them. The Hanford Generic ISB focuses on the institutional controls and safety requirements identified in DOE Order 5480.23, Nuclear Safety Analysis Reports. 14. Hanford Generic Interim Safety Basis International Nuclear Information System (INIS) Lavender, J.C. 1994-01-01 The purpose of this document is to identify WHC programs and requirements that are an integral part of the authorization basis for nuclear facilities that are generic to all WHC-managed facilities. The purpose of these programs is to implement the DOE Orders, as WHC becomes contractually obligated to implement them. The Hanford Generic ISB focuses on the institutional controls and safety requirements identified in DOE Order 5480.23, Nuclear Safety Analysis Reports 15. Utilization of an arc-heated jet for production of supersonic seeded beams of atomic nitrogen International Nuclear Information System (INIS) Bickes, R.W. Jr.; Newton, K.R.; Herrmann, J.M.; Bernstein, R.B. 1976-01-01 Intense supersonic beams of atomic nitrogen (>10 17 atoms sr -1 sec -1 ) have been produced from the dissociation of N 2 in an Ar arc (at temperatures in excess of 6000 K) using the arc-heated nozzle beam source of Young, Rodgers, and Knuth. Experiments characterizing the N 2 dissociation and the translational energies of the N, N 2 , and Ar components in the beams are described. Evidence is presented for the formation of atomic C as well as C 2 and CH from the pyrolysis of CH 4 and C 2 H 4 in the Ar arc 16. Hanford science and technology needs statements, 2000 Energy Technology Data Exchange (ETDEWEB) BERLIN, G.T. 1999-07-16 In the aftermath of the Cold War, the United States has begun addressing the environmental consequences of five decades of nuclear weapons production. In November 1989, DOE established the Office of Environmental Restoration and Waste Management (EM) as the central authority for cleaning up the DOE weapons complex legacy of pollution, for preventing further environmental contamination, and for instituting responsible environmental management. While performing its tasks, EM found that many aspects of its large and complex mission could not be achieved using existing science and technology or without incurring unreasonable costs, risks, or schedule impacts. Consequently, a process was developed to solicit needs from around the DOE complex and focus the science and technology resources of EM-50, the National Laboratories, private industry, and colleges and universities on those needs. This document describes those needs that the Hanford Site has identified as requiring additional science or technology to complete. 17. Hanford science and technology needs statements, 1999 Energy Technology Data Exchange (ETDEWEB) Berlin, G.T. 1998-09-30 In the aftermath of the Cold War, the US has begun addressing the environmental consequences of five decades of nuclear weapons production. In November 1989, DOE established the Office of Environmental Restoration and Waste Management (EM) as the central authority for cleaning up the DOE weapons complex legacy of pollution, for preventing further environmental contamination, and for instituting responsible environmental management. While performing its tasks, EM found that many aspects of its large and complex decisions could not be achieved using existing science and technology or without incurring unreasonable costs, risks, or schedule impacts. Consequently, a process was developed to solicit needs from around the DOE complex and focus the science and technology resources of EM-50, the National Laboratories, private industry, and collages and universities on those needs. This document describes those needs which the Hanford Site has identified as requiring additional science or technology to complete. 18. Hanford B Reactor Building Hazard Assessment Report International Nuclear Information System (INIS) Griffin, P. W. 1999-01-01 The 105-B Reactor (hereinafter referred to as B Reactor) is located in the 100 Area of the Hanford Site near Richland, Washington. The B Reactor is one of nine plutonium production reactors that were constructed in the 1940s during the Cold War Era. Construction of the B Reactor began June 7, 1943, and operation began on September 26, 1944. The Environmental Restoration Contractor was requested by RL to provide an assessment/characterization of the B Reactor building to determine and document the hazards that are present and could pose a threat to the environment and/or to individuals touring the building. This report documents the potential hazards, determines the feasibility of mitigating the hazards, and makes recommendations regarding areas where public tour access should not be permitted 19. Chemical pathways for the formation of ammonia in Hanford wastes International Nuclear Information System (INIS) Stock, L.M.; Pederson, L.R. 1997-12-01 This report reviews chemical reactions leading to the formation of ammonia in Hanford wastes. The general features of the chemistry of the organic compounds in the Hanford wastes are briefly outlined. The radiolytic and thermal free radical reactions that are responsible for the initiation and propagation of the oxidative degradation reactions of the nitrogen-containing complexants, trisodium HEDTA and tetrasodium EDTA, are outlined. In addition, the roles played by three different ionic reaction pathways for the oxidation of the same compounds and their degradation products are described as a prelude to the discussion of the formation of ammonia. The reaction pathways postulated for its formation are based on tank observations, laboratory studies with simulated and actual wastes, and the review of the scientific literature. Ammonia derives from the reduction of nitrite ion (most important), from the conversion of organic nitrogen in the complexants and their degradation products, and from radiolytic reactions of nitrous oxide and nitrogen (least important) 20. Chemical pathways for the formation of ammonia in Hanford wastes Energy Technology Data Exchange (ETDEWEB) Stock, L.M.; Pederson, L.R. 1997-12-01 This report reviews chemical reactions leading to the formation of ammonia in Hanford wastes. The general features of the chemistry of the organic compounds in the Hanford wastes are briefly outlined. The radiolytic and thermal free radical reactions that are responsible for the initiation and propagation of the oxidative degradation reactions of the nitrogen-containing complexants, trisodium HEDTA and tetrasodium EDTA, are outlined. In addition, the roles played by three different ionic reaction pathways for the oxidation of the same compounds and their degradation products are described as a prelude to the discussion of the formation of ammonia. The reaction pathways postulated for its formation are based on tank observations, laboratory studies with simulated and actual wastes, and the review of the scientific literature. Ammonia derives from the reduction of nitrite ion (most important), from the conversion of organic nitrogen in the complexants and their degradation products, and from radiolytic reactions of nitrous oxide and nitrogen (least important). 1. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 The Hanford Environmental Information System (HEIS) is a consolidated set of automated resources that effectively manage the data gathered during environmental monitoring and restoration of the Hanford Site. HEIS includes an integrated database that provides consistent and current data to all users and promotes sharing of data by the entire user community. HEIS is an information system with an inclusive database. Although the database is the nucleus of the system, HEIS also provides user access software: query-by-form data entry, extraction, and browsing facilities; menu-driven reporting facilities; an ad hoc query facility; and a geographic information system (GIS). These features, with the exception of the GIS, are described in this manual set. Because HEIS contains data from the entire Hanford Site, many varieties of data are included and have.been divided into subject areas. Related subject areas comprise several volumes of the manual set. The manual set includes a data dictionary that lists all of the fields in the HEIS database, with their definitions and a cross reference of their locations in the database; definitions of data qualifiers for analytical results; and a mapping between the HEIS software functions and the keyboard keys for each of the supported terminals or terminal emulators 2. Hanford whole body counting manual International Nuclear Information System (INIS) Palmer, H.E.; Rieksts, G.A.; Lynch, T.P. 1990-06-01 This document describes the Hanford Whole Body Counting Program as it is administered by Pacific Northwest Laboratory (PNL) in support of the US Department of Energy--Richland Operations Office (DOE-RL) and its Hanford contractors. Program services include providing in vivo measurements of internally deposited radioactivity in Hanford employees (or visitors). Specific chapters of this manual deal with the following subjects: program operational charter, authority, administration, and practices, including interpreting applicable DOE Orders, regulations, and guidance into criteria for in vivo measurement frequency, etc., for the plant-wide whole body counting services; state-of-the-art facilities and equipment used to provide the best in vivo measurement results possible for the approximately 11,000 measurements made annually; procedures for performing the various in vivo measurements at the Whole Body Counter (WBC) and related facilities including whole body counts; operation and maintenance of counting equipment, quality assurance provisions of the program, WBC data processing functions, statistical aspects of in vivo measurements, and whole body counting records and associated guidance documents. 16 refs., 48 figs., 22 tabs 3. Atlantic Richfield Hanford Company chemical operator training program International Nuclear Information System (INIS) Zumhoff, R.G. 1975-01-01 Formal training and testing of Chemical Operators at Hanford were initiated as part of a negotiated union settlement in 1966. Consequently, it was agreed that 25 percent of the chemical operator force would receive a higher rated job (Lead Nuclear Chemical Operator) provided they satisfactorily completed a training program including testing. The training and testing program was developed in two parts. The first covered subjects of a general nature and was applicable to an operator's duties no matter what the assignment. Part II was more specifically oriented to the presently assigned work area. Renewed interest in retraining and requalification of all chemical operators was taken in 1971. This evolved from a Company concern that a program be developed to assure the fact that operators were qualified to do their assigned jobs, and an Atomic Energy Commission request for an outline of a retraining and requalification program for chemical operators. Building upon the experience gained in the LNCO (Lead Nuclear Chemical Operator) program, the two part format is retained. The use of video tapes is used to complement the manuals. An arrangement where an operator can view a lecture-type presentation is provided in seven plant locations. A small studio for in-house production of the video tapes is available to the training Specialists. A script is developed from a training manual by condensing the information into 20-minute presentations. A prime objective of each tape is to highlight the safety and control aspects that accompany operator responsibilities in each of these areas. Testing is also handled on a two part basis; one test covers the fundamentals and a separate test is designed for each of the plant subjects. A walk-through examination is also performed for the plant portion. Operators are required to be requalified on emergency procedures on an annual basis and at two-year intervals in the other areas. (U.S.) 4. Draft interim close-out report Fitzner-Eberhardt Arid Lands Ecology Reserve remedial action, Hanford, Washington. Draft A International Nuclear Information System (INIS) 1994-10-01 The Fitzner-Eberhardt Arid Lands Ecology (ALE) Reserve consists of 120 square miles of shrub-steppe land on the western edge of the Hanford Site. It is located south of Highway 240 and east of the point where the Yakima River borders the site. The land was set aside as a natural research area in 1967 by the Atomic Energy Commission. With the recent change in mission at Hanford from plutonium production to environmental cleanup, much attention has been given to releasing clean tracts of land for other uses. The ALE Reserve is one such tract of land. Consistent with the Record of Decision (ROD) and Preliminary Assessment Screening (PAS), sites identified as having the potential for contamination on the ALE Reserve have been investigated, characterized, and remediated where necessary to comply with MTCA cleanup levels. Hazardous substances released to the environment have either been removed from the site or, in one case when removal was not practical, capped in place. Removal of all of the contaminated soil is supported by the confirmatory sampling results. The existing areas of contamination in the ALE Reserve were considered to be small. In March of 1993, Department of Energy (DOE), the Environmental Protection Agency and the Washington State Department of Ecology signed an Agreement in Principle in which they agreed to expedite cleanup of the ALE Reserve. Cleanup activities and a draft closeout report were to be completed by October 1994. Additionally, DOE proposed to mitigate hazards which may pose a physical threat to wildlife or humans 5. Production and detection of cold anti-hydrogen atoms A first step towards high precision CPT test CERN Document Server Variola, A; Bonomi, G; Boutcha, A; Bowe, P; Carraro, C; Cesar, C L; Charlton, M; Doser, Michael; Filippini, V; Fontana, A; Fujiwara, M C; Funakoshi, R; Genova, P; Hangst, J S; Hayano, R S; Jørgensen, L V; Lagomarsino, V; Landua, Rolf; Lindelöf, D; Lodi-Rizzini, E; Macri, M; Madsen, N; Manuzio, G; Montagna, P; Pruys, H S; Regenfus, C; Rotondi, A; Riedler, P; Testera, G; Van der Werf, D P 2003-01-01 Observations of anti-hydrogen in small quantities have been reported at CERN and at FermiLab, but these experiments were not suited to spectroscopy experiments. In 2002 the ATHENA collaboration reported the production and detection of very low energy anti-hydrogen atoms produced in cryogenic environment. This is the first major step in the study of antiatom's internal structure and it can lead to a high precision test of the CPT fundamental symmetry. The method of production and detection of cold anti-hydrogen will be introduced. The absolute rate of anti-hydrogen production and the signal to background ratio in the ATHENA experiment will be discussed. (7 refs) . 6. Hot oxygen atoms: Their generation and chemistry. [Production by sputtering; reaction with butenes Energy Technology Data Exchange (ETDEWEB) Ferrieri, R.A.; Chu, Yung Y.; Wolf, A.P. 1987-01-01 Oxygen atoms with energies between 1 and 10 eV have been produced through ion beam sputtering from metal oxide targets. Argon ion beams were used on Ta/sub 2/O/sub 5/ and V/sub 2/O/sub 5/. Results show that some control may be exerted over the atom's kinetic energy by changing the target. Reactions of the hot O(/sup 3/P) with cis- and trans-butenes were investigated. (DLC) 7. Hanford Waste Vitrification Plant Quality Assurance Program description for high-level waste form development and qualification. Revision 3, Part 2 Energy Technology Data Exchange (ETDEWEB) 1993-08-01 The Hanford Waste Vitrification Plant Project has been established to convert the high-level radioactive waste associated with nuclear defense production at the Hanford Site into a waste form suitable for disposal in a deep geologic repository. The Hanford Waste Vitrification Plant will mix processed radioactive waste with borosilicate material, then heat the mixture to its melting point (vitrification) to forin a glass-like substance that traps the radionuclides in the glass matrix upon cooling. The Hanford Waste Vitrification Plant Quality Assurance Program has been established to support the mission of the Hanford Waste Vitrification Plant. This Quality Assurance Program Description has been written to document the Hanford Waste Vitrification Plant Quality Assurance Program. 8. Deep Atomic Binding (DAB) Hypothesis: A New Approach of Fission Product Chemistry International Nuclear Information System (INIS) Ajlouni, Abdul-Wali M.S. 2006-01-01 Former studies assumed that, after fission process occurs, the highly ionized new born atoms (20-22 positive charge), ionize the media in which they pass through before becoming stable atoms in a manner similar to 4-MeV ?-particles. Via ordinary chemical reactions with the surroundings, each stable atom has a probability to form chemical compound. Since there are about 35 different elemental atoms created through fission processes, a large number of chemical species were suggested to be formed. But, these suggested chemical species were not found in the environment after actual releases of FP during accidents like TMI (USA, 1979), and Chernobyl (former USSR, 1986), also the models based on these suggested reactions and species could not interpret the behavior of these actual species. It is assumed here that the ionization states of the new born atoms and the long term high temperature were not dealt with in an appropriate way and they were the reasons of former models failure. Our new approach of Deep Atomic Binding (DAB) based on the following: 1-The new born atoms which are highly ionized, 10-12 electrons associated with each nucleus, having a large probability to create bonds between them to form molecules. These bonds are at the L, or M shells, and we call it DAB. 2-The molecules stay in the reactor at high temperatures for long periods, so they undergo many stages of composition and decomposition to form giant molecules. By applying DAB approach, field data from Chernobyl, TMI and nuclear detonations could be interpreted with a wide coincidence resulted. (author) 9. Chemical composition analysis and product consistency tests supporting refinement of the Nepheline Model for the high aluminum Hanford glass composition region Energy Technology Data Exchange (ETDEWEB) Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Mcclane, D. L. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL) 2016-03-01 In this report, Savannah River National Laboratory provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated high level waste (HLW) glasses fabricated by Pacific Northwest National Laboratory (PNNL) as part of an ongoing nepheline crystallization study. The results of these analyses will be used to improve the ability to predict crystallization of nepheline as a function of composition and heat treatment for glasses formulated at high alumina concentrations. 10. Chemical composition analysis and product consistency tests supporting refinement of the Nepheline model for the high aluminum Hanford Glass composition region Energy Technology Data Exchange (ETDEWEB) Fox, K. M. [Savannah River Site (SRS), Aiken, SC (United States); Edwards, T. B. [Savannah River Site (SRS), Aiken, SC (United States); Mcclane, D. L. [Savannah River Site (SRS), Aiken, SC (United States) 2016-02-17 In this report, SRNL provides chemical analyses and Product Consistency Test (PCT) results for a series of simulated HLW glasses fabricated by Pacific Northwest National Laboratory (PNNL) as part of an ongoing nepheline crystallization study. The results of these analyses will be used to improve the ability to predict crystallization of nepheline as a function of composition and heat treatment for glasses formulated at high alumina concentrations. 11. Mechanism of pulse discharge production of iodine atoms from CF3I molecules for a chemical oxygen-iodine laser International Nuclear Information System (INIS) Kochetov, I V; Napartovich, A P; Vagin, N P; Yuryshev, N N 2009-01-01 The pulsed chemical oxygen-iodine laser (COIL) development is aimed at many new applications. Pulsed electric discharge is most effective in turning COIL operation into the pulse mode by instant production of iodine atoms. A numerical model is developed for simulations of the pulsed COIL initiated by an electric discharge. The model comprises a system of kinetic equations for neutral and charged species, electric circuit equation, gas thermal balance equation and the photon balance equation. Reaction rate coefficients for processes involving electrons are found by solving the electron Boltzmann equation, which is re-calculated in a course of computations when plasma parameters changed. The processes accounted for in the Boltzmann equation include excitation and ionization of atoms and molecules, dissociation of molecules, electron attachment processes, electron-ion recombination, electron-electron collisions, second-kind collisions and stepwise excitation of molecules. The last processes are particularly important because of a high singlet oxygen concentration in gas flow from the singlet oxygen chemical generator. Results of numerical simulations are compared with experimental laser pulse waveforms. It is concluded that there is satisfactory agreement between theory and the experiment. The prevailing mechanism of iodine atom formation from the CF 3 I donor in a very complex kinetic system of the COIL medium under pulse discharge conditions, based on their detailed numerical modelling and by comparing these results both with experimental results of other authors and their own experiments, is established. The dominant iodine atom production mechanism for conditions under study is the electron-impact dissociation of CF 3 I molecules. It was proved that in the conditions of the experiment the secondary chemical reactions with O atoms play an insignificant role. 12. Production of hollow atoms by high brightness x-ray sources and its applications International Nuclear Information System (INIS) Moribayashi, Kengo 2004-01-01 We study x-ray emissions from the (multi-)inner-shell states and hollow atoms of Si ions excited by high intensity x-ray sources. It is found that the x-ray number from multi-inner-shell excited states (1s 2 2s 2 2p k 3s 2 3p 2 , k=1-4) and hollow atoms (1s 2 2s 2 3p 2 ) is affected greatly by the high intensity short-pulse x-rays and little by weak intensity post-long pulse x-rays. The ratio of the x-ray intensities from hollow atoms to those from the multi-inner-shell excited states becomes almost independent of the pulses and dependent on the intensities of x-ray sources. This ratio may be used for the measurement of intensities of high intensity short pulse x-ray sources. (author) 13. Vascular Plants of the Hanford Site International Nuclear Information System (INIS) Sackschewsky, Michael R.; Downs, Janelle L. 2001-01-01 This report provides an updated listing of the vascular plants present on and near the U.S. Department of Energy Hanford Site. This document is an update of a listing of plants prepared by Sackschewdky et al. in 1992. Since that time there has been a significant increase in the botanical knowledge of the Hanford Site. The present listing is based on an examination of herbarium collections held at PNNL, at WSU-Tri Cities, WSU-Pullman, Brigham Young University, and The University of Washington, and on examination of ecological literature derived from the Hanford and Benton county areas over the last 100 years. Based on the most recent analysis, there are approximately 725 different plant species that have been documented on or around the Hanford Site. This represents an approximate 20% increase in the number of species reported within Sackschewsky et al. (1992). This listing directly supports DOE and contractor efforts to assess the potential impacts of Hanford Site operations 14. Hanford Patrol Academy demolition sites closure plan Energy Technology Data Exchange (ETDEWEB) 1993-09-30 The Hanford Site is owned by the U.S. Government and operated by the U.S. Department of Energy, Richland Operations Office. Westinghouse Hanford Company is a major contractor to the U.S. Department of Energy, Richland Operations Office and serves as co-operator of the Hanford Patrol Academy Demolition Sites, the unit addressed in this paper. This document consists of a Hanford Facility Dangerous Waste Part A Permit Application, Form 3 (Revision 4), and a closure plan for the site. An explanation of the Part A Form 3 submitted with this closure plan is provided at the beginning of the Part A section. This Hanford Patrol Academy Demolition Sites Closure Plan submittal contains information current as of December 15, 1994. 15. Hanford Site baseline risk assessment methodology International Nuclear Information System (INIS) 1993-03-01 This methodology has been developed to prepare human health and environmental evaluations of risk as part of the Comprehensive Environmental Response, Compensation, and Liability Act remedial investigations (RIs) and the Resource Conservation and Recovery Act facility investigations (FIs) performed at the Hanford Site pursuant to the Hanford Federal Facility Agreement and Consent Order referred to as the Tri-Party Agreement. Development of the methodology has been undertaken so that Hanford Site risk assessments are consistent with current regulations and guidance, while providing direction on flexible, ambiguous, or undefined aspects of the guidance. The methodology identifies Site-specific risk assessment considerations and integrates them with approaches for evaluating human and environmental risk that can be factored into the risk assessment program supporting the Hanford Site cleanup mission. Consequently, the methodology will enhance the preparation and review of individual risk assessments at the Hanford Site 16. Hanford Environmental Management Program implementation plan International Nuclear Information System (INIS) 1988-08-01 The Hanford Environmental Management Program (HEMP) was established to facilitate compliance with the applicable environmental statues, regulations, and standards on the Hanford Site. The HEMP provides a structured approach to achieve environmental management objectives. The Hanford Environmental Management Program Plan (HEMP Plan) was prepared as a strategic level planning document to describe the program management, technical implementation, verification, and communications activities that guide the HEMP. Four basic program objectives are identified in the HEMP Plan as follows: establish ongoing monitoring to ensure that Hanford Site operations comply with environmental requirements; attain regulatory compliance through the modification of activities; mitigate any environmental consequences; and minimize the environmental impacts of future operations at the Hanford Site. 2 refs., 24 figs., 27 tabs 17. Hanford Site Risk Assessment Methodology. Revision 3 International Nuclear Information System (INIS) 1995-05-01 This methodology has been developed to prepare human health and ecological evaluations of risk as part of the Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA) remedial investigations (RI) and the Resource conservation and Recovery Act of 1976 (RCRA) facility investigations (FI) performed at the Hanford Site pursuant to the hanford Federal Facility Agreement and Consent Order (Ecology et al. 1994), referred to as the Tri-Party Agreement. Development of the methodology has been undertaken so that Hanford Site risk assessments are consistent with current regulations and guidance, while providing direction on flexible, ambiguous, or undefined aspects of the guidance. The methodology identifies site-specific risk assessment considerations and integrates them with approaches for evaluating human and ecological risk that can be factored into the risk assessment program supporting the Hanford Site cleanup mission. Consequently, the methodology will enhance the preparation and review of individual risk assessments at the Hanford Site 18. Monte Carlo simulation study of the positron contribution to displacements per atom production in YBCO superconductors International Nuclear Information System (INIS) Pinera, I.; Cruz, C.M.; Abreu, Y.; Leyva, A. 2008-01-01 The contribution from positrons to the displacements per atom (dpa) distribution induced by the gamma irradiation on YBCO superconducting slabs is presented. The procedure implemented previously by the authors was adapted to take into account the contribution from positrons to dpa induced by the gamma radiation. The results show that, when positrons are considered in the atom displacement process, the total dpa almost doubles at 10 MeV of incident gamma radiation. At that energy positrons contribute 7% more to the total dpa than electrons, although electrons maintain having the highest contribution up to about 8 MeV. 19. Pretreatment of Hanford PUREX Plant first-cycle waste International Nuclear Information System (INIS) Gibson, M.W.; Gerboth, D.M.; Peters, B.B. 1987-04-01 A process has been developed to pretreat neutralized, first-cycle high-level waste from the fuels reprocessing facility at the Hanford Site. The process separates solids from the supernate liquid, which contains soluble salts. The solids, including most of the fission products and transuranic elements, may then be vitrified for disposal, while the low-level supernate stream may be processed into a less expensive grout waste form. The process also includes ion exchange treatment of the separated supernate stream to remove radiocesium. A flow sheet based on these operations was completed to support a planned demonstration of the process in the Hanford Site B Plant canyon facility. 5 refs., 2 figs., 5 tabs 20. Radioactivity of food products in the region of the ''Kozloduj'' atomic power station in the pre-exploitation period International Nuclear Information System (INIS) Jotov, M.; Petkov, T.; Zlatanova, R.; Boyadzhiev, A. 1976-01-01 The pre-exploitation status of the region of the ''Kozloduj'' atomic power stationand its torch zone, regarding the concentration of the biologically most dangerous artificial isotopes in the basic food products was determined. The tested foods were milk, meat, wheat, fish and grapes produced at the most important and basic production regions in the controled region. The radioactive isotope concentrations in the analyzed food produxts are reported. They will serve as a basis in assessing any eventual additional contamination of the region as a result of the work of the ''Kozloduj'' electric power station. (author) 1. Atomic Number Dependence of Hadron Production at Large Transverse Momentum in 300 GeV Proton--Nucleus Collisions Science.gov (United States) Cronin, J. W.; Frisch, H. J.; Shochet, M. J.; Boymond, J. P.; Mermod, R.; Piroue, P. A.; Sumner, R. L. 1974-07-15 In an experiment at the Fermi National Accelerator Laboratory we have compared the production of large transverse momentum hadrons from targets of W, Ti, and Be bombarded by 300 GeV protons. The hadron yields were measured at 90 degrees in the proton-nucleon c.m. system with a magnetic spectrometer equipped with 2 Cerenkov counters and a hadron calorimeter. The production cross-sections have a dependence on the atomic number A that grows with P{sub 1}, eventually leveling off proportional to A{sup 1.1}. 2. The Hanford Site: An anthology of early histories International Nuclear Information System (INIS) Gerber, M.S. 1993-10-01 This report discusses the following topics: Memories of War: Pearl Harbor and the Genesis of the Hanford Site; safety has always been promoted at the Hanford Site; women have an important place in Hanford Site history; the boom and bust cycle: A 50-year historical overview of the economic impacts of Hanford Site Operations on the Tri-Cities, Washington; Hanford's early reactors were crucial to the sites's history; T-Plant made chemical engineering history; the UO 3 plant has a long history of service. PUREX Plant: the Hanford Site's Historic Workhorse. PUREX Plant Waste Management was a complex challenge; and early Hanford Site codes and jargon 3. HANFORD SITE RIVER PROTECTION PROJECT (RPP) TANK FARM CLOSURE International Nuclear Information System (INIS) JARAYSI, M.N.; SMITH, Z.; QUINTERO, R.; BURANDT, M.B.; HEWITT, W. 2006-01-01 The U. S. Department of Energy, Office of River Protection and the CH2M HILL Hanford Group, Inc. are responsible for the operations, cleanup, and closure activities at the Hanford Tank Farms. There are 177 tanks overall in the tank farms, 149 single-shell tanks (see Figure 1), and 28 double-shell tanks (see Figure 2). The single-shell tanks were constructed 40 to 60 years ago and all have exceeded their design life. The single-shell tanks do not meet Resource Conservation and Recovery Act of 1976 [1] requirements. Accordingly, radioactive waste is being retrieved from the single-shell tanks and transferred to double-shell tanks for storage prior to treatment through vitrification and disposal. Following retrieval of as much waste as is technically possible from the single-shell tanks, the Office of River Protection plans to close the single-shell tanks in accordance with the Hanford Federal Facility Agreement and Consent Order [2] and the Atomic Energy Act of 1954 [3] requirements. The double-shell tanks will remain in operation through much of the cleanup mission until sufficient waste has been treated such that the Office of River Protection can commence closing the double-shell tanks. At the current time, however, the focus is on retrieving waste and closing the single-shell tanks. The single-shell tanks are being managed and will be closed in accordance with the pertinent requirements in: Resource Conservation and Recovery Act of 1976 and its Washington State-authorized Dangerous Waste Regulations [4], US DOE Order 435.1 Radioactive Waste Management [5], the National Environmental Policy Act of 1969 [6], and the Comprehensive Environmental Response, Compensation, and Liability Act of 1980 [7]. The Hanford Federal Facility Agreement and Consent Order, which is commonly referred to as the Tri-Party Agreement or TPA, was originally signed by Department of Energy, the State of Washington, and the U. S. Environmental Protection Agency in 1989. Meanwhile, the 4. Hanford performance evaluation program for Hanford site analytical services International Nuclear Information System (INIS) Markel, L.P. 1995-09-01 The U.S. Department of Energy (DOE) Order 5700.6C, Quality Assurance, and Title 10 of the Code of Federal Regulations, Part 830.120, Quality Assurance Requirements, states that it is the responsibility of DOE contractors to ensure that ''quality is achieved and maintained by those who have been assigned the responsibility for performing the work.'' Hanford Analytical Services Quality Assurance Plan (HASQAP) is designed to meet the needs of the Richland Operations Office (RL) for maintaining a consistent level of quality for the analytical chemistry services provided by contractor and commmercial analytical laboratory operations. Therefore, services supporting Hanford environmental monitoring, environmental restoration, and waste management analytical services shall meet appropriate quality standards. This performance evaluation program will monitor the quality standards of all analytical laboratories supporting the Hanforad Site including on-site and off-site laboratories. The monitoring and evaluation of laboratory performance can be completed by the use of several tools. This program will discuss the tools that will be utilized for laboratory performance evaluations. Revision 0 will primarily focus on presently available programs using readily available performance evaluation materials provided by DOE, EPA or commercial sources. Discussion of project specific PE materials and evaluations will be described in section 9.0 and Appendix A 5. Waste minimization -- Hanfords strategy for sustainability Energy Technology Data Exchange (ETDEWEB) Merry, D.S. 1998-01-30 The Hanford Site cleanup activity is an immense and challenging undertaking, which includes characterization and decommissioning of 149 single-shell storage tanks, treating waste stored in 28 double-shell tanks, safely disposing of over 2,100 metric tons of spent nuclear fuel stored onsite, removing thousands of structures, and dealing with significant solid waste, groundwater, and land restoration issues. The Pollution Prevention/Waste Minimization (P2/WMin) Program supports the Hanford Site mission to safely clean up and manage legacy waste and to develop and deploy science and technology in many ways. Once such way is through implementing and documenting over 231 waste reduction projects during the past five years, resulting in over $93 million in cost savings/avoidances. These savings/avoidances allowed other high priority cleanup work to be performed. Another way is by exceeding the Secretary of Energys waste reduction goals over two years ahead of schedule, thus reducing the amount of waste to be stored, treated and disposed. Six key elements are the foundation for these sustained P2/WMin results. 6. Understanding Molecular-Ion Neutral Atom Collisions for the Production of Ultracold Molecular Ions Science.gov (United States) 2014-02-03 SECURITY CLASSIFICATION OF: This project was superseded and replaced by another ARO-funded project of the same name, which is still continuing. The goal...cooled atoms," IOTA -COST Workshop on molecular ions, Arosa, Switzerland. 5. E.R. Hudson, "Sympathetic cooling of molecules with laser cooled 7. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 This report discusses the procedures that establish the configuration control processes for the Hanford Environmental Information System (HEIS) software. The procedures also provide the charter and function of the HEIS Configuration Control Board (CCB) for maintaining software. The software configuration control items covered under these procedures include the HEIS software and database structure. The configuration control processes include both administrative and audit functions. The administrative role includes maintaining the overall change schedule, ensuring consistency of proposed changes, negotiating change plan adjustments, setting priorities, and tracking the status of changes. The configuration control process audits to ensure that changes are performed to applicable standards 8. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) Schreck, R.I. 1994-01-01 The Hanford Environmental Information System (HEIS) Subject Area manuals are designed as reference guides, that is, each chapter provides the information needed to make best use of each subject area, its tables, and reporting capabilities. Each subject area is documented in a chapter in one of the subject area manuals. Because these are reference manuals, most of the information is also available in the online help system as well. See Section 5.4.2 of the HEIS User's Guide (DOE-RL 1994a) for a detailed description of the online help 9. Strontium-90 at the Hanford Site and its ecological implications International Nuclear Information System (INIS) RE Peterson; TM Poston 2000-01-01 Strontium-90, a radioactive contaminant from historical operations at the U.S. Department of Energy (DOE) Hanford Site, enters the Columbia River at several locations associated with former plutonium production reactors at the Site. Strontium-90 is of concern to humans and the environment because of its moderately long half-life (29.1 years), its potential for concentrating in bone tissue, and its relatively high energy of beta decay. Although strontium-90 in the environment is not a new issue for the Hanford Site, recent studies of near-river vegetation along the shoreline near the 100 Areas raised public concern about the possibility of strontium-90-contaminated groundwater reaching the riverbed and fall chinook salmon redds. To address these concerns, DOE asked Pacific Northwest National Laboratory (PNNL) to prepare this report on strontium-90, its distribution in groundwater, how and where it enters the river, and its potential ecological impacts, particularly with respect to fall chinook salmon. The purpose of the report is to characterize groundwater contaminants in the near-shore environment and to assess the potential for ecological impact using salmon embryos, one of the most sensitive ecological indicators for aquatic organisms. Section 2.0 of the report provides background information on strontium-90 at the Hanford Site related to historical operations. Public access to information on strontium-90 also is described. Section 3.0 focuses on key issues associated with strontium-90 contamination in groundwater that discharges in the Hanford Reach. The occurrence and distribution of fall chinook salmon redds in the Hanford Reach and characteristics of salmon spawning are described in Section 4.0. Section 5.0 describes the regulatory standards and criteria used to set action levels for strontium-90. Recommendations for initiating additional monitoring and remedial action associated with strontium-90 contamination at the Hanford Site are presented in Section 6 10. Strontium-90 at the Hanford Site and its ecological implications Energy Technology Data Exchange (ETDEWEB) RE Peterson; TM Poston 2000-05-22 Strontium-90, a radioactive contaminant from historical operations at the U.S. Department of Energy (DOE) Hanford Site, enters the Columbia River at several locations associated with former plutonium production reactors at the Site. Strontium-90 is of concern to humans and the environment because of its moderately long half-life (29.1 years), its potential for concentrating in bone tissue, and its relatively high energy of beta decay. Although strontium-90 in the environment is not a new issue for the Hanford Site, recent studies of near-river vegetation along the shoreline near the 100 Areas raised public concern about the possibility of strontium-90-contaminated groundwater reaching the riverbed and fall chinook salmon redds. To address these concerns, DOE asked Pacific Northwest National Laboratory (PNNL) to prepare this report on strontium-90, its distribution in groundwater, how and where it enters the river, and its potential ecological impacts, particularly with respect to fall chinook salmon. The purpose of the report is to characterize groundwater contaminants in the near-shore environment and to assess the potential for ecological impact using salmon embryos, one of the most sensitive ecological indicators for aquatic organisms. Section 2.0 of the report provides background information on strontium-90 at the Hanford Site related to historical operations. Public access to information on strontium-90 also is described. Section 3.0 focuses on key issues associated with strontium-90 contamination in groundwater that discharges in the Hanford Reach. The occurrence and distribution of fall chinook salmon redds in the Hanford Reach and characteristics of salmon spawning are described in Section 4.0. Section 5.0 describes the regulatory standards and criteria used to set action levels for strontium-90. Recommendations for initiating additional monitoring and remedial action associated with strontium-90 contamination at the Hanford Site are presented in Section 6 11. Hanford Waste Vitrification Plant hydrogen generation International Nuclear Information System (INIS) King, R.B.; King, A.D. Jr.; Bhattacharyya, N.K. 1996-02-01 The most promising method for the disposal of highly radioactive nuclear wastes is a vitrification process in which the wastes are incorporated into borosilicate glass logs, the logs are sealed into welded stainless steel canisters, and the canisters are buried in suitably protected burial sites for disposal. The purpose of the research supported by the Hanford Waste Vitrification Plant (HWVP) project of the Department of Energy through Battelle Pacific Northwest Laboratory (PNL) and summarized in this report was to gain a basic understanding of the hydrogen generation process and to predict the rate and amount of hydrogen generation during the treatment of HWVP feed simulants with formic acid. The objectives of the study were to determine the key feed components and process variables which enhance or inhibit the.production of hydrogen. Information on the kinetics and stoichiometry of relevant formic acid reactions were sought to provide a basis for viable mechanistic proposals. The chemical reactions were characterized through the production and consumption of the key gaseous products such as H 2 . CO 2 , N 2 0, NO, and NH 3 . For this mason this research program relied heavily on analyses of the gases produced and consumed during reactions of the HWVP feed simulants with formic acid under various conditions. Such analyses, used gas chromatographic equipment and expertise at the University of Georgia for the separation and determination of H 2 , CO, CO 2 , N 2 , N 2 O and NO 12. Hanford Site solid waste acceptance criteria International Nuclear Information System (INIS) Willis, N.P.; Triner, G.C. 1991-09-01 Westinghouse Hanford Company manages the Hanford Site solid waste treatment, storage, and disposal facilities for the US Department of Energy Field Office, Richland under contract DE-AC06-87RL10930. These facilities include radioactive solid waste disposal sites, radioactive solid waste storage areas and hazardous waste treatment, storage, and/or disposal facilities. This manual defines the criteria that must be met by waste generators for solid waste to be accepted by Westinghouse Hanford Company for treatment, storage and/or disposal facilities. It is to be used by all waste generators preparing radioactive solid waste for storage or disposal at the Hanford Site facilities and for all Hanford Site generators of hazardous waste. This manual is also intended for use by Westinghouse Hanford Company solid waste technical staff involved with approval and acceptance of solid waste. The criteria in this manual represent a compilation of state and federal regulations; US Department of Energy orders; Hanford Site requirements; and other rules, regulations, guidelines, and standards as they apply to management of solid waste. Where appropriate, these requirements are included in the manual by reference. It is the intent of this manual to provide guidance to the waste generator in meeting the applicable requirements 13. Hanford 200 Areas Development Plan Energy Technology Data Exchange (ETDEWEB) Rinne, C.A.; Daly, K.S. 1993-08-01 The purpose of the Hanford 200 Areas Development Plan (Development Plan) is to guide the physical development of the 200 Areas (which refers to the 200 East Area, 200 West Area, and 200 Area Corridor, located between the 200 East and 200 West Areas) in accordance with US Department of Energy (DOE) Order 4320.lB (DOE 1991a) by performing the following: Establishing a land-use plan and setting land-use categories that meet the needs of existing and proposed activities. Coordinating existing, 5-year, and long-range development plans and guiding growth in accordance with those plans. Establishing development guidelines to encourage cost-effective development and minimize conflicts between adjacent activities. Identifying site development issues that need further analysis. Integrating program plans with development plans to ensure a logical progression of development. Coordinate DOE plans with other agencies [(i.e., Washington State Department of Ecology (Ecology) and US Environmental Protection Agency (EPA)]. Being a support document to the Hanford Site Development Plan (DOE-RL 1990a) (parent document) and providing technical site information relative to the 200 Areas. 14. Cancer mortality in Hanford workers International Nuclear Information System (INIS) Marks, S.; Gilbert, E.S.; Breitenstein, B.D. 1978-01-01 Personnel and radiation exposure data for past and present employees of the Hanford plant have been collected and analysed for a possible relationship of exposure to mortality. The occurrence of death in workers was established by the Social Security Administration and the cause of death obtained from death certificates. Mortality from all causes, all cancer cases and specific cancer types was related to the population at risk. Standardized mortality ratios were calculated for white males, using age- and calendar year-specific mortality rates for the U.S. population in the calculation of expected deaths. This analysis showed a substantial 'healthy worker effect' and no significantly high standardized mortality ratios for specific disease categories. A test for association of mortality with levels of radiation exposure revealed no correlation for all causes and all cancer. In carrying out this test, adjustment was made for age and calendar year of death, length of employment and occupational category. A statistically significant test for trend was obtained for multiple myeloma and carcinoma of the pancreas. However, in view of the absence of such a correlation for diseases more commonly associated with radiation exposure such as myeloid leukaemia, as well as the small number of deaths in higher exposure groups, the results cannot be considered definitive. Any conclusions based on these associations should be viewed in relation to the results of other studies. These results are compared with those of other investigators who have analysed the Hanford data. (author) 15. Hanford transuranic storage corrosion review International Nuclear Information System (INIS) Nelson, J.L.; Divine, J.R. 1980-12-01 The rate of atmospheric corrosion of the transuranic (TRU) waste drums at the US Department of Energy's Hanford Project, near Richland, Washington, was evaluated by Pacific Northwest Laboratory (PNL). The rate of corrosion is principally contingent upon the effects of humidity, airborne pollutants, and temperature. Results of the study indicate that actual penetration of barrels due to atmospheric corrosion will probably not occur within the 20-year specified recovery period. Several other US burial sites were surveyed, and it appears that there is sufficient uncertainty in the available data to prevent a clearcut statement of the corrosion rate at a specific site. Laboratory and site tests are recommended before any definite conclusions can be made. The corrosion potential at the Hanford TRU waste site could be reduced by a combination of changes in drum materials (for example, using galvanized barrels instead of the currently used mild steel barrels), environmental exposure conditions (for example, covering the barrels in one of numerous possible ways), and storage conditions 16. Hanford 200 Areas Development Plan International Nuclear Information System (INIS) Rinne, C.A.; Daly, K.S. 1993-08-01 The purpose of the Hanford 200 Areas Development Plan (Development Plan) is to guide the physical development of the 200 Areas (which refers to the 200 East Area, 200 West Area, and 200 Area Corridor, located between the 200 East and 200 West Areas) in accordance with US Department of Energy (DOE) Order 4320.lB (DOE 1991a) by performing the following: Establishing a land-use plan and setting land-use categories that meet the needs of existing and proposed activities. Coordinating existing, 5-year, and long-range development plans and guiding growth in accordance with those plans. Establishing development guidelines to encourage cost-effective development and minimize conflicts between adjacent activities. Identifying site development issues that need further analysis. Integrating program plans with development plans to ensure a logical progression of development. Coordinate DOE plans with other agencies [(i.e., Washington State Department of Ecology (Ecology) and US Environmental Protection Agency (EPA)]. Being a support document to the Hanford Site Development Plan (DOE-RL 1990a) (parent document) and providing technical site information relative to the 200 Areas 17. Determination of metallic impurities in raw materials for radioisotope production by atomic absorption spectroscopy International Nuclear Information System (INIS) Roca, M.; Alvarez, F.; Capdevila, C. 1969-01-01 Atomic absorption spectrometry has been used for the determination of traces of calcium in scandium oxide, copper in zinc, iron in cobalt oxide, manganese In ferric oxide, nickel in copper and zinc in gallium oxide. The influences on the sensitivities arising from the hollow cathode currents, the gas pressures and the acid concentrations have been considered. A study of the interferences from the metallic matrices has also been performed, the interference due to the absorption of the manganese radiation by the atoms of iron being the most outstanding . In order to remove the interfering elements and increase sensitivity, pre-concentration methods have been tested. The addition methods has also been used. (Author) 14 refs 18. Production of atomic negative ion beams of the Group IA elements International Nuclear Information System (INIS) Alton, G.D.; Mills, G.D. 1988-01-01 A method has been developed which enables the direct sputter generation of atomic negative ion beams of all members of the Group IA elements (Li, Na, K, Rb, and Cs). The method consists of the use of sputter samples formed by pressing mixtures of the carbonates of the Group IA elements and 10% (atomic) Cu, Ag, or other metal powder. The following intensities are typical of those observed from carbonate samples subjected to /approximately/3 KeV cesium ion bombardment: Li - : ≥0.5 μA; Na - : ≥0.5 μA; K - : ≥0.5 μA; Rb - : ≥0.5 μA; Cs - : ≥0.2 μA. 7 refs., 2 figs., 1 tab 19. Understanding Molecular Ion-Neutral Atom Collisions for the Production of Ultracold Molecular Ions Science.gov (United States) 2016-06-06 2012): 0. doi: 10.1103/PhysRevLett.109.223002 Kuang Chen, Scott T. Sullivan, Wade G. Rellergert, Eric R. Hudson. Measurement of the Coulomb Logarithm...or fellowships for further studies in science, mathematics, engineering or technology fields: Student Metrics This section only applies to graduating...clouds of Ba+ ions and Ca atoms. Due to the strong Coulomb interaction, the Ba+ ions quickly cool the molecular ion translation motion, while the 20. An Integrated Biological Control System At Hanford International Nuclear Information System (INIS) Johnson, A.R.; Caudill, J.G.; Giddings, R.F.; Rodriguez, J.M.; Roos, R.C.; Wilde, J.W. 2010-01-01 In 1999 an integrated biological control system was instituted at the U.S. Department of Energy's Hanford Site. Successes and changes to the program needed to be communicated to a large and diverse mix of organizations and individuals. Efforts at communication are directed toward the following: Hanford Contractors (Liquid or Tank Waste, Solid Waste, Environmental Restoration, Science and Technology, Site Infrastructure), General Hanford Employees, and Hanford Advisory Board (Native American Tribes, Environmental Groups, Local Citizens, Washington State and Oregon State regulatory agencies). Communication was done through direct interface meetings, individual communication, where appropriate, and broadly sharing program reports. The objectives of the communication efforts was to have the program well coordinated with Hanford contractors, and to have the program understood well enough that all stakeholders would have confidence in the work performed by the program to reduce or elimate spread of radioactive contamination by biotic vectors. Communication of successes and changes to an integrated biological control system instituted in 1999 at the Department of Energy's Hanford Site have required regular interfaces with not only a diverse group of Hanford contractors (i.e., those responsible for liquid or tank waste, solid wastes, environmental restoration, science and technology, and site infrastructure), and general Hanford employees, but also with a consortium of designated stake holders organized as the Hanford Advisory Board (i.e., Native American tribes, various environmental groups, local citizens, Washington state and Oregon regulatory agencies, etc.). Direct interface meetings, individual communication where appropriate, and transparency of the biological control program were the methods and outcome of this effort. 1. AN INTEGRATED BIOLOGICAL CONTROL SYSTEM AT HANFORD Energy Technology Data Exchange (ETDEWEB) JOHNSON AR; CAUDILL JG; GIDDINGS RF; RODRIGUEZ JM; ROOS RC; WILDE JW 2010-02-11 In 1999 an integrated biological control system was instituted at the U.S. Department of Energy's Hanford Site. Successes and changes to the program needed to be communicated to a large and diverse mix of organizations and individuals. Efforts at communication are directed toward the following: Hanford Contractors (Liquid or Tank Waste, Solid Waste, Environmental Restoration, Science and Technology, Site Infrastructure), General Hanford Employees, and Hanford Advisory Board (Native American Tribes, Environmental Groups, Local Citizens, Washington State and Oregon State regulatory agencies). Communication was done through direct interface meetings, individual communication, where appropriate, and broadly sharing program reports. The objectives of the communication efforts was to have the program well coordinated with Hanford contractors, and to have the program understood well enough that all stakeholders would have confidence in the work performed by the program to reduce or elimated spread of radioactive contamination by biotic vectors. Communication of successes and changes to an integrated biological control system instituted in 1999 at the Department of Energy's Hanford Site have required regular interfaces with not only a diverse group of Hanford contractors (i.e., those responsible for liquid or tank waste, solid wastes, environmental restoration, science and technology, and site infrastructure), and general Hanford employees, but also with a consortium of designated stake holders organized as the Hanford Advisory Board (i.e., Native American tribes, various environmental groups, local citizens, Washington state and Oregon regulatory agencies, etc.). Direct interface meetings, individual communication where appropriate, and transparency of the biological control program were the methods and outcome of this effort. 2. Hanford Waste Vitrification Plant Project advanced conceptual design summary report International Nuclear Information System (INIS) Anderson, T.D. 1988-11-01 The Hanford Waste Vitrification Plant (HWVP) will immobilize Hanford defense liquid high-level waste in borosilicate glass in preparation for shipment to a geologic repository. The shipment of the waste to the repository will satisfy an objective in the President's Defense Waste Management Plan. The glass product will be cast into stainless steel canisters, which will be sealed and stored at Hanford until they are shipped. This document summarizes work performed during the Advance Conceptual Design (ACD) of the HWVP. In the Reference Conceptual Design phase, which preceded the ACD, a number of design issues were identified with the potential to improve cost effectiveness, safety, constructibility, and operability. The ACD addressed and evaluated these design issues. Implementation of recommendations derived from ACD work will occur in subsequent design phases. The next design phase is preliminary design which will be followed by detailed design and construction. Net potential cost improvements of more than$36.9M were identified along with improvements in safety, constructibility, and operability. No negative schedule impacts will result from implementation of the improvements. 11 refs., 5 figs., 3 tabs 3. Hanford low-level waste process chemistry testing data package International Nuclear Information System (INIS) Smith, H.D.; Tracey, E.M.; Darab, J.G.; Smith, P.A. 1996-03-01 Recently, the Tri-Party Agreement (TPA) among the State of Washington Department of Ecology, U.S. Department of Energy (DOE) and the US Environmental Protection Agency (EPA) for the cleanup of the Hanford Site was renegotiated. The revised agreement specifies vitrification as the encapsulation technology for low level waste (LLW). A demonstration, testing, and evaluation program underway at Westinghouse Hanford Company to identify the best overall melter-system technology available for vitrification of Hanford Site LLW to meet the TPA milestones. Phase I is a open-quotes proof of principleclose quotes test to demonstrate that a melter system can process a simulated highly alkaline, high nitrate/nitrite content aqueous LLW feed into a glass product of consistent quality. Seven melter vendors were selected for the Phase I evaluation: joule-heated melters from GTS Duratek, Incorporated (GDI); Envitco, Incorporated (EVI); Penberthy Electomelt, Incorporated (PEI); and Vectra Technologies, Incorporated (VTI); a gas-fired cyclone burner from Babcock ampersand Wilcox (BCW); a plasma torch-fired, cupola furnace from Westinghouse Science and Technology Center (WSTC); and an electric arc furnace with top-entering vertical carbon electrodes from the U.S. Bureau of Mines (USBM) 4. Hanford Waste Vitrification Plant Clean Air Act permit application International Nuclear Information System (INIS) 1990-04-01 This document briefly describes the Hanford Site and provides a general overview of the Hanford Waste Vitrification Plant (HWVP). Other topics include sources of emissions, facility operating parameters, facility emissions, pollutant and radionuclide control technology and air quality. The HWVP will convert mixed wastes (high-activity radioactive and hazardous liquid wastes) to a solid vitrified form (borosilicate glass) for disposal. Mixed wastes pretreated in the Hanford Site B Plant will be pumped into double- shell tanks in the 200 East Area for interim storage. This pretreated mixed waste will be batch transferred from interim storage to the HWVP facility, where the waste will be concentrated by evaporation, treated with chemicals, and mixed with glass-forming materials. The mixture will then be continuously fed into an electrically heated glass melter. The molten glass will be poured into canisters that will be cooled, sealed, decontaminated, and stored until the vitrified product can be transferred to a geologic repository. 25 refs., 18 figs., 32 tabs 5. Registration for the Hanford Site: Sources of radioactive emissions International Nuclear Information System (INIS) Silvia, M.J. 1993-04-01 This Registration Application serves to renew the registration for all Hanford Site sources of radioactive air emissions routinely reported to the State of Washington Department of Health (DOH). The current registration expires on August 15, 1993. The Application is submitted pursuant to the Washington Administrative Code (WAC) Chapter 246--247, and is consistent with guidance provided by DOH for renewal. The Application subdivides the Hanford Site into six major production, processing or research areas. Those six areas are in the 100 Area, 200 East Area, 200 West Area, 300 Area, 400 Area, and 600 Area. Each major group of point sources within the six areas listed above is represented by a Source Registration for Radioactive Air Emissions form. Annual emissions. for the sources are listed in the ''Radionuclide Air Emissions Report for the Hanford Site,'' published annually. It is a requirement that the following Statement of Compliance be provided: ''The radioactive air emissions from the above sources do meet the emissions standards contained in Chapter 173-480-040 WAC, Ambient Air Quality Standards and Emissions Limits for Radionuclides. As the Statement of Compliance pertains to this submittal, the phrase ''above sources'' is to be understood as meaning the combined air emissions from all sources registered by this submittal 6. Atmospheric reactions of methylcyclohexanes with Cl atoms and OH radicals: determination of rate coefficients and degradation products. Science.gov (United States) Ballesteros, Bernabé; Ceacero-Vega, Antonio A; Jiménez, Elena; Albaladejo, José 2015-04-01 As the result of biogenic and anthropogenic activities, large quantities of chemical compounds are emitted into the troposphere. Alkanes, in general, and cycloalkanes are an important chemical class of hydrocarbons found in diesel, jet and gasoline, vehicle exhaust emissions, and ambient air in urban areas. In general, the primary atmospheric fate of organic compounds in the gas phase is the reaction with hydroxyl radicals (OH). The oxidation by Cl atoms has gained importance in the study of atmospheric reactions because they may exert some influence in the boundary layer, particularly in marine and coastal environments, and in the Arctic troposphere. The aim of this paper is to study of the atmospheric reactivity of methylcylohexanes with Cl atoms and OH radicals under atmospheric conditions (in air at room temperature and pressure). Relative kinetic techniques have been used to determine the rate coefficients for the reaction of Cl atoms and OH radicals with methylcyclohexane, cis-1,4-dimethylcyclohexane, trans-1,4-dimethylcyclohexane, and 1,3,5-trimethylcyclohexane at 298 ± 2 K and 720 ± 5 Torr of air by Fourier transform infrared) spectroscopy and gas chromatography-mass spectrometry (GC-MS) in two atmospheric simulation chambers. The products formed in the reaction under atmospheric conditions were investigated using a 200-L Teflon bag and employing the technique of solid-phase microextraction coupled to a GC-MS. The rate coefficients obtained for the reaction of Cl atoms with the studied compounds are the following ones (in units of 10(-10) cm(3) molecule(-1) s(-1)): (3.11 ± 0.16), (2.89 ± 0.16), (2.89 ± 0.26), and (2.61 ± 0.42), respectively. For the reactions with OH radicals the determined rate coefficients are (in units of 10(-11) cm(3) molecule(-1) s(-1)): (1.18 ± 0.12), (1.49 ± 0.16), (1.41 ± 0.15), and (1.77 ± 0.23), respectively. The reported error is twice the standard deviation. A detailed 7. Overview of the Hanford risk management plan International Nuclear Information System (INIS) Halverson, T.G. 1998-01-01 The Project Hanford Management Contract called for the enhancement of site-wide decision processes, and development of a Hanford Risk Management Plan to adopt or develop a risk management system for the Hanford Site. This Plan provides a consistent foundation for Site issues and addresses site-wide management of risks of all types. It supports the Department of Energy planning and sitewide decision making policy. Added to this requirement is a risk performance report to characterize the risk management accomplishments. This paper presents the development of risk management within the context of work planning and performance. Also discussed are four risk elements which add value to the context 8. Hanford's Radioactive Mixed Waste Disposal Facility International Nuclear Information System (INIS) McKenney, D.E. 1995-01-01 The Radioactive Mixed Waste Disposal Facility, is located in the Hanford Site Low-Level Burial Grounds and is designated as Trench 31 in the 218-W-5 Burial Ground. Trench 31 is a Resource Conservation and Recovery Act compliant landfill and will receive wastes generated from both remediation and waste management activities. On December 30, 1994, Westinghouse Hanford Company declared readiness to operate Trench 31, which is the Hanford Site's (and the Department of Energy complex's) first facility for disposal of low-level radioactive mixed wastes 9. High level waste at Hanford: Potential for waste loading maximization International Nuclear Information System (INIS) Hrma, P.R.; Bailey, A.W. 1995-09-01 The loading of Hanford nuclear waste in borosilicate glass is limited by phase-related phenomena, such as crystallization or formation of immiscible liquids, and by breakdown of the glass structure because of an excessive concentration of modifiers. The phase-related phenomena cause both processing and product quality problems. The deterioration of product durability determines the ultimate waste loading limit if all processing problems are resolved. Concrete examples and mass-balance based calculations show that a substantial potential exists for increasing waste loading of high-level wastes that contain a large fraction of refractory components 10. Surface recombination of oxygen atoms in O2 plasma at increased pressure: II. Vibrational temperature and surface production of ozone Science.gov (United States) Lopaev, D. V.; Malykhin, E. M.; Zyryanov, S. M. 2011-01-01 Ozone production in an oxygen glow discharge in a quartz tube was studied in the pressure range of 10-50 Torr. The O3 density distribution along the tube diameter was measured by UV absorption spectroscopy, and ozone vibrational temperature TV was found comparing the calculated ab initio absorption spectra with the experimental ones. It has been shown that the O3 production mainly occurs on a tube surface whereas ozone is lost in the tube centre where in contrast the electron and oxygen atom densities are maximal. Two models were used to analyse the obtained results. The first one is a kinetic 1D model for the processes occurring near the tube walls with the participation of the main particles: O(3P), O2, O2(1Δg) and O3 molecules in different vibrational states. The agreement of O3 and O(3P) density profiles and TV calculated in the model with observed ones was reached by varying the single model parameter—ozone production probability (\\gamma_{O_{3}}) on the quartz tube surface on the assumption that O3 production occurs mainly in the surface recombination of physisorbed O(3P) and O2. The phenomenological model of the surface processes with the participation of oxygen atoms and molecules including singlet oxygen molecules was also considered to analyse \\gamma_{O_{3}} data obtained in the kinetic model. A good agreement between the experimental data and the data of both models—the kinetic 1D model and the phenomenological surface model—was obtained in the full range of the studied conditions that allowed consideration of the ozone surface production mechanism in more detail. The important role of singlet oxygen in ozone surface production was shown. The O3 surface production rate directly depends on the density of physisorbed oxygen atoms and molecules and can be high with increasing pressure and energy inputted into plasma while simultaneously keeping the surface temperature low enough. Using the special discharge cell design, such an approach opens up the 11. Surface recombination of oxygen atoms in O2 plasma at increased pressure: II. Vibrational temperature and surface production of ozone International Nuclear Information System (INIS) Lopaev, D V; Malykhin, E M; Zyryanov, S M 2011-01-01 Ozone production in an oxygen glow discharge in a quartz tube was studied in the pressure range of 10-50 Torr. The O 3 density distribution along the tube diameter was measured by UV absorption spectroscopy, and ozone vibrational temperature T V was found comparing the calculated ab initio absorption spectra with the experimental ones. It has been shown that the O 3 production mainly occurs on a tube surface whereas ozone is lost in the tube centre where in contrast the electron and oxygen atom densities are maximal. Two models were used to analyse the obtained results. The first one is a kinetic 1D model for the processes occurring near the tube walls with the participation of the main particles: O( 3 P), O 2 , O 2 ( 1 Δ g ) and O 3 molecules in different vibrational states. The agreement of O 3 and O( 3 P) density profiles and T V calculated in the model with observed ones was reached by varying the single model parameter-ozone production probability (γ O 3 ) on the quartz tube surface on the assumption that O 3 production occurs mainly in the surface recombination of physisorbed O( 3 P) and O 2 . The phenomenological model of the surface processes with the participation of oxygen atoms and molecules including singlet oxygen molecules was also considered to analyse γ O 3 data obtained in the kinetic model. A good agreement between the experimental data and the data of both models-the kinetic 1D model and the phenomenological surface model-was obtained in the full range of the studied conditions that allowed consideration of the ozone surface production mechanism in more detail. The important role of singlet oxygen in ozone surface production was shown. The O 3 surface production rate directly depends on the density of physisorbed oxygen atoms and molecules and can be high with increasing pressure and energy inputted into plasma while simultaneously keeping the surface temperature low enough. Using the special discharge cell design, such an approach opens up 12. Risk management study for the retired Hanford Site facilities: Qualitative risk evaluation for the retired Hanford Site facilities International Nuclear Information System (INIS) Coles, G.A.; Shultz, M.V.; Taylor, W.E. 1993-09-01 This document provides a risk evaluation of the 100 and 200 Area retired, surplus facilities on the Hanford Site. Also included are the related data that were compiled by the risk evaluation team during investigations performed on the facilities. Results are the product of a major effort performed in fiscal year 1993 to produce qualitative information that characterizes certain risks associated with these facilities. The retired facilities investigated for this evaluation are located in the 100 and 200 Areas of the 1,450-km 2 (570-mi 2 ) Hanford Site. The Hanford Site is a semiarid tract of land in southeastern Washington State. The nearest population center is Richland, Washington, (population 32,000) 30-km (20 mi) southeast of the 200 Area. During walkdown investigations of these facilities, data on real and potential hazards that threatened human health or safety or created potential environmental release issues were identified by the risk evaluation team. Using these findings, the team categorized the identified hazards by facility and evaluated the risk associated with each hazard. The factors contributing to each risk, and the consequence and likelihood of harm associated with each hazard also are included in this evaluation 13. Risk management study for the retired Hanford Site facilities: Qualitative risk evaluation for the retired Hanford Site facilities. Volume 3 Energy Technology Data Exchange (ETDEWEB) Coles, G.A.; Shultz, M.V.; Taylor, W.E. 1993-09-01 This document provides a risk evaluation of the 100 and 200 Area retired, surplus facilities on the Hanford Site. Also included are the related data that were compiled by the risk evaluation team during investigations performed on the facilities. Results are the product of a major effort performed in fiscal year 1993 to produce qualitative information that characterizes certain risks associated with these facilities. The retired facilities investigated for this evaluation are located in the 100 and 200 Areas of the 1,450-km{sup 2} (570-mi{sup 2}) Hanford Site. The Hanford Site is a semiarid tract of land in southeastern Washington State. The nearest population center is Richland, Washington, (population 32,000) 30-km (20 mi) southeast of the 200 Area. During walkdown investigations of these facilities, data on real and potential hazards that threatened human health or safety or created potential environmental release issues were identified by the risk evaluation team. Using these findings, the team categorized the identified hazards by facility and evaluated the risk associated with each hazard. The factors contributing to each risk, and the consequence and likelihood of harm associated with each hazard also are included in this evaluation. 14. Investigation of nitrogen atom production in Ar/N2 and He/N2 surface wave plasmas International Nuclear Information System (INIS) Tabbal, M.; Kazopoulo, M.; Christidis, T.; Isber, S. 2000-01-01 Full text: There is presently great interest in nitrogen plasmas for surface coating processes. Such as the deposition of nitride thin films and surface treatment of materials. Indeed, nitrogen plasmas have been used to nitride the surface of ferrous and non-ferrous materials in order to improve their surface properties such as resistance to corrosion and hardness. Moreover, the design and development of nitrogen atom sources could be essential for the synthesis of gallium nitride (GaN), a wide band-gap semiconductor whose properties have revolutionized the microelectronics and optoelectronics industries. Correlations have been established between the density of active species in the process, namely atomic nitrogen (N) produced by the discharge and GaN film properties. Thus, it is of fundamental importance to investigate the N-atom production mechanisms in such discharges. N-atom production has been studied in pure N 2 surface-wave plasmas (SWP), as a function of operating parameters, namely gas pressure and electrical power. These studies indicate that the increase in the gas temperature (T g ) limits the N-atom production. One possible way of enhancing the N 2 dissociation rate ([N]/[N 2 ]) in the plasma could be the use of gas mixtures such as Ar/N 2 or He/N 2 . the aim of this paper is to characterize an Ar/N 2 and He/N 2 surface-wave discharge (SWD) by optical emission spectroscopy (OES), in order to determine the optimal plasma conditions in terms of [N]/[N 2 ]. The plasma is generated by a radio frequency (40.68 MHz) wave launcher. The effect of mixing N 2 with Ar and He on the production of N-atoms in the plasma was investigated at varying experimental conditions, such as operating pressure (4.5 and 7.5 Torr), electrical power (40 to 120 W), at a total gas flow of 250 sccm. It was found that [N]/[N 2 ] increases with the partial pressure of Ar in the mixture by a factor of about 8 at 120W. Such an enhancement is reduced at lower incident powers. On the 15. Radioactive air emissions notice of construction and application for approval to construct the Hanford Waste Vitrification Plant International Nuclear Information System (INIS) 1992-10-01 The Hanford Site is owned by the US Government and operated by the US Department of Energy, Richland Field Office. The Hanford Site manages and produces dangerous waste and mixed waste. (containing both radioactive and dangerous components). The US Department of Energy, Richland Field Office, currently stores mixed waste, resulting from various processing operations, in underground storage tanks. The Hanford Waste Vitrification Plant will be constructed and operated to process the high-activity fraction of mixed waste stored in these underground tanks. The Hanford Waste Vitrification Plant will solidify pretreated tank waste into a glass product that will be packaged for disposal in a national repository. Emissions from the Hanford Waste Vitrification Plant will be regulated by both the federal and state Clean Air Acts. The proposed Hanford Waste Vitrification Plant represents a new source of radioactive air emissions. Construction of the plant will require approval from both federal and state agencies. The Notice of Construction and Application for Approval to Construct the Hanford Waste Vitrification Plant contains information required under Title 40 of the Code of Federal Regulations, Chapter 61; and Chapter 246-247 of the Washington Administrative Code for a proposed new source of radioactive air emissions. The document contents are based on information contained in the Hanford Waste Vitrification Plant Reference Conceptual Design Report, the Hanford Waste Vitrification Plant Preliminary Safety Analysis Report, Revision 0, and subsequent design changes made before August 1, 1992. The contents of this document may be modified to include more specific information generated during subsequent detailed design phases. Modifications will be submitted for regulatory review and approval, as appropriate 16. Hanford Laboratories monthly activities report, November 1964 Energy Technology Data Exchange (ETDEWEB) 1964-12-15 This is the monthly report for the Hanford Laboratories Operation, November 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research. 17. Hanford Laboratories monthly activities report, March 1963 Energy Technology Data Exchange (ETDEWEB) 1963-04-15 This is the monthly report for the Hanford Laboratories Operation March 1963. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, operations research and synthesis, programming, and radiation protection operation are discussed. 18. Hanford Site baseline risk assessment methodology International Nuclear Information System (INIS) 1992-03-01 This report describes risk assessment methodology associated with the remedial action programs at the Hanford Reservation. Topics addressed include human health evaluation, pollutant and radionuclide transport through the environment, and environmental transport pathways 19. Hanford Laboratories monthly activities report, December 1963 Energy Technology Data Exchange (ETDEWEB) 1964-01-15 The monthly report for the Hanford Laboratories Operation, December 1963. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research, and applied mathematics, and programming operations are discussed. 20. Hanford Environmental Information System Configuration Management Plan International Nuclear Information System (INIS) 1996-06-01 The Hanford Environmental Information System (HEIS) Configuration Management Plan establishes the software and data configuration control requirements for the HEIS and project-related databases maintained within the Environmental Restoration Contractor's data management department 1. Hanford Laboratories monthly activities report, October 1963 Energy Technology Data Exchange (ETDEWEB) 1963-11-15 This is the monthly report for the Hanford Laboratories Operation, October 1963. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 2. Hanford Laboratories monthly activities report, January 1964 Energy Technology Data Exchange (ETDEWEB) 1964-02-14 This is the monthly report for the Hanford Laboratories Operation, January 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, applied mathematics, programming operation, and radiation protection are discussed. 3. Hanford Laboratories monthly activities report, August 1963 Energy Technology Data Exchange (ETDEWEB) 1963-09-16 This is the monthly report for the Hanford Laboratories Operation, August 1963. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 4. Hanford Laboratories monthly activities report, May 1964 Energy Technology Data Exchange (ETDEWEB) 1964-06-15 This is the monthly report for the Hanford Laboratories Operation, May 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, applied mathematics, programming operation, and radiation protection are discussed. 5. Hanford Laboratories monthly activities report, January 1963 Energy Technology Data Exchange (ETDEWEB) 1963-02-15 This is the monthly report for the Hanford Laboratories Operation January 1963. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, operations research and synthesis, programming, and radiation protection operation are discussed. 6. Hanford Laboratories monthly activities report, September 1963 Energy Technology Data Exchange (ETDEWEB) 1963-10-15 This is the monthly report for the Hanford Laboratories Operation, September 1963. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 7. Hanford Laboratories monthly activities report, July 1963 Energy Technology Data Exchange (ETDEWEB) 1963-08-15 This is the monthly report for the Hanford Laboratories Operation, July 1963. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 8. Hanford Laboratories monthly activities report, May 1963 Energy Technology Data Exchange (ETDEWEB) 1963-06-14 The monthly report for the Hanford Laboratories Operation, May 1963. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research, and applied mathematics, and programming operation are discussed. 9. Hanford Laboratories monthly activities report, February 1964 Energy Technology Data Exchange (ETDEWEB) 1964-03-16 This is the monthly report for the Hanford Laboratories Operation, February, 1964. Reactor fuels, chemistry, dosimetry, separation process, reactor technology financial activities, biology operation, physics and instrumentation research, employee relations, applied mathematics, programming, and radiation protection are discussed. 10. Hanford Laboratories monthly activities report, June 1963 Energy Technology Data Exchange (ETDEWEB) 1963-07-15 This is the monthly report for the Hanford Laboratories Operation, June 1963. Metallurgy, reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, visits, biology operation, physics and instrumentation research, and employee relations are discussed. 11. Continuing study of mortality in Hanford workers International Nuclear Information System (INIS) Marks, S.; Gilbert, E.S. 1979-10-01 The mortality of workers at the Hanford Plant in southeastern Washington who have been exposed to penetrating external ionizing radiation is studied. Deaths are analyzed statistically and compared to standardized mortality ratios. Cancer deaths in particular are examined 12. Hanford Laboratories monthly activities report, April 1964 Energy Technology Data Exchange (ETDEWEB) 1964-05-15 This is the monthly report for the Hanford Laboratories Operation, April 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, applied mathematics, programming operation, and radiation protection are discussed. 13. Hanford Laboratories monthly activities report, July 1964 Energy Technology Data Exchange (ETDEWEB) 1964-08-14 This is the monthly report for the Hanford Laboratories Operation, July 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, physics and instrumentation research, applied mathematics, programming operation, and radiation protection are discussed. 14. Hanford Laboratories monthly activities report, March 1964 Energy Technology Data Exchange (ETDEWEB) 1964-04-15 The monthly report for the Hanford Laboratories Operation, March 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research, and applied mathematics operation, and programming operations are discussed. 15. Hanford Laboratories monthly activities report, April, 1963 Energy Technology Data Exchange (ETDEWEB) 1963-05-15 This is the monthly report for the Hanford Laboratories Operation, April, 1963. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology financial activities, biology operation, physics and instrumentation research, employee relations, applied mathematics operation, programming, and radiation protection operation discussed. 16. Hanford Laboratories monthly activities report, August 1964 Energy Technology Data Exchange (ETDEWEB) 1964-09-15 The monthly report for the Hanford Laboratories Operation, August 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research, and applied mathematics, and programming operations are discussed. 17. Hanford Laboratories monthly activities report, October 1964 Energy Technology Data Exchange (ETDEWEB) 1964-11-16 The monthly report for the Hanford Laboratories Operation, October 1964. Reactor fuels, chemistry, dosimetry, separation processes, reactor technology, financial activities, biology operation, and physics and instrumentation research, and applied mathematics operations are discussed. 18. The place of Turkish Atomic Energy Authority in nuclear energy production International Nuclear Information System (INIS) Sanalan, Y. 1994-01-01 Turkish Atomic Energy Authority (TAEA), established in 1982 by the Act no: 2690, is a governmental organization directly under the direct supervision of the Prime Minister. The objective of the establisment of TAEA is the peaceful utilization of atomic energy, regarding the national policy and the related plans and programs, for the benefits of State. The main duties of TAEA, as stated in related Act, can be summarized as: to determine and progress the basis of the national policy and the related plans and programs and to submit them to be approved by the Prime Minister; to execute and to support scientific, technical and administrative studies; to give approval, permission and license to the nuclear installations; to enlighten the public in nuclear matters; to establish research and educate the personnel in the nuclear field If Turkey would participate in a Nuclear Energy Program, especially, TAEA should perform its own duties properly and in this respect, as an Authority, we should have objectives that have to be determined as State Policy 19. Integrated Monitoring Plan for the Hanford Groundwater Monitoring Project International Nuclear Information System (INIS) Hartman, Mary J.; Dresel, P. Evan; Lindberg, Jon W.; Newcomer, Darrell R.; Thornton, Edward C. 2000-01-01 Groundwater is monitored at the Hanford Site to fulfill a variety of state and federal regulations, including the Atomic Energy Act of 1954; the Resource Conservation and Recovery Act of 1976; the Comprehensive Environmental Response, Compensation, and Liability Act of 1980; and Washington Administrative Code. Separate monitoring plans are prepared for various requirements, but sampling is coordinated and data are shared among users to avoid duplication of effort. The U.S. Department of Energy manages these activities through the Hanford Groundwater Monitoring Project. This document is an integrated monitoring plan for the groundwater project. It documents well and constituent lists for monitoring required by the Atomic Energy Act of 1954 and its implementing orders; includes other, established monitoring plans by reference; and appends a master well/constituent/ frequency matrix for the entire site. The objectives of monitoring fall into three general categories: plume and trend tracking, treatment/ storage/disposal unit monitoring, and remediation performance monitoring. Criteria for selecting Atomic Energy Act of 1954 monitoring networks include locations of wells in relation to known plumes or contaminant sources, well depth and construction, historical data, proximity to the Columbia River, water supplies, or other areas of special interest, and well use for other programs. Constituent lists were chosen based on known plumes and waste histories, historical groundwater data, and, in some cases, statistical modeling. Sampling frequencies were based on regulatory requirements, variability of historical data, and proximity to key areas. For sitewide plumes, most wells are sampled every 3 years. Wells monitoring specific waste sites or in areas of high variability will be sampled more frequently 20. Integrated Monitoring Plan for the Hanford Groundwater Monitoring Project International Nuclear Information System (INIS) Newcomer, D.R.; Thornton, E.C.; Hartman, M.J.; Dresel, P.E. 1999-01-01 Groundwater is monitored at the Hanford Site to fulfill a variety of state and federal regulations, including the Atomic Energy Act of 1954 the Resource Conservation and Recovery Act of 1976 the Comprehensive Environmental Response, Compensation, and Liability Act of 1980; and Washington Administrative Code. Separate monitoring plans are prepared for various requirements, but sampling is coordinated and data are shared among users to avoid duplication of effort. The US Department of Energy manages these activities through the Hanford Groundwater Monitoring Project. This document is an integrated monitoring plan for the groundwater project. It documents well and constituent lists for monitoring required by the Atomic Energy Act of 1954 and its implementing orders; includes other, established monitoring plans by reference; and appends a master well/constituent/frequency matrix for the entire site. The objectives of monitoring fall into three general categories plume and trend tracking, treatment/storage/disposal unit monitoring, and remediation performance monitoring. Criteria for selecting Atomic Energy Act of 1954 monitoring networks include locations of wells in relation to known plumes or contaminant sources, well depth and construction, historical data, proximity to the Columbia River, water supplies, or other areas of special interest, and well use for other programs. Constituent lists were chosen based on known plumes and waste histories, historical groundwater data, and, in some cases, statistical modeling. Sampling frequencies were based on regulatory requirements, variability of historical data, and proximity to key areas. For sitewide plumes, most wells are sampled every 3 years. Wells monitoring specific waste sites or in areas of high variability will be sampled more frequently 1. Hanford Nuclear Energy Center study International Nuclear Information System (INIS) Harty, H. 1976-01-01 Studies of a Nuclear Energy Center (NEC) at Hanford have not revealed any insurmountable technical problems, but problems have been identified that appear to be more difficult to resolve than for dispersed siting. Major technical developments in meteorology, and probably in seismology, are needed before an environmental report or safety analysis report could be prepared for an NEC. It would be helpful in further NEC studies if licensing requirements (and related criteria) were defined for them. An NEC will likely cause a step change in the amount of planning and involvement of regional groups in the energy picture compared to dispersed siting. The tools that must be developed for analysis of NECs will probably be used for evaluating dispersed siting in greater detail. NECs will probably bring about the use of dry or wet/dry cooling before it is required in equivalent amount for dispersed plants 2. Hanford cultural resources management plan Energy Technology Data Exchange (ETDEWEB) Chatters, J.C. (ed.) 1989-06-01 As a federal agency, the US Department of Energy (DOE) has been directed by Congress and the President to provide leadership in the preservation of prehistoric, historical, and cultural resources on lands it administers, to manage these in a spirit of stewardship for future generations, and to protect and preserve the rights of Native Americans to religious freedom. The purpose of this document is to describe how the DOE-Richland Operations (DOE-RL) will meet those responsibilities on the Hanford Site, pursuant to guidelines for Agency Responsibilities under the Historic Preservation Act (FR 53:31, February 17, 1988). This document is intended for multiple uses. Among other things, the text is designed as a manual for cultural resource managers to follow and as an explanation of the process of cultural resource regulatory compliance for the DOE-RL and Site contractors. 10 refs., 17 figs., 11 tabs. 3. Hanford 300 Area Development Plan International Nuclear Information System (INIS) Daly, K.S.; Seiler, S.W.; Hail, J.C. 1991-09-01 The purpose of the Hanford 300 Area Development Plan (Development Plan) is to guide the physical development of the 300 Area in accordance with US Department of Energy (DOE) Order 4320.1B (DOE 1991b) by performing the following: (1) Establishing a land use plan, setting land use categories that meet the needs of existing and proposed activities; (2) Coordinating existing, 5-yr, and long-range development plans and guiding growth in accordance with those plans; (3) Establishing development guidelines to encourage cost-effective development and minimize conflicts between adjacent activities; (4) Identifying site development issues that need further analysis; Integrating program plans with development plans to ensure a logical progression of development; and, (6) Integrating DOE plans with local agency plans (i.e., city, country, state, and Tri-Cities Science and Technology Park plans) 4. Aluminum precipitation from Hanford DSSF International Nuclear Information System (INIS) Borgen, D.; Frazier, P.; Staton, G. 1994-01-01 A series of pilot scale tests using simulated Double Shell Slurry Feed (DSSF) showed that well-settled aluminum precipitate can be produced in Hanford double shell tank (DST) high level waste by slow neutralization with carbon dioxide. This pretreatment could provide an early grout feed and free tank space, as well as facilitate downstream processes such as ion exchange by providing a less caustic feed. A total of eight test runs were completed using a 10-ft tall 3-in i.d. glass column. The 10-ft height corresponds to about one third of the vertical height of a DST, hence providing a reasonable basis for extrapolating the observed precipitate settling and compaction to the actual waste tank environment. Four runs (three with a simplified simulant and one with a chemically complete simulant) produced well settled precipitates averaging 1.5 to 2 feet high. Aluminum gel rather than settled precipitate resulted from one test where neutralization was too rapid 5. Hanford Environmental Information System (HEIS) International Nuclear Information System (INIS) 1994-01-01 The purpose of the Tank Characterization Data subject area of the Hanford Environmental Information System (HEIS) is to manage data acquired from waste tank characterization efforts. Tank samples provide the data stored in this subject area. Also included are data from tank inventories. These data are analyzed to determine disposal requirements, such as suitability for grout or vitrification. The data provide the basis for developing safety analyses and closure plans, and for establishing and verifying compliance with waste acceptance specifications. Two major sources of data make up the tank characterization data subject area: Data from single-shell and double-shell tank core samples -- core sampling analytical results include physical properties, radionuclides, major chemicals, and hazardous components; and data from waste tank supernatant samples. Four types of data are stored in the TCD subject area. Qualifiers for TCD analytical result data are listed in Appendix A. Data loading and verification procedures are described in Appendix B 6. Hanford Nuclear Energy Center study Energy Technology Data Exchange (ETDEWEB) Harty, H. 1976-03-16 Studies of a Nuclear Energy Center (NEC) at Hanford have not revealed any insurmountable technical problems, but problems have been identified that appear to be more difficult to resolve than for dispersed siting. Major technical developments in meteorology, and probably in seismology, are needed before an environmental report or safety analysis report could be prepared for an NEC. It would be helpful in further NEC studies if licensing requirements (and related criteria) were defined for them. An NEC will likely cause a step change in the amount of planning and involvement of regional groups in the energy picture compared to dispersed siting. The tools that must be developed for analysis of NECs will probably be used for evaluating dispersed siting in greater detail. NECs will probably bring about the use of dry or wet/dry cooling before it is required in equivalent amount for dispersed plants. 7. Hanford: The evolution of a dinosaur International Nuclear Information System (INIS) Fulton, J. 1995-01-01 This article describes how the Westinghouse Hanford Company is reinventing the US DOE's Hanford Site, turning a 1940s-era dinosaur into a 1990s-style business. The major topics covered include the following: breaking the logjam by ending the inefficient cost-plus days; Concentrating resources on resolving urgent safety issues; contract reform with more incentive, greater risk; finally reengineering: the next step 8. Hanford Environmental Dose Reconstruction Project monthly report International Nuclear Information System (INIS) Finch, S.M. 1991-10-01 The objective of the Hanford Environmental Dose Reconstruction Project is to estimate the radiation doeses that individuals and populations could have received from nuclear operations at Hanford since 1944. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed, from release to impact on humans (dose estimates): Source terms; environmental transport; environmental monitoring data; demographics, agriculture, food habits; environmental pathways and dose estimates 9. Hanford soil partitioning and vapor extraction study International Nuclear Information System (INIS) Yonge, D.; Hossain, A.; Cameron, R.; Ford, H.; Storey, C. 1996-07-01 This report describes the testing and results of laboratory experiments conducted to assist the carbon tetrachloride soil vapor extraction project operating in the 200 West Area of the Hanford Site in Richland, Washington. Vapor-phase adsorption and desorption testing was performed using carbon tetrachloride and Hanford Site soils to estimate vapor-soil partitioning and reasonably achievable carbon tetrachloride soil concentrations during active vapor extractions efforts at the 200 West Area. (CCl 4 is used in Pu recovery from aqueous streams.) 10. Hanford Site Waste management units report International Nuclear Information System (INIS) 1992-01-01 This report summarizes the operable units in several areas of the Hanford Site Waste Facility. Each operable unit has several waste units (crib, ditch, pond, etc.). The operable units are summarized by describing each was unit. Some of the descriptions are unit name, unit type, waste category start data, site description, etc. The descriptions will vary for each waste unit in each operable unit and area of the Hanford Site 11. Environmental surveillance at Hanford for CY 1977 International Nuclear Information System (INIS) Houston, J.R.; Blumer, P.J. 1978-04-01 Environmental data collected during 1977 show continued compliance by Hanford with all applicable state and federal regulations. Data were collected for most environmental media including air, Columbia River water, external radiation, foodstuffs (milk, beef, eggs, poultry, and produce) and wildlife (deer, fish, game birds, and oysters from Willapa Bay), as well as soil and vegetation samples. In general, offsite levels of radionuclides attributable to Hanford operations during 1977 were indistinguishable from background levels 12. Hanford Site Tank Waste Remediation System International Nuclear Information System (INIS) 1993-05-01 The US Department of Energy's (DOE) Hanford Site in southeastern Washington State has the most diverse and largest amount of highly radioactive waste of any site in the US. High-level radioactive waste has been stored in large underground tanks since 1944. A Tank Waste Remediation System Program has been established within the DOE to safely manage and immobilize these wastes in anticipation of permanent disposal in a geologic repository. The Hanford Site Tank Waste Remediation System Waste Management 1993 Symposium Papers and Viewgraphs covered the following topics: Hanford Site Tank Waste Remediation System Overview; Tank Waste Retrieval Issues and Options for their Resolution; Tank Waste Pretreatment - Issues, Alternatives and Strategies for Resolution; Low-Level Waste Disposal - Grout Issue and Alternative Waste Form Technology; A Strategy for Resolving High-Priority Hanford Site Radioactive Waste Storage Tank Safety Issues; Tank Waste Chemistry - A New Understanding of Waste Aging; Recent Results from Characterization of Ferrocyanide Wastes at the Hanford Site; Resolving the Safety Issue for Radioactive Waste Tanks with High Organic Content; Technology to Support Hanford Site Tank Waste Remediation System Objectives 13. NHC's contribution to cleanup of the Hanford Site International Nuclear Information System (INIS) Chauve, H.D. 1998-01-01 The one billion dollars per year Project Hanford Management Contract (PHMC), managed by Fluor Daniel Hanford, calls for cleanup of the Hanford Site for the Department of Energy. Project Hanford comprises four major subprojects, each managed by a different major contractor. Numatec Hanford Corporation (NHC) is a fifth major subcontractor which provides energy and technology to each of the Hanford projects. NHC draws on the experience and capabilities of its parent companies, COGEMA and SGN, and relies on local support from its sister Company in Richland, COGEMA Engineering Corporation, to bring the best commercial practices and new technology to the Project 14. Production of radionuclides and radiopharmaceuticals at the Argentine Atomic Energy National Commission International Nuclear Information System (INIS) Mitta, A.E.A.; Bonetto, Oscar; Kurcbart, Horacio; Mancini, Alberto; Marquez, R.O.; Palcos, M.C.; Quihillalt, E.L.; Salas, G.N.B. de; Suner, A.A.; Troparevsky, M.L.P. de. 1978-03-01 The production of radionuclides and radiopharmaceuticals at the CNEA is described, as well as the preparation of reagent's sets, and informations is given on the preparation of Tc-99m and In-113m generators. Some figures of the production of 1974-1976 are given. (author) [es 15. DEVELOPMENT OF HIGH-DENSITY U/AL DISPERSION PLATES FOR MO-99 PRODUCTION USING ATOMIZED URANIUM POWDER Directory of Open Access Journals (Sweden) HO JIN RYU 2013-12-01 Full Text Available Uranium metal particle dispersion plates have been proposed as targets for Molybdenum-99 (Mo-99 production to improve the radioisotope production efficiency of conventional low enriched uranium targets. In this study, uranium powder was produced by centrifugal atomization, and miniature target plates containing uranium particles in an aluminum matrix with uranium densities up to 9 g-U/cm3 were fabricated. Additional heat treatment was applied to convert the uranium particles into UAlx compounds by a chemical reaction of the uranium particles and aluminum matrix. Thus, these target plates can be treated with the same alkaline dissolution process that is used for conventional UAlx dispersion targets, while increasing the uranium density in the target plates 16. Laboratory testing of ozone oxidation of Hanford site waste International Nuclear Information System (INIS) Delegard, C.H.; Stubbs, A.M.; Bolling, S.D.; Colby, S.A. 1994-01-01 Organic constituents in radioactive waste stored in underground tanks at the U.S. Department of Energy's Hanford Site provoke safety concerns arising from their low-temperature reactions with nitrate and nitrite oxidants. Destruction of the organics would eliminate both safety problems. Oxone oxidation was investigated to destroy organic species present in simulated and genuine waste from Hanford Site Tank 241-SY-101. Bench-scale tests showed high-shear mixing apparatus achieved efficient gas-to-solution mass transfer and utilization of the ozone reagent. Oxidations of nitrite (to form nitrate) and organic species were observed. The organics formed carbonate and oxalate as well as nitrate and nitrogen gas from organic nitrogen. Formate, acetate and oxalate were present both in source waste and as reaction intermediates. Metal species oxidations also were observed directly or inferred by solubilities. Chemical precipitations of metal ions such as strontium and americium occurred as the organic species were destroyed by ozone. Reaction stoichiometries were consistent with the reduction of one oxygen atom per ozone molecule 17. National Environmental Policy Act source guide for the Hanford Site International Nuclear Information System (INIS) Jansky, M.T. 1998-01-01 This Source Guide will assist those working with the National Environmental Policy Act (NEPA) of 1969 to become more familiar with the environmental assessments (EA) and environmental impact statements (EIS) that apply to specific activities and facilities on the Hanford Site. This document should help answer questions concerning NEPA coverage, history, processes, and the status of many of the buildings and units on and related to the Hanford Site. This document summarizes relevant EAs and EISs by briefly outlining the proposed action of each document and the decision made by the US Department of Energy (DOE) or its predecessor agencies, the US Atomic Energy Commission (AEC) and the US Energy Research and Development Administration (ERDA). The summary includes the proposed action alternatives and current status of the proposed action. If a decision officially was stated by the DOE, as in a finding of no significant impact (FONSI) or a record of decision (ROD), and the decision was located, a summary is provided. Not all federal decisions, such as FONSIs and RODS, can be found in the Federal Register (FR). For example, although significant large-action FONSIs can be found in the FR, some low-interest FONSIs might have been published elsewhere (i.e., local newspapers) 18. Radionuclide releases to the atmosphere from Hanford Operations, 1944--1972. Hanford Environmental Dose Reconstruction Project Energy Technology Data Exchange (ETDEWEB) Heeb, C.M. 1994-05-01 The purpose of the Hanford Environmental Dose Reconstruction Project is to estimate the radiation dose that individuals could have received as a result of radionuclide emissions since 1944 from the Hanford Site. The first step in determining dose is to estimate the amount and timing of radionuclide releases to air and water. This report provides the air release information. 19. Collaboration in long-term stewardship at DOE Hanford Site Energy Technology Data Exchange (ETDEWEB) Moren, R. J.; Zeisloft, J. H.; Feist, E. T.; Brown, D.; Grindstaff, K. D. 2013-01-10 The U.S. Department of Energy's (DOE) Hanford Site comprises approximately 1,517 km{sup 2} (586 mi{sup 2}) of land in southeastern Washington. The site was established in 1943 as part of the Manhattan Project to produce plutonium for the nation's nuclear weapons program. As the Cold War era came to an end, the mission of the site transitioned from weapons production to environmental cleanup. As the River Corridor area of the site cleanup is completed, the mission for that portion of the site will transition from active cleanup to continued protection of environment through the Long-Term Stewardship (LTS) Program. The key to successful transition from cleanup to LTS is the unique collaboration among three (3) different DOE Programs and three (3) different prime contractors with each contractor having different contracts. The LTS Program at the site is a successful model of collaboration resulting in efficient resolution of issues and accelerated progress that supports DOE's Richland Office 2015 Vision for the Hanford Site. The 2015 Vision for the Hanford Site involves shrinking the active cleanup footprint of the surface area of the site to approximately 20 mi{sup 2} on the Central Plateau. Hanford's LTS Program is defined in DOE's planning document, Hanford Long-Term Stewardship Program Plan, DOE/RL-2010-35 Rev 1. The Plan defines the relationship and respective responsibilities between the federal cleanup projects and the LTS Program along with their respective contractors. The LTS Program involves these different parties (cleanup program and contractors) who must work together to achieve the objective for transition of land parcels. Through the collaborative efforts with the prime contractors on site over the past two years, 253.8 km{sup 2} (98 mi{sup 2}) of property has been successfully transitioned from the cleanup program to the LTS Program upon completion of active surface cleanup. Upcoming efforts in the near term will include transitioning another large 20. Advances In Hexavalent Chromium Removal At Hanford International Nuclear Information System (INIS) Neshem, D.O.; Riddelle, J. 2012-01-01 At the Hanford Site, chromium was used as a corrosion inhibitor in the reactor cooling water and was introduced into the groundwater as a result of planned and unplanned discharges from reactors during plutonium production since 1944. Beginning in 1995, groundwater treatment methods were evaluated leading to the use of pump and treat facilities with ion exchange using Dowex 21 K, a regenerable strong base anion exchange resin. This required regeneration of the resin, which is currently performed offsite. Resin was installed in a 4 vessel train, with resin removal required from the lead vessel approximately once a month. In 2007, there were 8 trains (32 vessels) in operation. In 2008, DOE recognized that regulatory agreements would require significant expansion in the groundwater chromium treatment capacity. Previous experience from one of the DOE project managers led to identification of a possible alternative resin, and the contractor was requested to evaluate alternative resins for both cost and programmatic risk reductions. Testing was performed onsite in 2009 and 2010, using a variety of potential resins in two separate facilities with groundwater from specific remediation sites to demonstrate resin performance in the specific groundwater chemistry at each site. The testing demonstrated that a weak base anion single-use resin, ResinTech SIR-700, was effective at removing chromium, had a significantly higher capacity, could be disposed of efficiently on site, and would eliminate the complexities and programmatic risks from sampling, packaging, transportation and return of resin for regeneration. This resin was installed in Hanford's newest groundwater treatment facility, called 100-DX, which began operations in November, 2010, and used in a sister facility, 100-HX, which started up in September of 2011. This increased chromium treatment capacity to 25 trains (100 vessels). The resin is also being tested in existing facilities that utilize Dowex 21 K for 1. Deep Atomic Binding (DAB) Approach in Interpretation of Fission Products Behavior in Terrestrial and Water Ecosystems International Nuclear Information System (INIS) Ajlouni, Abdul-Wali M.S. 2006-01-01 A large number of studies and models were established to explain the fission products (FP) behavior within terrestrial and water ecosystems, but a number of behaviors were non understandable, which always attributed to unknown reasons. According to DAB hypothesis, almost all fission products behaviors in terrestrial and water ecosystems could be interpreted in a wide coincidence. The gab between former models predictions, and field behavior of fission products after accidents like Chernobyl have been explained. DAB represents a tool to reduce radio-phobia as well as radiation protection expenses. (author) 2. Multiple H3+ fragment production in single collision of fast Hn+ clusters with He atoms International Nuclear Information System (INIS) Farizon, B.; Farizon, M.; Gaillard, M.J.; Gerlic, E.; Ouaskit, S. 1994-09-01 The production of H 3 + ions resulting from single collisions of mass-selected ionic hydrogen clusters, H n + (n=9,25,31), with helium at high velocity (1.55 times the Bohr velocity) has been studied. A strong double H 3 + ion production resulting from one incident cluster is observed. Moreover, evidence for a triple H 3 + fragment production is presented for n=25 and 31. Thus, in this energy range, the collision gives rise to multifragmentation processes. The formation of H 3 + ions takes place in the fragmentation of the multicharged cluster resulting from the collision. (authors) 3. Hanford Waste Vitrification Plant quality assurance program description: Overview and applications International Nuclear Information System (INIS) Caplinger, W.H. 1990-12-01 This document describes the Hanford Waste Vitrification Plant Project Quality Assurance Program. This program is being implemented to ensure the acceptability of high-level radioactive canistered waste forms produced by the Hanford Waste Vitrification Plant for disposal in a licensed federal repository. The Hanford Waste Vitrification Plant Quality Assurance Program is comprised of this Quality Assurance Program Description as well as the associated contractors' quality assurance programs. The objective of this Quality Assurance Program Description is to provide the Hanford Waste Vitrification Plant Project participants with guidance and direction for program implementation while satisfying the US Department of Energy Office of Civilian Radioactive Waste Management needs in repository licensing activities with regard to canistered waste forms. To accomplish this objective, this description will be prepared in three parts: Part 1 - Overview and applications document; Part 2 - Development and qualification of the canistered waste form; Part 3 - Production of canistered waste forms. Part 1 describes the background, strategy, application, and content of the Hanford Waste Vitrification Plant Quality Assurance Program. This Quality Assurance Program Description, when complete, is designed to provide a level of confidence in the integrity of the canistered waste forms. 8 refs 4. RCRA permitting strategies for the development of innovative technologies: Lessons from Hanford International Nuclear Information System (INIS) Gajewski, S.W.; Donaghue, J.F. 1994-01-01 The Hanford Site restoration is the largest waste cleanup operation in history. The Hanford plutonium production mission generated two-thirds of all the nuclear waste, by volume, in the Department of Energy (DOE) Complex. Cleanup challenges include not only large stored volumes of radioactive, hazardous, and mixed waste, but contaminated soil and groundwater and scores of major structures slated for decontamination, decommissioning, and demolition. DOE and its contractors will need to invent the technology required to do the job on a timetable driven by negotiated milestones, public concerns, and budgetary constraints. This paper will discuss the effort at Hanford to develop an integrated, streamlined strategy for compliance with the Resource Conservation and Recovery Act (RCRA) and the Comprehensive Environmental Response, Compensation and Liability Act (CERCLA) in the conduct of research, development, and demonstration (RD ampersand D) of innovative cleanup technologies. The aspects that will be discussed include the following: the genesis of the RD ampersand D permitting challenge at Hanford; permitting options in the existing regulatory framework; regulatory options that offered the best fit for Hanford RD ampersand D activities, and the problems associated with them; and conclusions and recommendations made to regulatory bodies 5. Hanford Waste Vitrification Plant capacity increase options International Nuclear Information System (INIS) Larson, D.E. 1996-04-01 Studies are being conducted by the Hanford Waste Vitrification Plant (HWVP) Project on ways to increase the waste processing capacity within the current Vitrification Building structural design. The Phase 1 study on remote systems concepts identification and extent of capacity increase was completed. The study concluded that the HWVP capacity could be increased to four times the current capacity with minor design adjustments to the fixed facility design, and the required design changes would not impact the current footprint of the vitrification building. A further increase in production capacity may be achievable but would require some technology development, verification testing, and a more systematic and extensive engineering evaluation. The primary changes included a single advance melter with a higher capacity, new evaporative feed tank, offgas quench collection tank, ejector venturi scrubbers, and additional inner canister closure station,a smear test station, a new close- coupled analytical facility, waste hold capacity of 400,000 gallon, the ability to concentrate out-of-plant HWVP feed to 90 g/L waste oxide concentration, and limited changes to the current base slab construction package 6. Characterization of the Hanford Site and environs Energy Technology Data Exchange (ETDEWEB) Cushing, C.E. (ed.) 1991-03-01 The US Department of Energy (DOE) proposes to site, construct, and operate a new production reactor (NPR) intended to produce materials for the US nuclear weapons program. The DOE has determined that this proposed action constitutes an action that may significantly affect the quality of the human environment; therefore, the DOE is preparing an environmental impact statement (EIS) to assess the potential impacts of the proposed action and reasonable alternatives on the human and natural environment. The NPR-EIS is being prepared in accordance with Section 102(2)(C) of the National Environmental Policy Act of 1969 (NEPA), as implemented in regulations (40 CFR 1500--1508) promulgated by the Council on Environmental Quality (CEQ). Information on the potentially affected environment at the Hanford Site and its environs was provided to ANL by PNL in various submissions during CY-1989, and some of that information was consolidated into this report, which is considered to be supporting documentation for the NPR-EIS. 93 refs., 35 figs., 46 tabs. 7. Hanford Engineer Works monthly report, July 1944 Energy Technology Data Exchange (ETDEWEB) 1944-08-08 This progress report discusses activities at Hanford Engineer Works for the month of July, 1944. The organization size increased rapidly in July, jumping from 1384 to 1886, an increase of 502 employees. Shortages existed on special lines such as electricians, instrument mechanics, and power operators. The opportunities to interview and employ Construction personnel improved as the Construction, organization moved past its peak employment figure. 269 houses were completed in Richland and 370 occupancies were reported, reducing the backlog of empty houses due to slow furniture movements. A total of 1530 houses or 35.6% of the village is now occupied. The overall canning yield increased from 63% to 75% while production leveled off at approximately two-thirds of design capacity. The metal machining operation and the test pile operated on a one shift-six day per week basis. The portion of the 100-B Area from the pump house through the filter plant and including the Power House was taken over by Operations on July 25. The Power House in the 200-W Area was started and flushing of process lines with water and steam got under way in the canyon. A new department was formed to control the Protection activities of the plant. 8. Description of a Multipurpose Processing and Storage Complex for the Hanford Site's radioactive material International Nuclear Information System (INIS) Nyman, D.H.; Wolfe, B.A.; Hoertkorn, T.R. 1993-05-01 The mission of the US Department of Energy's (DOE) Hanford Site has changed from defense nuclear materials production to that of waste management/disposal and environmental restoration. ne Multipurpose Processing and Storage Complex (MPSC) is being designed to process discarded waste tank internal hardware contaminated with mixed wastes, failed melters from the vitrification plant, and other Hanford Site high-level solid waste. The MPSC also will provide interim storage of other radioactive materials (irradiated fuel, canisters of vitrified high-level waste [HLW], special nuclear material [SNM], and other designated radioactive materials) 9. Immobilization of Radionuclides in the Hanford Vadose Zone by Incorporation in Solid Phases International Nuclear Information System (INIS) Brown, Gordon E. Jr.; Catalano, Jeffrey G.; Warner, Jeffrey A.; Samual Shaw; Daniel Grolimund 2005-01-01 The Department of Energy's Hanford Nuclear Site located in Washington State has accumulated over 2 million curies of radioactive waste from activities related to the production of plutonium (Ahearne, 1997). Sixty-seven of the single-shelled tanks located at the site are thought to have leaked, allowing between 2 and 4 million liters of waste fluids into the underlying vadose zone. The chemical processes employed at the Hanford Site to extract plutonium, as well as the need to minimize corrosion of the high-carbon steel storage tanks, resulted in uncharacterized hyperalkaline waste streams rich in radionuclides as well as other species including significant amounts of sodium and aluminum 10. HANFORD PLUTONIUM FINISHG PLAN (PFP) COMPLETES PLUTONIUM STABILIZATION KEY SAFETY ISSUES CLOSED International Nuclear Information System (INIS) GERBER, M.S. 2004-01-01 A long and intense effort to stabilize and repackage nearly 18 metric tons (MT) of plutonium-bearing leftovers from defense production and nuclear experiments concluded successfully in February, bringing universal congratulations to the Department of Energy's Hanford Site in southeast Washington State. The victorious stabilization and packaging endeavor at the Plutonium Finishing Plant (PFP), managed and operated by prime contractor Fluor Hanford, Inc., finished ahead of all milestones in Hanford's cleanup agreement with regulators, and before deadlines set by the Defense Nuclear Facilities Safety Board (DNFSB), a part of the federal Executive Branch that oversees special nuclear materials. The PFP stabilization and packaging project also completed under budget for its four-year tenure, and has been nominated for a DOE Secretarial Award. It won the Project of the Year Award in the local chapter competition of the Project Management Institute, and is being considered for awards at the regional and national level 11. DEWATERING TREATMENT SCALE-UP TESTING RESULTS OF HANFORD TANK WASTES International Nuclear Information System (INIS) TEDESCHI AR 2008-01-01 This report documents CH2M HILL Hanford Group Inc. (CH2M HILL) 2007 dryer testing results in Richland, WA at the AMEC Nuclear Ltd., GeoMelt Division (AMEC) Horn Rapids Test Site. It provides a discussion of scope and results to qualify the dryer system as a viable unit-operation in the continuing evaluation of the bulk vitrification process. A 10,000 liter (L) dryer/mixer was tested for supplemental treatment of Hanford tank low-activity wastes, drying and mixing a simulated non-radioactive salt solution with glass forming minerals. Testing validated the full scale equipment for producing dried product similar to smaller scale tests, and qualified the dryer system for a subsequent integrated dryer/vitrification test using the same simulant and glass formers. The dryer system is planned for installation at the Hanford tank farms to dry/mix radioactive waste for final treatment evaluation of the supplemental bulk vitrification process 12. Probabilistic safety assessment for Hanford high-level waste tanks International Nuclear Information System (INIS) MacFarlane, D.R.; Stack, D.S.; Kindinger, J.P.; Deremer, R.K. 1995-01-01 This paper gives results from the first comprehensive level-3 probabilistic safety assessment (PSA), including consideration of external events, for the Hanford tank farm (HTF). This work was sponsored by the U.S. Department of Energy/Environmental Restoration and Waste Management Division (DOE/EM). At the HTF, there are 177 underground tanks in 18 separate tank farms containing accumulated liquid/sludge/saltcake radioactive wastes from 50 yr of weapons materials production activities. The total waste volume is ∼60 million gal, containing ∼200 million Ci of radioactivity 13. Upgrades of Hanford Engineering Development Laboratory hot cell facilities International Nuclear Information System (INIS) Daubert, R.L.; DesChane, D.J. 1987-01-01 The Hanford Engineering Development Laboratory operates the 327 Postirradiation Testing Laboratory (PITL) and the 324 Shielded Materials Facility (SMF). These hot cell facilities provide diverse capabilities for the postirradiation examination and testing of irradiated reactor fuels and materials. The primary function of these facilities is to determine failure mechanisms and effects of irradiation on physical and mechanical properties of reactor components. The purpose of this paper is to review major equipment and facility upgrades that enhance customer satisfaction and broaden the engineering capabilities for more diversified programs. These facility and system upgrades are providing higher quality remote nondestructive and destructive examination services with increased productivity, operator comfort, and customer satisfaction 14. A history of solid waste packaging at the Hanford Site International Nuclear Information System (INIS) Duncan, D.R.; Weyns-Rollosson, D.I.; Pottmeyer, J.A.; Stratton, T.J. 1995-02-01 Since the initiation of the defense materials product mission, a total of more than 600,000 m 3 of radioactive solid waste has been stored or disposed at the US Department of Energy's (DOE) Hanford Site, located in southeastern Washington State. As the DOE complex prepares for its increasing role in environmental restoration and waste remediation, the characterization of buried and retrievably stored waste will become increasingly important. Key to this characterization is an understanding of the standards and specifications to which waste was packaged; the regulations that mandated these standards and specifications; the practices used for handling and packaging different waste types; and the changes in these practices with time 15. Hanford Waste Vitrification Plant Project Waste Form Qualification Program Plan International Nuclear Information System (INIS) Randklev, E.H. 1993-06-01 The US Department of Energy has created a waste acceptance process to help guide the overall program for the disposal of high-level nuclear waste in a federal repository. This Waste Form Qualification Program Plan describes the hierarchy of strategies used by the Hanford Waste Vitrification Plant Project to satisfy the waste form qualification obligations of that waste acceptance process. A description of the functional relationship of the participants contributing to completing this objective is provided. The major activities, products, providers, and associated scheduling for implementing the strategies also are presented 16. Antimatter atoms International Nuclear Information System (INIS) Anon. 1996-01-01 In january 1996, CERN broadcasted the information of the creation of nine anti-hydrogen atoms, observed through disintegration products. The experimental facility was CERN LEAR ring. An antiproton beam scattered a xenon jet, and the resulting antimatter was first selected by its insensitivity to beam bending magnets. Their disintegration was detected in thin NaI detectors, in which the anti-atoms are at once deprived from their positron. Then, magnetic and time-of-flight spectrometers are used. (D.L.) 17. Hanford science and technology needs statements document Energy Technology Data Exchange (ETDEWEB) Piper, L.L. 1997-12-31 This document is a compilation of the Hanford science and technology needs statements for FY 1998. The needs were developed by the Hanford Site Technology Coordination Group (STCG) with full participation and endorsement of site user organizations, stakeholders, and regulators. The purpose of this document is to: (a) provide a comprehensive listing of Hanford science and technology needs, and (b) identify partnering and commercialization opportunities with industry, other federal and state agencies, and the academic community. The Hanford STCG reviews and updates the needs annually. Once completed, the needs are communicated to DOE for use in the development and prioritization of their science and technology programs, including the Focus Areas, Cross-Cutting Programs, and the Environmental Management Science Program. The needs are also transmitted to DOE through the Accelerating Cleanup: 2006 Plan. The public may access the need statements on the Internet on: the Hanford Home Page (www.hanford.gov), the Pacific Rim Enterprise Centers web site (www2.pacific-rim.org/pacific rim), or the STCG web site at DOE headquarters (em-52.em.doegov/ifd/stcg/stcg.htm). This page includes links to science and technology needs for many DOE sites. Private industry is encouraged to review the need statements and contact the Hanford STCG if they can provide technologies that meet these needs. On-site points of contact are included at the ends of each need statement. The Pacific Rim Enterprise Center (206-224-9934) can also provide assistance to businesses interested in marketing technologies to the DOE. 18. Hanford Site Performance Report - March 1999 International Nuclear Information System (INIS) EDER, D.M. 2001-01-01 The purpose of the Hanford Site Performance Report is to provide the Department of Energy Richland Operations Office's (DOE-RL's) report of Hanford's performance by: U.S. Department of Energy, Richland Operations Office, Project Hanford Management Contract (PHMC) through Fluor Daniel Hanford, Inc. (FDH) and its subcontractors, Environmental Restoration Contract through Bechtel Hanford, Inc. (BHI), and its subcontractors, and Pacific Northwest National Laboratories (PNNL) for Science and Technology (S and T) Mission and support to the Environmental Management (EM). This report is published monthly with the intent of relating work performance and progress in the context of the Success Indicators and Critical Success Factors as outlined in the Hanford Strategic Plan. On a quarterly basis, the report also addresses performance and progress related to the Science and Technology Mission's Critical Outcomes as derived from the Hanford Strategic Plan. Section A of this report is the Executive Summary, encapsulating high-level data in this report into an overall brief. Summary information provided includes Notable Accomplishments, a performance profile with associated analyses, Critical Issues, Key Integration Activities, and a ''quick list'' of Upcoming Key Events. Section B of this report, the Site Summary section, provides Environmental Management performance data specifically organized to the pertinent Critical Success Factors and Success Indicators, and Science and Technology data in the context of the Critical Outcomes. The Site Summary demonstrates the various missions' overall progress against these strategic objectives. The information is presented in both narrative and graphical formats. The remaining sections provide performance data relative to each individual mission area (e.g., Waste Management, Spent Nuclear Fuels, etc.). The information provided in the Mission Area sections is at a level of greater detail than is presented in either the Executive Summary or 19. Hanford Site Performance Report - May 1999 International Nuclear Information System (INIS) EDER, D.M. 2001-01-01 The purpose of the Hanford Site Performance Report is to provide the Department of Energy Richland Operations Office's (DOE-RL's) report of Hanford's performance by: U. S. Department of Energy, Richland Operations Office, Project Hanford Management Contract (PHMC) through Fluor Daniel Hanford, Inc. (FDH) and its subcontractors, Environmental Restoration Contract through Bechtel Hanford, Inc. (BHI), and its subcontractors, and Pacific Northwest National Laboratories (PNNL) for Science and Technology (S and T) Mission and support to the Environmental Management (EM). This report is published monthly with the intent of relating work performance and progress in the context of the Success Indicators and Critical Success Factors as outlined in the Hanford Strategic Plan. On a quarterly basis, the report also addresses performance and progress related to the Science and Technology Mission's Critical Outcomes as derived from the Hanford Strategic Plan. Section A of this report is the Executive Summary, encapsulating high-level data in this report into an overall brief. Summary information provided includes Notable Accomplishments, a performance profile with associated analyses, Critical Issues, Key Integration Activities, and a ''quick list'' of Upcoming Key Events. Section B of this report, the Site Summary section, provides Environmental Management performance data specifically organized to the pertinent Critical Success Factors and Success Indicators, and Science and Technology data in the context of the Critical Outcomes. The Site Summary demonstrates the various missions' overall progress against these strategic objectives. The information is presented in both narrative and graphical formats. The remaining sections provide performance data relative to each individual mission area (e.g., Waste Management, Spent Nuclear Fuels, etc.). The information provided in the Mission Area sections is at a level of greater detail than is presented in either the Executive Summary or 20. Hanford Site Performance Report - April 1999 International Nuclear Information System (INIS) EDER, D.M. 2001-01-01 The purpose of the Hanford Site Performance Report is to provide the Department of Energy Richland Operations Office's (DOE-RL's) report of Hanford's performance by: U.S. Department of Energy, Richland Operations Office, Project Hanford Management Contract (PHMC) through Fluor Daniel Hanford, Inc. (FDH) and its subcontractors, Environmental Restoration Contract through Bechtel Hanford, Inc. (BHI), and its subcontractors, and Pacific Northwest National Laboratories (PNNL) for Science and Technology (S and T) Mission and support to the Environmental Management (EM). This report is published monthly with the intent of relating work performance and progress in the context of the Success Indicators and Critical Success Factors as outlined in the Hanford Strategic Plan. On a quarterly basis, the report also addresses performance and progress related to the Science and Technology Mission's Critical Outcomes as derived from the Hanford Strategic Plan. Section A of this report is the Executive Summary, encapsulating high-level data in this report into an overall brief. Summary information provided includes Notable Accomplishments, a performance profile with associated analyses, Critical Issues, Key Integration Activities, and a ''quick list'' of Upcoming Key Events. Section B of this report, the Site Summary section, provides Environmental Management performance data specifically organized to the pertinent Critical Success Factors and Success Indicators, and Science and Technology data in the context of the Critical Outcomes. The Site Summary demonstrates the various missions' overall progress against these strategic objectives. The information is presented in both narrative and graphical formats. The remaining sections provide performance data relative to each individual mission area (e.g., Waste Management, Spent Nuclear Fuels, etc.). The information provided in the Mission Area sections is at a level of greater detail than is presented in either the Executive Summary or
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5420413017272949, "perplexity": 6272.286823899885}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875146744.74/warc/CC-MAIN-20200227160355-20200227190355-00535.warc.gz"}
http://mathhelpforum.com/advanced-statistics/79639-help-poisson-other-type-questions.html
# Math Help - Help with Poisson and other type questions 1. ## Help with Poisson and other type questions Ive been trying to wrap my head around these questions for a few days now and i seem to be getting nowhere with it If anyone would take the time to help it would be greatly appreciated. Ive attached the pdf here because i thought it would be easiest. P.S. On another note does anyone know where i can find step by step derivations of mean and variance of negative binomial and Poisson random variables? Thanks again Attached Files 2. Originally Posted by bigrules_2 [snip] P.S. On another note does anyone know where i can find step by step derivations of mean and variance of negative binomial and Poisson random variables? Thanks again Poisson: http://www.mathhelpforum.com/math-he...tributiom.html Negative binomial: http://www.mathhelpforum.com/math-he...need-help.html contains what you need to do the calculations. 3. Originally Posted by bigrules_2 Ive been trying to wrap my head around these questions for a few days now and i seem to be getting nowhere with it If anyone would take the time to help it would be greatly appreciated. Ive attached the pdf here because i thought it would be easiest. [snip] This looks like an assignment that will contribute towards your final grade. MHF does not knowingly give help in such circumstances. Thread closed (pm me if you want to discuss this).
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8611655235290527, "perplexity": 392.492082460698}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440645151768.51/warc/CC-MAIN-20150827031231-00250-ip-10-171-96-226.ec2.internal.warc.gz"}
https://share.cocalc.com/share/fd64822bd6e169e1131379758d5f129f4e93ab39/sage_worksheets/ADS_Groups.sagews?viewer=share
Worksheets related to Applied Discrete Structures # Algebraic Structures using SageMath import sage.monoids.monoid M=Monoids().free(['a','b']) x=M('ab') x Error in lines 1-1 Traceback (most recent call last): File "/cocalc/lib/python2.7/site-packages/smc_sagews/sage_server.py", line 1044, in execute exec compile(block+'\n', '', 'single', flags=compile_flags) in namespace, locals File "", line 1, in <module> File "/cocalc/lib/python2.7/site-packages/smc_sagews/sage_salvus.py", line 3601, in displayhook _system_sys_displayhook(obj) File "sage/structure/sage_object.pyx", line 244, in sage.structure.sage_object.SageObject.__repr__ (build/cythonized/sage/structure/sage_object.c:2904) result = reprfunc() File "/ext/sage/sage-8.2_1604/local/lib/python2.7/site-packages/sage/monoids/indexed_free_monoid.py", line 117, in _repr_ return scalar_mult.join(P._repr_generator(g) + exp(v) for g,v in monomial) File "/ext/sage/sage-8.2_1604/local/lib/python2.7/site-packages/sage/monoids/indexed_free_monoid.py", line 117, in <genexpr> return scalar_mult.join(P._repr_generator(g) + exp(v) for g,v in monomial) ValueError: need more than 1 value to unpack Monoids().axioms() frozenset(['Associative', 'Unital']) ## Abelian Groups %html <p> Most standard algebraic stuctures are built into SageMath. Some do require importing a packages such <c>abelian_grps</c>. </p> Most standard algebraic stuctures are built into SageMath. Some do require importing a packages such Group. from sage.groups.group import Group import sage.groups.abelian_gps G=AbelianGroup([2,4]) G Multiplicative Abelian group isomorphic to C2 x C4 G Multiplicative Abelian group isomorphic to C2 x C4 G.is_cyclic() False G.cayley_table(names='elements') * 1 f1 f1^2 f1^3 f0 f0*f1 f0*f1^2 f0*f1^3 +---------------------------------------------------------------- 1| 1 f1 f1^2 f1^3 f0 f0*f1 f0*f1^2 f0*f1^3 f1| f1 f1^2 f1^3 1 f0*f1 f0*f1^2 f0*f1^3 f0 f1^2| f1^2 f1^3 1 f1 f0*f1^2 f0*f1^3 f0 f0*f1 f1^3| f1^3 1 f1 f1^2 f0*f1^3 f0 f0*f1 f0*f1^2 f0| f0 f0*f1 f0*f1^2 f0*f1^3 1 f1 f1^2 f1^3 f0*f1| f0*f1 f0*f1^2 f0*f1^3 f0 f1 f1^2 f1^3 1 f0*f1^2| f0*f1^2 f0*f1^3 f0 f0*f1 f1^2 f1^3 1 f1 f0*f1^3| f0*f1^3 f0 f0*f1 f0*f1^2 f1^3 1 f1 f1^2 sage.groups.abelian_gps.abelian_group? File: /ext/sage/sage-8.2_1604/local/lib/python2.7/site-packages/sage/groups/abelian_gps/abelian_group.py Docstring : Multiplicative Abelian Groups This module lets you compute with finitely generated Abelian groups of the form G = ZZ^r oplus ZZ_{k_1} oplus ... oplus ZZ_{k_t} It is customary to denote the infinite cyclic group ZZ as having order 0, so the data defining the Abelian group can be written as an integer vector vec{k} = (0, ..., 0, k_1, ..., k_t) where there are r zeroes and t non-zero values. To construct this Abelian group in Sage, you can either specify all entries of vec{k} or only the non-zero entries together with the total number of generators: sage: AbelianGroup([0,0,0,2,3]) Multiplicative Abelian group isomorphic to Z x Z x Z x C2 x C3 sage: AbelianGroup(5, [2,3]) Multiplicative Abelian group isomorphic to Z x Z x Z x C2 x C3 It is also legal to specify 1 as the order. The corresponding generator will be the neutral element, but it will still take up an index in the labelling of the generators: sage: G = AbelianGroup([2,1,3], names='g') sage: G.gens() (g0, 1, g2) Note that this presentation is not unique, for example ZZ_6 = ZZ_2 x ZZ_3. The orders of the generators vec{k}=(0,...,0,k_1,..., k_t) has previously been called invariants in Sage, even though they are not necessarily the (unique) invariant factors of the group. You should now use "gens_orders()" sage: J = AbelianGroup([2,0,3,2,4]); J Multiplicative Abelian group isomorphic to C2 x Z x C3 x C2 x C4 sage: J.gens_orders() # use this instead (2, 0, 3, 2, 4) sage: J.invariants() # deprecated (2, 0, 3, 2, 4) sage: J.elementary_divisors() # these are the "invariant factors" (2, 2, 12, 0) sage: for i in range(J.ngens()): ....: print((i, J.gen(i), J.gen(i).order())) # or use this form (0, f0, 2) (1, f1, +Infinity) (2, f2, 3) (3, f3, 2) (4, f4, 4) Background on invariant factors and the Smith normal form (according to section 4.1 of [C1]): An abelian group is a group A for which there exists an exact sequence ZZ^k rightarrow ZZ^ell rightarrow A rightarrow 1, for some positive integers k,ell with k<= ell. For example, a finite abelian group has a decomposition A = langle a_1rangle x ... x langle a_ellrangle , where ord(a_i)=p_i^{c_i}, for some primes p_i and some positive integers c_i, i=1,...,ell. GAP calls the list (ordered by size) of the p_i^{c_i} the *abelian invariants*. In Sage they will be called *invariants*. In this situation, k=ell and phi: ZZ^ell rightarrow A is the map phi(x_1,...,x_ell) = a_1^{x_1}...a_ell^{x_ell}, for (x_1,...,x_ell)in ZZ^ell. The matrix of relations M:ZZ^k rightarrow ZZ^ell is the matrix whose rows generate the kernel of phi as a ZZ-module. In other words, M=(M_{ij}) is a ell x ell diagonal matrix with M_{ii}=p_i^{c_i}. Consider now the subgroup Bsubset A generated by b_1 = a_1^{f_{1,1}}...a_ell^{f_{ell,1}}, ..., b_m = a_1^{f_{1,m}}...a_ell^{f_{ell,m}}. The kernel of the map phi_B: ZZ^m rightarrow B defined by phi_B(y_1,...,y_m) = b_1^{y_1}...b_m^{y_m}, for (y_1,...,y_m)in ZZ^m, is the kernel of the matrix F= ( begin{array}{cccc} f_{11} & f_{12} & ... & f_{1m} f_{21} & f_{22} & ... & f_{2m} vdots & & ddots & vdots f_{ell,1} & f_{ell,2} & ... & f_{ell,m} end{array} ), regarded as a map ZZ^mrightarrow (ZZ/p_1^{c_1}ZZ) x ... x (ZZ/p_ell^{c_ell}ZZ). In particular, Bcong ZZ^m/ker(F). If B=A then the Smith normal form (SNF) of a generator matrix of ker(F) and the SNF of M are the same. The diagonal entries s_i of the SNF S = diag[s_1,s_2,s_3, ... s_r,0,0,...0], are called *determinantal divisors* of F. where r is the rank. The {it invariant factors} of A are: s_1, s_2/s_1, s_3/s_2, ... s_r/s_{r-1}. Sage supports multiplicative abelian groups on any prescribed finite number n >= 0 of generators. Use the "AbelianGroup()" function to create an abelian group, and the "gen()" and "gens()" methods to obtain the corresponding generators. You can print the generators as arbitrary strings using the optional "names" argument to the "AbelianGroup()" function. EXAMPLE 1: We create an abelian group in zero or more variables; the syntax T(1) creates the identity element even in the rank zero case: sage: T = AbelianGroup(0,[]) sage: T Trivial Abelian group sage: T.gens() () sage: T(1) 1 EXAMPLE 2: An Abelian group uses a multiplicative representation of elements, but the underlying representation is lists of integer exponents: sage: F = AbelianGroup(5,[3,4,5,5,7],names = list("abcde")) sage: F Multiplicative Abelian group isomorphic to C3 x C4 x C5 x C5 x C7 sage: (a,b,c,d,e) = F.gens() sage: a*b^2*e*d a*b^2*d*e sage: x = b^2*e*d*a^7 sage: x a*b^2*d*e sage: x.list() [1, 2, 0, 1, 1] REFERENCES: * [C1] H. Cohen Advanced topics in computational number theory, Springer, 2000. * [C2] ----, A course in computational algebraic number theory, Springer, 1996. * [R] J. Rotman, An introduction to the theory of groups, 4th ed, Springer, 1995. Warning: Many basic properties for infinite abelian groups are not implemented. AUTHORS: * William Stein, David Joyner (2008-12): added (user requested) is_cyclic, fixed elementary_divisors. * David Joyner (2006-03): (based on free abelian monoids by David Kohel) * David Joyner (2006-05) several significant bug fixes * David Joyner (2006-08) trivial changes to docs, added random, fixed bug in how invariants are recorded * David Joyner (2006-10) added dual_group method * David Joyner (2008-02) fixed serious bug in word_problem * David Joyner (2008-03) fixed bug in trivial group case * David Loeffler (2009-05) added subgroups method * Volker Braun (2012-11) port to new Parent base. Use tuples for immutables. Rename invariants to gens_orders. %html <h2> Modular Arithmetic </h2> ## Modular Arithmetic Integers(5) Ring of integers modulo 5 a=2017 b=561 [q,r]=[a//b,a%b] [q,r] [3, 334] a=2017 b=561 print gcd(a,b) print xgcd(a,b) 1 (1, -173, 622) G=AbelianGroup(1,[14]) G.list() (1, f, f^2, f^3, f^4, f^5, f^6, f^7, f^8, f^9, f^10, f^11, f^12, f^13) for g in G: print str(g)+" has order "+str(g.order()) 1 has order 1 f has order 14 f^2 has order 7 f^3 has order 14 f^4 has order 7 f^5 has order 14 f^6 has order 7 f^7 has order 2 f^8 has order 7 f^9 has order 14 f^10 has order 7 f^11 has order 14 f^12 has order 7 f^13 has order 14 G= AdditiveAbelianGroup([2,2,2]) G.list() [(0, 0, 0), (1, 0, 0), (0, 1, 0), (1, 1, 0), (0, 0, 1), (1, 0, 1), (0, 1, 1), (1, 1, 1)] G.is_cyclic() False G= AdditiveAbelianGroup([23]); G Additive abelian group isomorphic to Z/23 G2=AdditiveAbelianGroup([2,4]);G2 Additive abelian group isomorphic to Z/2 + Z/4 + (0, 0) (0, 1) (0, 2) (0, 3) (1, 0) (1, 1) (1, 2) (1, 3) +-------------------------------------------------------- (0, 0)| (0, 0) (0, 1) (0, 2) (0, 3) (1, 0) (1, 1) (1, 2) (1, 3) (0, 1)| (0, 1) (0, 2) (0, 3) (0, 0) (1, 1) (1, 2) (1, 3) (1, 0) (0, 2)| (0, 2) (0, 3) (0, 0) (0, 1) (1, 2) (1, 3) (1, 0) (1, 1) (0, 3)| (0, 3) (0, 0) (0, 1) (0, 2) (1, 3) (1, 0) (1, 1) (1, 2) (1, 0)| (1, 0) (1, 1) (1, 2) (1, 3) (0, 0) (0, 1) (0, 2) (0, 3) (1, 1)| (1, 1) (1, 2) (1, 3) (1, 0) (0, 1) (0, 2) (0, 3) (0, 0) (1, 2)| (1, 2) (1, 3) (1, 0) (1, 1) (0, 2) (0, 3) (0, 0) (0, 1) (1, 3)| (1, 3) (1, 0) (1, 1) (1, 2) (0, 3) (0, 0) (0, 1) (0, 2) G2.is_cyclic() False a=[1878,1384,84,2021,784,1509,1740,1201,2363,1774,1865,33,1477,894,690,520,198,1349,1278,650] s =0 for t in a: s+=t s 23692 G=cartesian_product([Integers(32),Integers(27),Integers(25)]) def theta(x): return G((x%32,x%27,x%25)) theta(1000) (8, 1, 0) [theta(1878)+theta(1384),theta(1878+1384)] [(30, 22, 12), (30, 22, 12)] sum=G((0,0,0)) for t in a: sum+=theta(t) sum (12, 13, 17) l=list(sum) (7425*12 + 6400*13+ 7776* 17)%21600 2092 isum=crt([12,13,17],[32,27,25]) [isum,(s-isum)%(32*27*25)] [2092, 0] 32*27*25 21600 s =0 for t in a: s+=t s (s-isum)%21600 0 u=[crt([1,0,0],[32,27,25]),crt([0,1,0],[32,27,25]),crt([0,0,1],[32,27,25])] u [7425, 6400, 7776] G=DihedralGroup(5) gr=G.cayley_graph(simple="True") gr.plot(vertex_size=2) g=DiGraph({2:[1],1:[2,3]}) g.plot() triangle = SymmetricGroup(3) triangle.list() [(), (1,2), (1,2,3), (1,3,2), (2,3), (1,3)] triangle.cayley_graph(generators=[(1,2),(1,2,3)]).show() triangle.cayley_graph(generators=[(1,2),(1,3)]).show()
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.513709306716919, "perplexity": 5237.097315788591}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-34/segments/1534221212639.36/warc/CC-MAIN-20180817163057-20180817183057-00566.warc.gz"}
https://git.scc.kit.edu/Seitosh/Seitosh/-/blame/ba361d4041dddc11555f96d3457b254ab6dd1a8b/src/libs/libpsdxx/dpsdcomputer_cross_psd.cc
dpsdcomputer_cross_psd.cc 1.65 KB thomas.forbriger committed Jan 03, 2019 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 ``````/*! \file dpsdcomputer_cross_psd.cc * \brief compute cross power spectral density (implementation) * * ---------------------------------------------------------------------------- * * \author Thomas Forbriger * \date 03/01/2019 * * compute cross power spectral density (implementation) * * Copyright (c) 2019 by Thomas Forbriger (BFO Schiltach) * * ---- * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see . * ---- * * REVISIONS and CHANGES * - 03/01/2019 V1.0 Thomas Forbriger * * ============================================================================ */ `````` thomas.forbriger committed Jan 03, 2019 33 34 ``````#define PSDXX_DPSDCOMPUTER_CROSS_PSD_CC_VERSION \ "PSDXX_DPSDCOMPUTER_CROSS_PSD_CC V1.0" `````` thomas.forbriger committed Jan 03, 2019 35 `````` `````` thomas.forbriger committed Jan 03, 2019 36 ``````#include `````` thomas.forbriger committed Jan 03, 2019 37 `````` `````` thomas.forbriger committed Jan 03, 2019 38 ``````namespace psd { `````` thomas.forbriger committed Jan 03, 2019 39 `````` `````` thomas.forbriger committed Jan 06, 2019 40 41 42 `````` TDCISeries DPSDComputer::cross_psd(const TDISeries::Tcoc& s1, const TDISeries::Tcoc& s2) const `````` thomas.forbriger committed Jan 03, 2019 43 `````` { `````` thomas.forbriger committed Jan 07, 2019 44 45 `````` SpectralValues sv=this->processor(s1, s2, true, false); TDCISeries retval(sv.cpsd); `````` thomas.forbriger committed Jan 03, 2019 46 `````` return(retval); `````` thomas.forbriger committed Jan 06, 2019 47 `````` } // TDCISeries DPSDComputer::cross_psd(...) const `````` thomas.forbriger committed Jan 03, 2019 48 49 `````` } // namespace psd `````` thomas.forbriger committed Jan 03, 2019 50 51 `````` /* ----- END OF dpsdcomputer_cross_psd.cc ----- */``````
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9554081559181213, "perplexity": 4506.557075364882}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337537.25/warc/CC-MAIN-20221005042446-20221005072446-00005.warc.gz"}
https://worldwidescience.org/topicpages/s/system+optics+ii.html
#### Sample records for system optics ii 1. Okayama optical polarimetry and spectroscopy system (OOPS) II. Network-transparent control software. Science.gov (United States) Sasaki, T.; Kurakami, T.; Shimizu, Y.; Yutani, M. Control system of the OOPS (Okayama Optical Polarimetry and Spectroscopy system) is designed to integrate several instruments whose controllers are distributed over a network; the OOPS instrument, a CCD camera and data acquisition unit, the 91 cm telescope, an autoguider, a weather monitor, and an image display tool SAOimage. With the help of message-based communication, the control processes cooperate with related processes to perform an astronomical observation under supervising control by a scheduler process. A logger process collects status data of all the instruments to distribute them to related processes upon request. Software structure of each process is described. 2. Structure of multiphoton quantum optics. II. Bipartite systems, physical processes, and heterodyne squeezed states Science.gov (United States) dell'Anno, Fabio; de Siena, Silvio; Illuminati, Fabrizio 2004-03-01 Extending the scheme developed for a single mode of the electromagnetic field in the preceding paper [F. Dell’Anno, S. De Siena, and F. Illuminati, Phys. Rev. A 69, 033812 (2004)], we introduce two-mode nonlinear canonical transformations depending on two heterodyne mixing angles. They are defined in terms of Hermitian nonlinear functions that realize heterodyne superpositions of conjugate quadratures of bipartite systems. The canonical transformations diagonalize a class of Hamiltonians describing nondegenerate and degenerate multiphoton processes. We determine the coherent states associated with the canonical transformations, which generalize the nondegenerate two-photon squeezed states. Such heterodyne multiphoton squeezed states are defined as the simultaneous eigenstates of the transformed, coupled annihilation operators. They are generated by nonlinear unitary evolutions acting on two-mode squeezed states. They are non-Gaussian, highly nonclassical, entangled states. For a quadratic nonlinearity the heterodyne multiphoton squeezed states define two-mode cubic phase states. The statistical properties of these states can be widely adjusted by tuning the heterodyne mixing angles, the phases of the nonlinear couplings, as well as the strength of the nonlinearity. For quadratic nonlinearity, we study the higher-order contributions to the susceptibility in nonlinear media and we suggest possible experimental realizations of multiphoton conversion processes generating the cubic-phase heterodyne squeezed states. 3. UV-B and B-band Optical Flare Search in AR Lacertae, II Pegasi, and UX Arietis Star Systems Science.gov (United States) Vander Haagen, G. A. 2013-11-01 A high-cadence search was conducted on the known RS CVn-type flare stars AR Lac, II Peg, and UX Ari. Two optical flares were observed in the B-band on AR Lac at 5 milliseconds (ms) resolution for a rate of 0.04 fl/hr. Flare energy of the two B-band fast-flares ranged from 0.55 to 16.7 × 1033 ergs. The UV-B and B-band search of II Peg for 44.5 hours at 5 and 10 ms resolution and UV-B band search of UX Ari for 25.6 hours at 10 ms resolution detected no flare activity. 4. Optical absorption measurement system Science.gov (United States) Draggoo, Vaughn G.; Morton, Richard G.; Sawicki, Richard H.; Bissinger, Horst D. 1989-01-01 The system of the present invention contemplates a non-intrusive method for measuring the temperature rise of optical elements under high laser power optical loading to determine the absorption coefficient. The method comprises irradiating the optical element with a high average power laser beam, viewing the optical element with an infrared camera to determine the temperature across the optical element and calculating the absorption of the optical element from the temperature. 5. Flatland optics. II. Basic experiments. Science.gov (United States) Lohmann, A W; Wang, D; Pe'er, A; Friesem, A A 2001-05-01 In "Flatland optics: fundamentals" [J. Opt. Soc. Am. A 17, 1755 (2000)] we described the basic principles of two-dimensional (2D) optics and showed that a wavelength lambda in three-dimensional (3D) space (x,y,z) may appear in Flatland (x,z) as a wave with another wavelength, lambda = lambda/cosalpha. The tilt angle alpha can be modified by a 3D (Spaceland) individual who then is able to influence the 2D optics in a way that must appear to be magical to 2D Flatland individuals-in the spirit of E. A. Abbott's science fiction story [Flatland, a Romance of Many Dimensions, 6th ed. (Dover, New York, 1952)] of 1884. We now want to establish the reality or objectivity of the 2D wavelength lambda by some basic experiments similar to those that demonstrated roughly 200 years ago the wave nature of light. Specifically, we describe how to measure the 2D wavelength lambda by mean of five different arrangements that involve Young's biprism configuration, Talbot's self-imaging effect, measuring the focal length of a Fresnel zone plate, and letting light be diffracted by a double slit and by a grating. We also performed experiments with most of these arrangements. The results reveal that the theoretical wavelength, as predicted by our Flatland optics theory, does indeed coincide with the wavelength lambda as measured by Flatland experiments. Finally, we present an alternative way to understand Flatland optics in the spatial frequency domains of Flatland and Spaceland. 6. Future Optical Communications Systems Science.gov (United States) 2008-06-01 Figure 40: U.S. Multiple System Operator and Telco Carrier Network Infrastructure ......................... 43 Future Optical Communications Systems – vi...access networks deployed today. Future Optical Communications Systems – 43 An OIDA Forum Report Figure 40: U.S. Multiple System Operator and Telco 7. Optical timing receiver for the NASA Spaceborne Ranging System. Part II: high precision event-timing digitizer Energy Technology Data Exchange (ETDEWEB) Leskovar, Branko; Turko, Bojan 1978-08-01 Position-resolution capabilities of the NASA Spaceborne Laser Ranging System are essentially determined by the timeresolution capabilities of its optical timing receiver. The optical timing receiver consists of a fast photoelectric device; (e.g., photomultiplier or an avalanche photodiode detector), a timing discriminator, a high-precision event-timing digitizer, and a signal-processing system. The time-resolution capabilities of the receiver are determined by the photoelectron time spread of the photoelectric device, the time walk and resolution characteristics of the timing discriminator, and the resolution of the event-timing digitizer. It is thus necessary to evaluate available fast photoelectronic devices with respect to the time-resolution capabilities, and to develop a very low time walk timing discriminator and a high-resolution event-timing digitizer to be used in the high-resolution spaceborne laser ranging system receiver. This part of the report describes the development of a high precision event-timing digitizer. The event-timing digitizer is basically a combination of a very accurate high resolution real time digital clock and an interval timer. The timing digitizer is a high resolution multiple stop clock, counting the time up to 131 days in 19.5 ps increments. 8. Optical system design CERN Document Server Fischer, Robert F 2008-01-01 Honed for more than 20 years in an SPIE professional course taught by renowned optical systems designer Robert E. Fischer, Optical System Design, Second Edition brings you the latest cutting-edge design techniques and more than 400 detailed diagrams that clearly illustrate every major procedure in optical design. This thoroughly updated resource helps you work better and faster with computer-aided optical design techniques, diffractive optics, and the latest applications, including digital imaging, telecommunications, and machine vision. No need for complex, unnecessary mathematical derivations-instead, you get hundreds of examples that break the techniques down into understandable steps. For twenty-first century optical design without the mystery, the authoritative Optical Systems Design, Second Edition features: Computer-aided design use explained through sample problems Case studies of third-millennium applications in digital imaging, sensors, lasers, machine vision, and more New chapters on optomechanic... 9. Optical information-processing systems and architectures II; Proceedings of the Meeting, San Diego, CA, July 9-13, 1990 Science.gov (United States) Javidi, Bahram The present conference discusses topics in the fields of neural networks, acoustooptic signal processing, pattern recognition, phase-only processing, nonlinear signal processing, image processing, optical computing, and optical information processing. Attention is given to the optical implementation of an inner-product neural associative memory, optoelectronic associative recall via motionless-head/parallel-readout optical disk, a compact real-time acoustooptic image correlator, a multidimensional synthetic estimation filter, and a light-efficient joint transform optical correlator. Also discussed are a high-resolution spatial light modulator, compact real-time interferometric Fourier-transform processors, a fast decorrelation algorithm for permutation arrays, the optical interconnection of optical modules, and carry-free optical binary adders. 10. Stereoscopic optical viewing system Science.gov (United States) Tallman, C.S. 1986-05-02 An improved optical system which provides the operator with a stereoscopic viewing field and depth of vision, particularly suitable for use in various machines such as electron or laser beam welding and drilling machines. The system features two separate but independently controlled optical viewing assemblies from the eyepiece to a spot directly above the working surface. Each optical assembly comprises a combination of eye pieces, turning prisms, telephoto lenses for providing magnification, achromatic imaging relay lenses and final stage pentagonal turning prisms. Adjustment for variations in distance from the turning prisms to the workpiece, necessitated by varying part sizes and configurations and by the operator's visual accuity, is provided separately for each optical assembly by means of separate manual controls at the operator console or within easy reach of the operator. 11. Optical Complex Systems 2008 Science.gov (United States) Brun, Guillaume The Optical Complex Systems are more and more in the heart of various systems that industrial applications bring to everyday life. From environment up to spatial applications, OCS is also relevant in monitoring, transportation, robotics, life sciences, sub-marine, and even for agricultural purposes. 12. Detection of Extraterrestrial Life. Method II- Optical Rotatory Dispersion Science.gov (United States) 1963-01-01 The object of this study is to develop polarimetric methods to detect the presence of DNA (deoxyribonucleic acid) or its congeners in soil suspensions, and through these methods determine the existence of life (as known terrestrially) on other planets. The cotton region associated with optically active organic compounds is being used to detect and characterize the compounds above. An apparatus has been designed and assembled which can measure optical rotations in systems which strongly attenuate incident-polarized, monochromatic light. This instrument was used to measure the optical rotatory dispersion spectra of nucleosides, a polynucleotide, and proteins whose optical density at 260 microns approached 1.0. This work is discussed in the final report on Contract NASR-85, Detection of Extraterrestrial Life, Method II: Optical Rotatory Dispersion. Recent work in Melpar laboratories has reaffirmed these rotatory dispersion spectra. Based upon the analysis of the optical components associated with this apparatus, however, these measurements must be considered as qualitative rather than quantitative. The reason for this is discussed in greater detail subsequently in this report. In addition, an evaluation of the theoretical and instrumental aspects of making rotatory-dispersion measurements in the cotton region has resulted in a procedure for measuring optical rotation. 13. OPTICAL WIRELESS COMMUNICATION SYSTEM Directory of Open Access Journals (Sweden) JOSHUA L.Y. CHIENG 2016-02-01 Full Text Available The growing demand of bandwidth in this modern internet age has been testing the existing telecommunication infrastructures around the world. With broadband speeds moving towards the region of Gbps and Tbps, many researches have begun on the development of using optical wireless technology as feasible and future methods to the current wireless technology. Unlike the existing radio frequency wireless applications, optical wireless uses electromagnetic spectrums that are unlicensed and free. With that, this project aim to understand and gain better understanding of optical wireless communication system by building an experimental and simulated model. The quality of service and system performance will be investigated and reviewed. This project employs laser diode as the propagation medium and successfully transferred audio signals as far as 15 meters. On its quality of service, results of the project model reveal that the bit error rate increases, signal-to-noise ratio and quality factor decreases as the link distance between the transmitter and receiver increases. OptiSystem was used to build the simulated model and MATLAB was used to assist signal-to-noise ratio calculations. By comparing the simulated and experimental receiver’s power output, the experimental model’s efficiency is at 66.3%. Other than the system’s performance, challenges and factors affecting the system have been investigated and discussed. Such challenges include beam divergence, misalignment and particle absorption. 14. Optical systems in ergophthalmology Science.gov (United States) Kovalenko, Valentina; Besedovskaya, Valentina; Paloob, Tamara 1994-02-01 The important part of ergophthalmology is the problem of diagnosing and treatment of refraction errors, accommodation and visual disorders by means of the special optical systems. The using of our diagnostical approach helps to choose the right treatment strategy. Our therapeutical approach permits to normalize the muscle tonus and working capacity of eye accommodation apparatus and gives the possibility to obtain the stable positive results in treatment of the refraction amblyopia as well. 15. Final Report: Posttest Analysis of Omega II Optical Specimens Energy Technology Data Exchange (ETDEWEB) Newlander, C D; Fisher, J H 2007-01-30 Preliminary posttest analyses have been completed on optical specimens exposed during the Omega II test series conducted on 14 July 2006. The Omega Facility, located at the Laboratory for Laser Energetics (LLE) at the University of Rochester was used to produce X-ray environments through the interaction of intense pulsed laser radiation upon germanium-loaded silica aerogels. The optical specimen testing was supported by GH Systems through experiment design, pre- and post-test analyses, specimen acquisition, and overall technical experience. The test specimens were fabricated and characterized by Surface Optics Corporation (SOC), San Diego, CA and were simple protected gold coatings on silica substrates. Six test specimens were exposed, five filtered with thin beryllium foil filters, and one unfiltered which was exposed directly to the raw environment. The experimental objectives were: (1) demonstrate that tests of optical specimens could be performed at the Omega facility; (2) evaluate the use and survivability of beryllium foil filters as a function of thickness; (3) obtain damage data on optical specimens which ranged from no damage to damage; (4) correlate existing thermal response models with the damage data; (5) evaluate the use of the direct raw environment upon the specimen response and the ability/desirability to conduct sensitive optical specimen tests using the raw environment; and (6) initiate the development of a protocol for performing optical coatings/mirror tests. This report documents the activities performed by GH Systems in evaluating and using the environments provided by LLNL, the PUFFTFT analyses performed using those environments, and the calculated results compared to the observed and measured posttest data. 16. Optical system defect propagation in ABCD systems DEFF Research Database (Denmark) McKinley, W.G.; Yura, H.T.; Hanson, Steen Grüner 1988-01-01 We describe how optical system defects (tilt/jitter, decenter, and despace) propagate through an arbitrary paraxial optical system that can be described by an ABCD ray transfer matrix. A pedagogical example is given that demonstrates the effect of alignment errors on a typical optical system... 17. Optical Computing-Optical Components and Storage Systems Home; Journals; Resonance – Journal of Science Education; Volume 8; Issue 6. Optical Computing - Optical Components and Storage Systems ... Keywords. Advanced materials. optical switching. pulse shaping. optical storage device. high-performance computing. imaging; nanotechnology. photonics. telecommunications ... 18. Optimum Optical Systems CERN Document Server Autin, Bruno 2000-01-01 The characteristics of complex optical systems required by modern accelerators and storage rings result from the optimization of many variable function, usually of c2-type. The search for the minimum of the function is very sensitive to the initial values of the variables. The route towards a safe determination of the initial conditions is described. First, fully analytical solutions for thin lens models are derived using symbolic computing. Then, for thick lens systems, the equations become transcendent and numerical techniques are applied. It is shown that, under rather general conditions, the optimization converges towards a single solution located in the neighbourhood of the thin lens approximation. The example of betatron matching with a doublet is described, as well as the case of an isochronous periodic structure. 19. Intelligent Optical Systems Using Adaptive Optics Science.gov (United States) Clark, Natalie 2012-01-01 Until recently, the phrase adaptive optics generally conjured images of large deformable mirrors being integrated into telescopes to compensate for atmospheric turbulence. However, the development of smaller, cheaper devices has sparked interest for other aerospace and commercial applications. Variable focal length lenses, liquid crystal spatial light modulators, tunable filters, phase compensators, polarization compensation, and deformable mirrors are becoming increasingly useful for other imaging applications including guidance navigation and control (GNC), coronagraphs, foveated imaging, situational awareness, autonomous rendezvous and docking, non-mechanical zoom, phase diversity, and enhanced multi-spectral imaging. The active components presented here allow flexibility in the optical design, increasing performance. In addition, the intelligent optical systems presented offer advantages in size and weight and radiation tolerance. 20. The CLASS blazar survey - II. Optical properties NARCIS (Netherlands) Caccianiga, A; Marcha, MJ; Anton, S; Mack, KH; Neeser, MJ 2002-01-01 This paper presents the optical properties of the objects selected in the CLASS blazar survey. Because an optical spectrum is now available for 70 per cent of the 325 sources present in the sample, a spectral classification, based on the appearance of the emission/absorption lines, is possible. A 1. Ocean Optics Instrumentation Systems Data.gov (United States) Federal Laboratory Consortium — FUNCTION: Provides instrumentation suites for a wide variety of measurements to characterize the ocean’s optical environment. These packages have been developed to... 2. Optical performance monitoring in coherent optical OFDM systems. Science.gov (United States) Shieh, William; Tucker, Rodney S; Chen, Wei; Yi, Xingwen; Pendock, Graeme 2007-01-22 Optical performance monitoring is an indispensable feature for optical systems and networks. In this paper, we propose the concept of optical performance monitoring through channel estimation by receiver signal processing. We show that in coherent-optical-orthogonal-frequency-division- multiplexed (CO-OFDM) systems, critical optical system parameters including fiber chromatic dispersion, Q value, and optical signal-to-noise ratio (OSNR) can be accurately monitored without resorting to separate monitoring devices. 3. Optical properties of infrared FELs from the FELI Facility II Energy Technology Data Exchange (ETDEWEB) Saeki, K.; Okuma, S.; Oshita, E. [Free Electron Laser Institute, Osaka (Japan)] [and others 1995-12-31 The FELI Facility II has succeeded in infrared FEL oscillation at 1.91 {mu} m using a 68-MeV, 40-A electron beam from the FELI S-band linac in February 27, 1995. The FELI Facility II is composed of a 3-m vertical type undulator ({lambda}u=3.8cm, N=78, Km a x=1.4, gap length {ge}20mm) and a 6.72-m optical cavity. It can cover the wavelength range of 1-5{mu}m. The FELs can be delivered from the optical cavity to the diagnostics room through a 40-m evacuated optical pipeline. Wavelength and cavity length dependences of optical properties such as peak power, average power, spectrum width, FEL macropulse, FEL transverse profile are reported. 4. Atmospheric Optical Communication Systems. Science.gov (United States) 1981-02-01 found on vendor data sheets are the optical quality of the transmitter optics and the reflection characteristics of the receiver. These factors are...rotational absorpiton-line chaacer- isi-Is of the spec!Zc =oiecnLes.. Scaes-ng is dependenz on the -- ber density, size A stnabaon. shape, and indez - of...Simer and W!tte. These models are defned in terms of the atmosphe-_iO Press=-* and density, tanprsoe, and ozone and water vapor concenratios as a 5. Integrated Procurement Management System, Version II Science.gov (United States) Collier, L. J. 1985-01-01 Integrated Procurement Management System, Version II (IPMS II) is online/ batch system for collecting developing, managing and disseminating procurementrelated data at NASA Johnson Space Center. Portions of IPMS II adaptable to other procurement situations. 6. Astronomical optical interferometry, II: Astrophysical results Directory of Open Access Journals (Sweden) Jankov S. 2011-01-01 Full Text Available Optical interferometry is entering a new age with several ground- based long-baseline observatories now making observations of unprecedented spatial resolution. Based on a great leap forward in the quality and quantity of interferometric data, the astrophysical applications are not limited anymore to classical subjects, such as determination of fundamental properties of stars; namely, their effective temperatures, radii, luminosities and masses, but the present rapid development in this field allowed to move to a situation where optical interferometry is a general tool in studies of many astrophysical phenomena. Particularly, the advent of long-baseline interferometers making use of very large pupils has opened the way to faint objects science and first results on extragalactic objects have made it a reality. The first decade of XXI century is also remarkable for aperture synthesis in the visual and near-infrared wavelength regimes, which provided image reconstructions from stellar surfaces to Active Galactic Nuclei. Here I review the numerous astrophysical results obtained up to date, except for binary and multiple stars milliarcsecond astrometry, which should be a subject of an independent detailed review, taking into account its importance and expected results at microarcsecond precision level. To the results obtained with currently available interferometers, I associate the adopted instrumental settings in order to provide a guide for potential users concerning the appropriate instruments which can be used to obtain the desired astrophysical information. 7. Astronomical Optical Interferometry. II. Astrophysical Results Science.gov (United States) Jankov, S. 2011-12-01 Optical interferometry is entering a new age with several ground-based long-baseline observatories now making observations of unprecedented spatial resolution. Based on a great leap forward in the quality and quantity of interferometric data, the astrophysical applications are not limited anymore to classical subjects, such as determination of fundamental properties of stars; namely, their effective temperatures, radii, luminosities and masses, but the present rapid development in this field allowed to move to a situation where optical interferometry is a general tool in studies of many astrophysical phenomena. Particularly, the advent of long-baseline interferometers making use of very large pupils has opened the way to faint objects science and first results on extragalactic objects have made it a reality. The first decade of XXI century is also remarkable for aperture synthesis in the visual and near-infrared wavelength regimes, which provided image reconstructions from stellar surfaces to Active Galactic Nuclei. Here I review the numerous astrophysical results obtained up to date, except for binary and multiple stars milli-arcsecond astrometry, which should be a subject of an independent detailed review, taking into account its importance and expected results at micro-arcsecond precision level. To the results obtained with currently available interferometers, I associate the adopted instrumental settings in order to provide a guide for potential users concerning the appropriate instruments which can be used to obtain the desired astrophysical information. 8. Fiber-optic communication systems CERN Document Server Agrawal, Govind P 2010-01-01 This book provides a comprehensive account of fiber-optic communication systems. The 3rd edition of this book is used worldwide as a textbook in many universities. This 4th edition incorporates recent advances that have occurred, in particular two new chapters. One deals with the advanced modulation formats (such as DPSK, QPSK, and QAM) that are increasingly being used for improving spectral efficiency of WDM lightwave systems. The second chapter focuses on new techniques such as all-optical regeneration that are under development and likely to be used in future communication systems. All othe 9. Holographic Optics For Vision Systems Science.gov (United States) Freeman, Michael H. 1989-05-01 The human visual system is often equated to a photographic camera. This is a poor analogy because the differences are far greater than the similarities. The processing of the human visual system is complex and non-linear so that even optical transfer function concepts must be applied with caution. Holographic optics offers some extra degrees of freedom with respect to refractive optics. Unlike refractive optics, diffractive effects are not, in the first order, dependent on material and geometric shape and require no significant volume. On the other hand they may suffer from fractional efficiencies and strong wavelength dependencies. The Pilkington 'Diffrax' lens invented by the author is an example of a product which steers between the disadvantages and maximises the advantages to provide the world's first diffractive bifocal contact lens. Indications for other visual applications are not very propitious although time and development may show this to be incorrect. This paper will review the interaction between the preferences and antipathies of the human visual system and the optical effects of diffractive systems. Data.gov (United States) National Aeronautics and Space Administration — This proposal is for a flexible navigation system for deep space operations that does not require GPS measurements. The navigation solution is computed using an... 11. Chemo-optical micro-sensing systems NARCIS (Netherlands) Lambeck, Paul 1991-01-01 Chemo-optical microsensing systems, whether based on fiber optics or integrated optics, show promising prospects. The physical principles underlying chemo-optical waveguide sensors are discussed with the accent on linear evanescent field sensors. The role of the chemo-optical transduction layer is 12. Optical detection in microfluidic systems DEFF Research Database (Denmark) Mogensen, Klaus Bo; Kutter, Jörg Peter 2009-01-01 Optical detection schemes continue to be favoured for measurements in microfluidic systems. A selection of the latest progress mainly within the last two years is critically reviewed. Emphasis is on integrated solutions, such as planar waveguides, coupling schemes to the outside world, evanescent... 13. Handbook of Optical Systems, Volume 1, Fundamentals of Technical Optics Science.gov (United States) Gross, Herbert 2005-04-01 Herbert Gross, born in 1955, joined Carl Zeiss in 1982 after finishing his physics degree as specialist for optical design. Since 1995 he has been working as head of the department of optical design, while also teaching as a lecturer in Aalen and Lausanne. The new handbook is an intuitive, didactically elegant approach to the subject of optical systems and is not competed by any other work on the market. The selected board of authors, all reputed industrial experts, guarantee the timeliness of the well coordinated, coherent chapters. This first volume of the handbook introduces readers to the basics of geometrical and technical optics. For an understanding of optical systems, it is necessary to be familiar with the paraxial optics, the methods of ray tracing, the notations of geometrical optics and the description of optical systems. Very often the simple geometrical model is not sufficient to understand complex systems, therefore the wave optical model and the effects of light sources and receivers are discussed here too. Special components, such as gratings, prisms or aspherical lenses are described in detail to provide an understanding of modern complex systems. A short introduction into aberrations and the testing of optical systems allow readers to consider and control the quality of optical systems. 14. CP symmetry in optical systems CERN Document Server Dana, Brenda; Malomed, Boris A 2015-01-01 We introduce a model of a dual-core optical waveguide with opposite signs of the group-velocity-dispersion (GVD) in the two cores, and a phase-velocity mismatch between them. The coupler is embedded into an active host medium, which provides for the linear coupling of a gain-loss type between the two cores. The same system can be derived, without phenomenological assumptions, by considering the three-wave propagation in a medium with the quadratic nonlinearity, provided that the depletion of the second-harmonic pump is negligible. This linear system offers an optical realization of the charge-parity ($\\mathcal{CP}$) symmetry, while the addition of the intra-core cubic nonlinearity breaks the symmetry. By means of direct simulations and analytical approximations, it is demonstrated that the linear system generates expanding Gaussian states, while the nonlinear one gives rise to broad oscillating solitons, as well as a general family of stable stationary gap solitons. 15. Teaching Optics and Systems Engineering With Adaptive Optics Workbenches CERN Document Server Harrington, David; Hunter, Lisa; Max, Claire; Hoffmann, Mark; Pitts, Mark; Armstrong, J D 2010-01-01 Adaptive optics workbenches are fully functional optical systems that can be used to illustrate and teach a variety of concepts and cognitive processes. Four systems have been funded, designed and constructed by various institutions and people as part of education programs associated with the Center for Adaptive Optics, the Professional Development Program and the Institute for Science and Engineer Educators. Activities can range from first-year undergraduate explorations to professional level training. These workbenches have been used in many venues including the Center for Adaptive Optics AO Summer School, the Maui Community College hosted Akamai Maui Short Course, classrooms, training of new staff in laboratories and other venues. The activity content has focused on various elements of systems thinking, characterization, feedback and system control, basic optics and optical alignment as well as advanced topics such as phase conjugation, wave-front sensing and correction concepts and system design. The work... 16. Systems and methods for enhancing optical information Energy Technology Data Exchange (ETDEWEB) DeVore, Peter Thomas Setsuda; Chou, Jason T. 2018-01-02 An Optical Information Transfer Enhancer System includes a first system for producing an information bearing first optical wave that is impressed with a first information having a first information strength wherein the first optical wave has a first shape. A second system produces a second optical wave. An information strength enhancer module receives the first and said second optical waves and impresses the first optical wave upon the second optical wave via cross-phase modulation (XPM) to produce an information-strength-enhanced second optical wave having a second information strength that is greater than the first information strength of the first optical wave. Following a center-wavelength changer by an Optical Information Transfer Enhancer System improves its performance. 17. Optical Energy Transfer and Conversion System Science.gov (United States) Stone, William C. (Inventor); Hogan, Bartholomew P. (Inventor) 2015-01-01 An optical power transfer system comprising a fiber spooler, a fiber optic rotary joint mechanically connected to the fiber spooler, and an electrical power extraction subsystem connected to the fiber optic rotary joint with an optical waveguide. Optical energy is generated at and transferred from a base station through fiber wrapped around the spooler, through the rotary joint, and ultimately to the power extraction system at a remote mobility platform for conversion to another form of energy. 18. Optical fiber data transfer system Science.gov (United States) Mcmillan, S. H. 1988-01-01 This Phase 2 effort applies the results of Phase 1 to design and fabricate an optical slip ring system for a helicopter rotor blade/wind tunnel application. In this application, there are two assemblies: one on the rotating portion of the mechanical system, one on the stationary portion. The assembly on the rotating portion digitizes and encodes 128 transducer signals from various parts of the blade, and optically transfers data across the noncontacting coupling. Two complete identical independent channels are provided. On the stationary side, the signals are decoded and one channel is transmitted in digital form to a computer for recording and analysis. The second channel reconstructs the analog transducer signals for real time observation. In the opposite direction, eight signal channels enable control signals to be passed from the stationary to the rotating part of the system. Power to the rotor mounted electronics is supplied via power slip rings. The advantages of the optical over the traditional electro-mechanical slip ring method of data transfer across a rotating joint are long life, low-maintenance, immunity to crosstalk, and wider bandwidth. Successful completion of this effort demonstrated that this method is practical and reliable, and can be implemented under difficult conditions of available space, power, environment, and stringent performance and equipment life requirements. 19. Optical Manipulation System Using a Plurality of Optical Traps DEFF Research Database (Denmark) 2006-01-01 The present invention relates to an optical manipulation system (10) for generation of a plurality of optical traps for manipulation of micro-objects including nano-objects using electromagnetic radiation forces in a micro-object manipulation volume (14), the system comprising a spatially modulated... 20. Straightness measurements by use of a reflection confocal optical system. Science.gov (United States) Matsuda, K; Roy, M; O'Byrne, J W; Fekete, P W; Eiju, T; Sheppard, C J 1999-09-01 Straightness measurement is a very important technique in the field of mechanical engineering. A particular application for straightness measurement is high-accuracy machining on a diamond-turning lathe. We propose a novel, to our knowledge, optical method for measuring the straightness of motion, and its mathematical analysis is outlined. The technique is based on measurement of the lateral displacement of point images by use of reflection confocal optical systems. The advantages of this method are that (i) the lateral displacements in the direction of the two axes perpendicular to the optical axis can be measured, (ii) the rotation angles around all three axes can be measured, and (iii) reflection optical systems are more compact in length than are transmission optical systems. 1. Cleaning of contaminated XUV-optics at BESSY II CERN Document Server Eggenstein, F; Zeschke, T; Gudat, W 2001-01-01 Carbon contaminations as observed on XUV-optics can be removed by an in situ plasma discharge process. The method developed at BESSY is based on waterfree oxygen/argon mixture and avoids water contamination of the UHV-equipment. The radio frequency based plasma cleaning method has been used at several undulator beamlines at BESSY II with a gain in flux at the carbon K-edge. At the UE56-I-plane grating monochromator, a gain in flux up to a factor 20 is observed. No loss in flux has been observed across the whole energy ranges of the 'cleaned' beamlines. 2. 77 FR 27081 - II-VI, Incorporated, Infrared Optics-Saxonburg Division, Saxonburg, Pennsylvania; Notice of... Science.gov (United States) 2012-05-08 ... Employment and Training Administration II-VI, Incorporated, Infrared Optics--Saxonburg Division, Saxonburg... former workers of II-VI, Incorporated, Infrared Optics--Saxonburg Division, Saxonburg, Pennsylvania... employment related to the production of infrared and CO 2 laser optics, and related materials. The initial... 3. THE CONNECTIONS BETWEEN THE UV AND OPTICAL Fe ii EMISSION LINES IN TYPE 1 AGNs Energy Technology Data Exchange (ETDEWEB) Kovacević-Dojcinović, Jelena; Popović, Luka Č., E-mail: jkovacevic@aob.bg.ac.rs, E-mail: lpopovic@aob.bg.ac.rs [Astronomical Observatory, Volgina 7, 11060 Belgrade (Serbia) 2015-12-15 We investigate the spectral properties of the UV (λλ2650–3050 Å) and optical (λλ4000–5500 Å) Fe ii emission features in a sample of 293 Type 1 active galactic nuclei (AGNs) from the Sloan Digital Sky Survey database. We explore different correlations between their emission line properties, as well as the correlations with other emission lines from the spectral range. We find several interesting correlations and outline the most interesting results as follows. (i) There is a kinematical connection between the UV and optical Fe ii lines, indicating that the UV and optical Fe ii lines originate from the outer part of the broad line region, the so-called intermediate line region. (ii) The unexplained anticorrelations of the optical Fe ii equivalent width (EW Fe ii{sub opt}) versus EW [O iii] 5007 Å and EW Fe ii{sub opt} versus FWHM Hβ have not been detected for the UV Fe ii lines. (iii) The significant averaged redshift in the UV Fe ii lines, which is not present in optical Fe ii, indicates an inflow in the UV Fe ii emitting clouds, and probably their asymmetric distribution. (iv) Also, we confirm the anticorrelation between the intensity ratio of the optical and UV Fe ii lines and the FWHM of Hβ, and we find the anticorrelations of this ratio with the widths of Mg ii 2800 Å, optical Fe ii, and UV Fe ii. This indicates a very important role for the column density and microturbulence in the emitting gas. We discuss the starburst activity in high-density regions of young AGNs as a possible explanation of the detected optical Fe ii correlations and intensity line ratios of the UV and optical Fe ii lines. 4. Recent Advances in Photonic Devices for Optical Computing and the Role of Nonlinear Optics-Part II Science.gov (United States) Abdeldayem, Hossin; Frazier, Donald O.; Witherow, William K.; Banks, Curtis E.; Paley, Mark S. 2007-01-01 The twentieth century has been the era of semiconductor materials and electronic technology while this millennium is expected to be the age of photonic materials and all-optical technology. Optical technology has led to countless optical devices that have become indispensable in our daily lives in storage area networks, parallel processing, optical switches, all-optical data networks, holographic storage devices, and biometric devices at airports. This chapters intends to bring some awareness to the state-of-the-art of optical technologies, which have potential for optical computing and demonstrate the role of nonlinear optics in many of these components. Our intent, in this Chapter, is to present an overview of the current status of optical computing, and a brief evaluation of the recent advances and performance of the following key components necessary to build an optical computing system: all-optical logic gates, adders, optical processors, optical storage, holographic storage, optical interconnects, spatial light modulators and optical materials. 5. Optical fibre gas detections systems Science.gov (United States) Culshaw, Brian 2016-05-01 This tutorial review covers the principles of and prospects for fibre optic sensor technology in gas detection. Many of the potential benefits common to fibre sensor technology also apply in the context of gas sensing - notably long distance - many km - access to multiple remote measurement points; invariably intrinsic safety; access to numerous important gas species and often uniquely high levels of selectivity and/or sensitivity. Furthermore, the range of fibre sensor network architectures - single point, multiple point and distributed - enable unprecedented flexibility in system implementation. Additionally, competitive technologies and regulatory issues contribute to final application potential. 6. Laboratory testing & measurement on optical imaging systems CSIR Research Space (South Africa) Theron, B 2013-04-01 Full Text Available The purpose of the workshop was for participants to interactively discuss (with regard to optical imaging or optical imaging systems): Local end-user needs; What those needs imply for associated new & existing laboratory testing & measurement... 7. Optical CDMA system using bacteriorhodopsin for optical data storage Science.gov (United States) Bae; Yang; Jin; Lee; Park 1999-11-01 An optical CDMA (code division multiple access) system for the optical data storage using bacteriorhodopsin (BR) is reported as an application of the BR materials. The desired signal of multiple input can be recorded and reconstructed by use of orthogonal codes. An experimental setup is proposed and demonstrated. 8. Prototype system tests of the Belle II PXD DAQ system Energy Technology Data Exchange (ETDEWEB) Fleischer, Soeren; Gessler, Thomas; Kuehn, Wolfgang; Lange, Jens Soeren; Muenchow, David; Spruck, Bjoern [II. Physikalisches Institut, Justus-Liebig-Universitaet Giessen (Germany); Liu, Zhen' An; Xu, Hao; Zhao, Jingzhou [Institute of High Energy Physics, Chinese Academy of Sciences (China); Collaboration: II PXD Collaboration 2012-07-01 The data acquisition system for the Belle II DEPFET Pixel Vertex Detector (PXD) is designed to cope with a high input data rate of up to 21.6 GB/s. The main hardware component will be AdvancedTCA-based Compute Nodes (CN) equipped with Xilinx Virtex-5 FX70T FPGAs. The design for the third Compute Node generation was completed recently. The xTCA-compliant system features a carrier board and 4 AMC daughter boards. First test results of a prototype board will be presented, including tests of (a) The high-speed optical links used for data input, (b) The two 2 GB DDR2-chips on the board and (c) Output of data via ethernet, using UDP and TCP/IP with both hardware and software protocol stacks. 9. Knowledge-based environment for optical system design Science.gov (United States) Johnson, R. Barry 1991-01-01 Optical systems are extensively utilized by industry government and military organizations. The conceptual design engineering design fabrication and testing of these systems presently requires significant time typically on the order of 3-5 years. The Knowledge-Based Environment for Optical System Design (KB-OSD) Program has as its principal objectives the development of a methodology and tool(s) that will make a notable reduction in the development time of optical system projects reduce technical risk and overall cost. KB-OSD can be considered as a computer-based optical design associate for system engineers and design engineers. By utilizing artificial intelligence technology coupled with extensive design/evaluation computer application programs and knowledge bases the KB-OSD will provide the user with assistance and guidance to accomplish such activities as (i) develop system level and hardware level requirements from mission requirements (ii) formulate conceptual designs (iii) construct a statement of work for an RFP (iv) develop engineering level designs (v) evaluate an existing design and (vi) explore the sensitivity of a system to changing scenarios. The KB-OSD comprises a variety of computer platforms including a Stardent Titan supercomputer numerous design programs (lens design coating design thermal materials structural atmospherics etc. ) data bases and heuristic knowledge bases. An important element of the KB-OSD Program is the inclusion of the knowledge of individual experts in various areas of optics and optical system engineering. This knowledge is obtained by KB-OSD knowledge engineers performing 10. Propulsion Systems for Aircraft. Aerospace Education II. Instructional Unit II. Science.gov (United States) Elmer, James D. This curriculum guide accompanies another publication in the Aerospace Education II series entitled "Propulsion Systems for Aircraft." The guide includes specific guidelines for teachers on each chapter in the textbook. Suggestions are included for objectives (traditional and behavioral), suggested outline, orientation, suggested key… 11. Optical supermicrosensor responses for simple recognition and sensitive removal of Cu (II) Ion target. Science.gov (United States) El-Safty, Sherif A; Ismail, Adel A; Shahat, Ahmed 2011-02-15 The field of optical chemosensor technology demands a simple yet general design for fast, sensitive, selective, inexpensive, and specific recognition of a broad range of toxic metal ions. The suitable accommodation of chromogenic receptors onto ordered porous carriers have led to selective and sensitive chemosensors of target species. In this study, we offer real evidence on the potential use of two- and three-dimensional (2D and 3D) ordered supermicroporous monoliths as selective shape and size carriers for immobilizing the chromogenic probe. Among all the chemosensors, 3D supermicropore has exhibited easy accessibility of target ions, such as ion transports and high affinity responses of receptor-metal analyte binding events. This leads to an optical color signal that is easily generated and transduced even at trace levels of Cu(II) target ions. The supermicrosensors have shown the ability to create Cu(II) ion-sensing responses up to nanomolar concentrations (∼10(-9) mol/dm(3)) with rapid response time (in the order of seconds). Supermicrosensors have the ability to create easily modified sensing systems with multiple regeneration/reuse cycles of sensing systems of Cu(II) analytes. The simple treatment using ClO(4)(-) anion as a stripping agent has removed effectively the Cu(II) ions and formed a "metal-free" probe surface. The supermicrosensors have exhibited the specificity behavior permitting Cu(II) ion-selective determination in real-life samples, such as in wastewater, despite the presence of active component species. Extensive analytical results indicate that the use of the supermicrosensor as Cu(II) ion strips for field screening can be a time- and cost-alternative tool to current effective laboratory assays. Copyright © 2010 Elsevier B.V. All rights reserved. 12. Analysis of optical amplifier noise in coherent optical communication systems with optical image rejection receivers DEFF Research Database (Denmark) Jørgensen, Bo Foged; Mikkelsen, Benny; Mahon, Cathal J. 1992-01-01 A detailed theoretical analysis of optical amplifier noise in coherent optical communication systems with heterodyne receivers is presented. The analysis quantifies in particular how optical image rejection receiver configurations reduce the influence of optical amplifier noise on system performa......A detailed theoretical analysis of optical amplifier noise in coherent optical communication systems with heterodyne receivers is presented. The analysis quantifies in particular how optical image rejection receiver configurations reduce the influence of optical amplifier noise on system...... performance. Two types of optical image rejection receivers are investigated: a novel, all-optical configuration and the conventional, microwave-based configuration. The analysis shows that local oscillator-spontaneous emission beat noise (LO-SP), signal-spontaneous emission beat noise (S-SP), and spontaneous......-spontaneous beat noise (SP-SP) can all be reduced by 3 dB, thereby doubling the dynamic range of the optical amplifier. A 2.5-dB improvement in dynamic range has been demonstrated experimentally with the all-optical image rejection configuration. The implications of the increased dynamic range thus obtained... 13. New Photonic System for Optical Packet Switching Directory of Open Access Journals (Sweden) F. Rudge Barbosa 2003-08-01 Full Text Available Fast optical switching (ms timebase is realized by using a RF frequency tone inserted in the optical packet that carries a digital payload. By using a highly selective RF filtering for optical packet header frequency recognition, we have obtained excellent performance in optical switching function.. The RF header is detected at optical node input, and signals the node switching control, which instantly directs the packet to a prescribed output. No electronic processing of the digital payload is performed. The optical circuit is noise-free, has very low crosstalk, and is extremely selective in header frequency detection. BER measurements for payload consistently yield figures as low as 10-12 . This system is applicable to optical metropolitan and access networks, and is fully compatible with DWDM systems. 14. Capillary Electrophoresis - Optical Detection Systems Energy Technology Data Exchange (ETDEWEB) Sepaniak, M. J. 2001-08-06 Molecular recognition systems are developed via molecular modeling and synthesis to enhance separation performance in capillary electrophoresis and optical detection methods for capillary electrophoresis. The underpinning theme of our work is the rational design and development of molecular recognition systems in chemical separations and analysis. There have been, however, some subtle and exciting shifts in our research paradigm during this period. Specifically, we have moved from mostly separations research to a good balance between separations and spectroscopic detection for separations. This shift is based on our perception that the pressing research challenges and needs in capillary electrophoresis and electrokinetic chromatography relate to the persistent detection and flow rate reproducibility limitations of these techniques (see page 1 of the accompanying Renewal Application for further discussion). In most of our work molecular recognition reagents are employed to provide selectivity and enhance performance. Also, an emerging trend is the use of these reagents with specially-prepared nano-scale materials. Although not part of our DOE BES-supported work, the modeling and synthesis of new receptors has indirectly supported the development of novel microcantilevers-based MEMS for the sensing of vapor and liquid phase analytes. This fortuitous overlap is briefly covered in this report. Several of the more significant publications that have resulted from our work are appended. To facilitate brevity we refer to these publications liberally in this progress report. Reference is also made to very recent work in the Background and Preliminary Studies Section of the Renewal Application. 15. Optical Potential Field Mapping System Science.gov (United States) Reid, Max B. (Inventor) 1996-01-01 The present invention relates to an optical system for creating a potential field map of a bounded two dimensional region containing a goal location and an arbitrary number of obstacles. The potential field mapping system has an imaging device and a processor. Two image writing modes are used by the imaging device, electron deposition and electron depletion. Patterns written in electron deposition mode appear black and expand. Patterns written in electron depletion mode are sharp and appear white. The generated image represents a robot's workspace. The imaging device under processor control then writes a goal location in the work-space using the electron deposition mode. The black image of the goal expands in the workspace. The processor stores the generated images, and uses them to generate a feedback pattern. The feedback pattern is written in the workspace by the imaging device in the electron deposition mode to enhance the expansion of the original goal pattern. After the feedback pattern is written, an obstacle pattern is written by the imaging device in the electron depletion mode to represent the obstacles in the robot's workspace. The processor compares a stored image to a previously stored image to determine a change therebetween. When no change occurs, the processor averages the stored images to produce the potential field map. 16. Silicon retina for optical tracking systems Science.gov (United States) Strohbehn, K.; Jenkins, R. E.; Sun, X.; Andreou, A. G. 1993-01-01 There are a host of position sensors, such as quadcells and CCD's, which are candidates for detecting optical position errors and providing error signals for a mirror positioning loop. We are developing a novel, very high bandwidth, biologically inspired position sensor for optical position tracking systems. We present recent test results and design issues for the use of biologically inspired silicon retinas for spaceborne optical position tracking systems. 17. Optics Toolbox: An Intelligent Relational Database System For Optical Designers Science.gov (United States) Weller, Scott W.; Hopkins, Robert E. 1986-12-01 Optical designers were among the first to use the computer as an engineering tool. Powerful programs have been written to do ray-trace analysis, third-order layout, and optimization. However, newer computing techniques such as database management and expert systems have not been adopted by the optical design community. For the purpose of this discussion we will define a relational database system as a database which allows the user to specify his requirements using logical relations. For example, to search for all lenses in a lens database with a F/number less than two, and a half field of view near 28 degrees, you might enter the following: FNO expert system as a program which contains expert knowledge, can ask intelligent questions, and can form conclusions based on the answers given and the knowledge which it contains. Most expert systems store this knowledge in the form of rules-of-thumb, which are written in an English-like language, and which are easily modified by the user. An example rule is: IF require microscope objective in air and require NA > 0.9 THEN suggest the use of an oil immersion objective The heart of the expert system is the rule interpreter, sometimes called an inference engine, which reads the rules and forms conclusions based on them. The use of a relational database system containing lens prototypes seems to be a viable prospect. However, it is not clear that expert systems have a place in optical design. In domains such as medical diagnosis and petrology, expert systems are flourishing. These domains are quite different from optical design, however, because optical design is a creative process, and the rules are difficult to write down. We do think that an expert system is feasible in the area of first order layout, which is sufficiently diagnostic in nature to permit useful rules to be written. This first-order expert would emulate an expert designer as he interacted with a customer for the first time: asking the right questions, forming 18. All-optical switching and limiting properties of a Ru (II) Schiff-base complex for nonlinear optical applications Science.gov (United States) Manjunatha, K. B.; Rajarao, Ravindra; Umesh, G.; Ramachandra Bhat, B.; Poornesh, P. 2017-08-01 A salen-based ruthenium (Ru) (II) complex was synthesized for possible use in nonlinear optical device applications. The Ru complex was doped in a polymer matrix to fabricate films using a low-cost spin-coating technique. The third-order nonlinear optical parameters of the complex were investigated by Z-scan and degenerate four-wave mixing techniques. The study reveals two-order enhancement of third-order optical susceptibility χ (3) and exhibits superior limiting capability due to a reverse saturable absorption process. All-optical switching action for the films indicates that the sample can function as an optical inverter or a NOT gate. Hence, the Ru (II) Schiff-base complex materializes as a possible candidate for use in nonlinear optical devices. 19. ADAPTIVE OPTICAL SYSTEMS LIGHTING EQUIPMENT VEHICLES OpenAIRE S. P. Sernov; D. V. Balokhonov; T. V. Kolontaeva; A. V. Zhuravok 2011-01-01 This article describes the main principles of implementation of modern adaptive signal lighting equipment of vehicles, provides an analysis of optical systems are used, the necessity of the use of LEDs. We present the design of adaptive optical system, rear combination LED lamp of a vehicle with several levels of intensity, we discuss the algorithm of its work in different modes. 20. Electro-optical System Measures Aircraft Deflections Science.gov (United States) Fodale, B.; Hampton, H. R.; Seymour, H. R.; Deangelis, V. M. 1986-01-01 In-flight deflections of aircraft surfaces are measured accurately over wide range of airspeeds and attitudes by electro-optical instrumentation system. Deflections caused by aerodynamic and acceleration forces are measured on wings, stabilizers, helicopter rotors, and other surfaces. Electro-optical system replaces 70-millimeter motor-driven camera previously used to observe deflections. 1. Automatic polarization control in optical sampling system Science.gov (United States) Zhao, Zhao; Yang, Aiying; Feng, Lihui 2015-08-01 In an optical sampling system for high-speed optical communications, polarization controlling is one of the most important parts of the system, regardless of nonlinear optical sampling or linear optical sampling. A simple method based on variance calculation of sampled data is proposed in this paper to tune the wave plates in a motor-driven polarization controller. In the experiment, an optical sampling system base on SFG in PPLN is carried for a 10Gbit/s or beyond optical data signal. The results demonstrate that, with the proposed method, the error of estimated Q factor from the sampled data is least, and the tuning time of optimized polarization state is less than 30 seconds with the accuracy of +/-1°. 2. Combined x-ray/electron/optical Monte Carlo code based on PENELOPE and DETECT-II Science.gov (United States) Badano, Aldo; Sempau, Josep; Boswell, Jonathan S. 2005-04-01 We describe MANTIS (Monte carlo x-rAy electroN opTical Imaging Simulation), a tool for simulating imaging systems that tracks x rays, electrons, and optical photons in the same geometric model. The x-ray and electron transport and involved physics models are from the PENELOPE package and include elastic and inelastic scattering, and bremsstrahlung from 100 eV to 1 GeV. The optical transport and corresponding physics models are from DETECT-II and include Fresnel refraction and reflection at material boundaries, bulk absorption and scattering. X rays are generated using the flexible source description from PENELOPE. When x rays or electrons interact and deposit energy in the scintillator, the code generates a number of optical quanta at that location, according to a model for the conversion process. The optical photons are then tracked until they reach an absorption event that in some cases contributes to the electronic signal. We demonstrate the capabilities of the new tool with respect to x-ray source, object to be imaged, and detector models. Of particular importance is the improved geometric description of structured phosphors that can handle tilted columns in needle-like phosphor screens. Examples of the simulation output with respect to signal blur and pulse-height distributions of the scintillation light are discussed and compared with previously published experimental results. 3. Is optical Fe II emission related to the soft X-ray properties of quasars? Science.gov (United States) Wilkes, Belinda J.; Elvis, Martin; McHardy, Ian 1987-10-01 Radio-quiet quasars generally show broad, blended multiplets of Fe II emission in their optical and UV spectra. Radio-loud quasars also show UV Fe II emission, but their optical Fe II emission is generally weaker. No satisfactory theory connecting the generation of Fe II and radio emission has been found to explain this effect. A second, well-established distinction between the two clases of quasar is in their X-ray properties: radio-loud quasars are more X-ray luminous, and recent results have shown that they also have systematically flatter soft X-ray slopes. Here it is proposed that the second effect causes the first; i.e., that the primary factor controlling the optical Fe II emission is the soft X-ray spectrum. This proposition is supported by X-ray and optical data for nine quasars, which shows a correlation between the soft X-ray slope and the strength of the optical Fe II emission. One of these quasars (1803+676) is radio-quiet, and yet its optical spectrum shows no evidence for Fe II emission. This quasar is also unusual in that it has a flat X-ray spectrum. This further supports the proposal that the X-ray spectrum is important in determining the relative strengths of UV and optical Fe II emission. 4. Intraoperative OCT Imaging of the Argus II Retinal Prosthesis System. Science.gov (United States) Rachitskaya, Aleksandra V; Yuan, Alex; Marino, Meghan J; Reese, Jamie; Ehlers, Justis P 2016-11-01 Optimal placement of the Argus II Retinal Prosthesis System (Second Sight Medical Products, Sylmar, CA) is critical. Intraoperative optical coherence tomography (OCT) allows for intrasurgical visualization and confirmation of array placement. In this study, two different OCT systems were evaluated to assess the feasibility and utility of this technology during Argus II surgery. Intraoperative OCT was performed on five patients undergoing Argus II implantation at Cole Eye Institute from June 2015 to July 2016. The EnVisu portable OCT (Bioptigen, Morrisville, NC) and microscope-integrated RESCAN 700 (Zeiss, Oberkochen, Germany) intraoperative OCT systems were utilized. The EnVisu was used in three patients and the RESCAN 700 in three of the five patients. Following array tacking, intraoperative OCT was performed over the entire array including the edges and tack. Intraoperative OCT allowed for visualization of the array/retina interface. Microscope integration of the OCT system facilitated ease of focusing, real-time feedback, surgeon-directed OCT scanning to the areas of interest, and enhanced image quality at points of interest. Intraoperative imaging of the Argus II electrode array is feasible and provides information about electrode array-retina interface and distance to help guide a surgeon. Microscope integration of OCT appears to provide an optimal and efficient approach to intraoperative OCT during Argus II array placement. [Ophthalmic Surg Lasers Imaging Retina. 2016;47:999-1003.]. Copyright 2016, SLACK Incorporated. 5. LHC II system sensitivity to magnetic fluids CERN Document Server 2005-01-01 Experiments have been designed to reveal the influences of ferrofluid treatment and static magnetic field exposure on the photosynthetic system II, where the light harvesting complex (LHC II) controls the ratio chlorophyll a/ chlorophyll b (revealing, indirectly, the photosynthesis rate). Spectrophotometric measurement of chlorophyll content revealed different influences for relatively low ferrofluid concentrations (10-30 mul/l) in comparison to higher concentrations (70-100 mul/l). The overlapped effect of the static magnetic field shaped better the stimulatory ferrofluid action on LHC II system in young poppy plantlets. 6. Integrated Optical Interconnect Architectures for Embedded Systems CERN Document Server Nicolescu, Gabriela 2013-01-01 This book provides a broad overview of current research in optical interconnect technologies and architectures. Introductory chapters on high-performance computing and the associated issues in conventional interconnect architectures, and on the fundamental building blocks for integrated optical interconnect, provide the foundations for the bulk of the book which brings together leading experts in the field of optical interconnect architectures for data communication. Particular emphasis is given to the ways in which the photonic components are assembled into architectures to address the needs of data-intensive on-chip communication, and to the performance evaluation of such architectures for specific applications.   Provides state-of-the-art research on the use of optical interconnects in Embedded Systems; Begins with coverage of the basics for high-performance computing and optical interconnect; Includes a variety of on-chip optical communication topologies; Features coverage of system integration and opti... 7. Performance of the upgraded LTP-II at the ALS Optical Metrology Laboratory Energy Technology Data Exchange (ETDEWEB) Advanced Light Source; Yashchuk, Valeriy V; Kirschman, Jonathan L.; Domning, Edward E.; McKinney, Wayne R.; Morrison, Gregory Y.; Smith, Brian V.; Yashchuk, Valeriy V. 2008-07-14 The next generation of synchrotrons and free electron laser facilities requires x-ray optical systems with extremely high performance, generally of diffraction limited quality. Fabrication and use of such optics requires adequate, highly accurate metrology and dedicated instrumentation. Previously, we suggested ways to improve the performance of the Long Trace Profiler (LTP), a slope measuring instrument widely used to characterize x-ray optics at long spatial wavelengths. The main way is use of a CCD detector and corresponding technique for calibration of photo-response non-uniformity [J. L. Kirschman, et al., Proceedings of SPIE 6704, 67040J (2007)]. The present work focuses on the performance and characteristics of the upgraded LTP-II at the ALS Optical Metrology Laboratory. This includes a review of the overall aspects of the design, control system, the movement and measurement regimes for the stage, and analysis of the performance by a slope measurement of a highly curved super-quality substrate with less than 0.3 microradian (rms)slope variation. 8. Polishing optical fibers for the D0 ICD in Run II Energy Technology Data Exchange (ETDEWEB) Elizabeth Gallas and Jia Li 1998-12-01 The active element in the Run II D0 Inter Cryostat Detector (ICD) is an array of scintil- lator tiles. Charged and neutral particles produce light in the tiles which is transported to the photodetection system along optical fiber pathways. In general, building a tile/fiber detector requires very conscientious technical support and a high degree of quality control. Polishing fibers is one of the most delicate of tasks involved. This note describes methods used by the ICD group to polish the ends of fibers. These methods may be used in the ICD production as well as in prototype development. 9. Reciprocity-enhanced optical communication through atmospheric turbulence - part II: communication architectures and performance Science.gov (United States) Puryear, Andrew L.; Shapiro, Jeffrey H.; Parenti, Ronald R. 2012-10-01 Free-space optical communication provides rapidly deployable, dynamic communication links that are capable of very high data rates compared with those of radio-frequency systems. As such, free-space optical communication is ideal for mobile platforms, for platforms that require the additional security afforded by the narrow divergence of a laser beam, and for systems that must be deployed in a relatively short time frame. In clear-weather conditions the data rate and utility of free-space optical communication links are primarily limited by fading caused by micro-scale atmospheric temperature variations that create parts-per-million refractive-index fluctuations known as atmospheric turbulence. Typical communication techniques to overcome turbulence-induced fading, such as interleavers with sophisticated codes, lose viability as the data rate is driven higher or the delay requirement is driven lower. This paper, along with its companion [J. H. Shapiro and A. Puryear, "Reciprocity-Enhanced Optical Communication through Atmospheric Turbulence-Part I: Reciprocity Proofs and Far-Field Power Transfer"], present communication systems and techniques that exploit atmospheric reciprocity to overcome turbulence which are viable for high data rate and low delay requirement systems. Part I proves that reciprocity is exhibited under rather general conditions, and derives the optimal power-transfer phase compensation for far-field operation. The Part II paper presents capacity-achieving architectures that exploit reciprocity to overcome the complexity and delay issues that limit state-of-the art free-space optical communications. Further, this paper uses theoretical turbulence models to determine the performance—delay, throughput, and complexity—of the proposed architectures. 10. A system for obtaining an optical spectrum DEFF Research Database (Denmark) 2013-01-01 The present invention relates to a system for obtaining an optical spectrum 648 of analytes in a fluid sample, wherein a porous filter 602a, 602b is arranged so that the fluid sample may be placed onto a first region of the porous filter, and a SERS-active material 610a, 610b having a SERS...... the second region. Thereby, the porous filter enables that the fluid sample is filtered so that only sufficiently small entities in the fluid sample reach the second region where they may be probed so that an optical spectrum related to the analytes in the filtered sample may be obtained. The optical system...... also comprises a light source 634, a light detector 642, and the optical analysis system is arranged for obtaining the optical spectrum 648 of the analytes adjacent to the SERS-active material 610a, 610b exclusively from the second region of the porous filter 602a, 602b.... 11. Particle Systems and PDEs II CERN Document Server Soares, Ana 2015-01-01 This book focuses on mathematical problems concerning different applications in physics, engineering, chemistry and biology. It covers topics ranging from interacting particle systems to partial differential equations (PDEs), statistical mechanics and dynamical systems. The purpose of the second meeting on Particle Systems and PDEs was to bring together renowned researchers working actively in the respective fields, to discuss their topics of expertise and to present recent scientific results in both areas. Further, the meeting was intended to present the subject of interacting particle systems, its roots in and impacts on the field of physics, and its relation with PDEs to a vast and varied public, including young researchers. The book also includes the notes from two mini-courses presented at the conference, allowing readers who are less familiar with these areas of mathematics to more easily approach them. The contributions will be of interest to mathematicians, theoretical physicists and other researchers... 12. Mobile based optical form evaluation system Directory of Open Access Journals (Sweden) Asım Sinan YÜKSEL 2016-05-01 Full Text Available Optical forms that contain multiple-choice answers are widely used both for electing students and evaluating student achievements in education systems in our country and worldwide. Optical forms are evaluated by employing optical mark recognition techniques through optical readers. High cost of these machines, limited access to them, long waiting time for evaluation results make the process hard for educationists working in cities or countries. In this study, a mobile application was developed for the educationists who own mobile phones or tablets for the purpose of evaluating students' answer sheets quickly and independent of location and optical readers. Optical form recognition, reading and evaluation processes are done on the image of student's answer sheet that is taken with the mobile phone or tablet of educationist. The Android based mobile application that we developed has a user-friendly interface, high success rate and is the first of our knowledge application that operates on mobile platforms in this field. 13. Fiber optical parametric amplifiers in optical communication systems DEFF Research Database (Denmark) Marhic, Michel E.; Andrekson, Peter A.; Petropoulos, Periklis 2015-01-01 The prospects for using fiber optical parametric amplifiers (OPAs) in optical communication systems are reviewed. Phase-insensitive amplifiers (PIAs) and phase-sensitive amplifiers (PSAs) are considered. Low-penalty amplification at/or near 1 Tb/s has been achieved, for both wavelength- and time......-modulated signals. A PSA with 1.1-dB noise figure has been demonstrated, and preliminary wavelength-division multiplexing experiments have been performed with PSAs. 512Gb/s have been transmitted over 6,000km by periodic phase conjugation. Simulations indicate that PIAs could reach data rate x reach products... 14. Optical investigations and control of spindynamics in Mn doped II-VI quantum dots; Optische Untersuchung und Kontrolle der Spindynamik in Mn dotierten II-VI Quantenpunkten Energy Technology Data Exchange (ETDEWEB) Schmidt, Thomas 2009-05-13 The present thesis deals with the spin of charge carriers confined in CdSe/ZnSe quantum dots (QDs) closely linked to the polarization of emitted photons. II-VI material systems can be adequately mixed with the B-group element manganese. Such semimagnetic nanostructures offer a number of characteristic optical and electronic features. This is caused by an exchange interaction between the spin of optically excited carriers and the 3d electrons of the Mn ions. Within the framework of this thesis addressing of well defined spin states was realized by optical excitation of charge carriers. The occupation of different spin states was detected by the degree of polarization of the emitted photoluminescence (PL) light. For that purpose different optical methods of time-resolved and time-integrated spectroscopy as well as investigations in magnetic fields were applied. (orig.) 15. Tunable Optical Polymer Systems (TOPS) Science.gov (United States) 2001-05-01 Rochester AROMURI Background and Introduction • Nature of coloration - pigmentary colors due to absorption - structural colors due to interference...and theoretical characterization of color switching phenomena: Extensive material~ characterization is planned, including studies of the...multifunctional optical polymers during operation in model devices. Various types of color change phenomena will be exploited including those which rely on 16. Nonlinear Mixing in Optical Multicarrier Systems Science.gov (United States) Hameed, Mahmood Abdul Although optical fiber has a vast spectral bandwidth, efficient use of this bandwidth is still important in order to meet the ever increased capacity demand of optical networks. In addition to wavelength division multiplexing, it is possible to partition multiple low-rate subcarriers into each high speed wavelength channel. Multicarrier systems not only ensure efficient use of optical and electrical components, but also tolerate transmission impairments. The purpose of this research is to understand the impact of mixing among subcarriers in Radio-Over-Fiber (RoF) and high speed optical transmission systems, and experimentally demonstrate techniques to minimize this impact. We also analyze impact of clipping and quantization on multicarrier signals and compare bandwidth efficiency of two popular multiplexing techniques, namely, orthogonal frequency division multiplexing (OFDM) and Nyquist modulation. For an OFDM-RoF system, we present a novel technique that minimizes the RF domain signal-signal beat interference (SSBI), relaxes the phase noise limit on the RF carrier, realizes the full potential of optical heterodyne-based RF carrier generation, and increases the performance-to-cost ratio of RoF systems. We demonstrate a RoF network that shares the same RF carrier for both downlink and uplink, avoiding the need of an additional RF oscillator in the customer unit. For multi-carrier optical transmission, we first experimentally compare performance degradations of coherent optical OFDM and single-carrier Nyquist pulse modulated systems in a nonlinear environment. We then experimentally evaluate SSBI compensation techniques in the presence of semiconductor optical amplifier (SOA) induced nonlinearities for a multicarrier optical system with direct detection. We show that SSBI contamination can be significantly reduced from the data signal when the carrier-to-signal power ratio is sufficiently low. 17. ADAPTIVE OPTICAL SYSTEMS LIGHTING EQUIPMENT VEHICLES Directory of Open Access Journals (Sweden) S. P. Sernov 2011-01-01 Full Text Available This article describes the main principles of implementation of modern adaptive signal lighting equipment of vehicles, provides an analysis of optical systems are used, the necessity of the use of LEDs. We present the design of adaptive optical system, rear combination LED lamp of a vehicle with several levels of intensity, we discuss the algorithm of its work in different modes. 18. Optics and multilayer coatings for EUVL systems Energy Technology Data Exchange (ETDEWEB) Soufli, R; Bajt, S; Hudyma, R M; Taylor, J S 2008-03-21 EUV lithography (EUVL) employs illumination wavelengths around 13.5 nm, and in many aspects it is considered an extension of optical lithography, which is used for the high-volume manufacturing (HVM) of today's microprocessors. The EUV wavelength of illumination dictates the use of reflective optical elements (mirrors) as opposed to the refractive lenses used in conventional lithographic systems. Thus, EUVL tools are based on all-reflective concepts: they use multilayer (ML) coated optics for their illumination and projection systems, and they have a ML-coated reflective mask. 19. Can optical afterglows be used to discriminate between Type I and Type II GRBs? OpenAIRE Kann, D. A. 2008-01-01 The precise localization of short/hard (Type I) gamma-ray bursts (GRBs) in recent years has answered many questions but raised even more. I present some results of a systematic study of the optical afterglows of long/soft (Type II) and short/hard (Type I) GRBs, focusing on the optical luminosity as another puzzle piece in the classification of GRBs. 20. Synthesis, optical and thermal behaviour of palladium(II) complexes ... shows that all the ligands and Pd(II) complexes exhibit enantiotropic mesophases. The ligands with n-octyloxy and n-decyloxy flexible chains exhibit the nematic (N) and smectic A (SmA) phases whilst the Pd(II) complexes show exclusive SmA phase. The SmA phase observed in Pd(II) complexes can be supported by the ... 1. Fluid Mud in Energetic Systems: FLUMES II Science.gov (United States) 2014-09-30 FINAL REPORT Fluid Mud in Energetic Systems: FLUMES II Gail C. Kineke Dept of...the dynamics of fluid mud and its role in the transport and deposition of sediment in coastal environments. In particular, we sought greater...understanding of the processes that influence the formation and maintenance of fluid mud in energetic environments. OBJECTIVES The research is a process 2. NSLS-II RF Cryogenic System Energy Technology Data Exchange (ETDEWEB) Rose, J.; Dilgen, T.; Gash, B.; Gosman, J.; Mortazavi, P.; Papu, J.; Ravindranath, V.; Sikora, R.; Sitnikov, A.; Wilhelm, H.; Jia, Y.; Monroe, C. 2015-05-03 The National Synchrotron Light Source II is a 3 GeV X-ray user facility commissioned in 2014. A new helium refrigerator system has been installed and commissioned to support the superconducting RF cavities in the storage ring. Special care was taken to provide very stable helium and LN2 pressures and flow rates to minimize microphonics and thermal effects at the cavities. Details of the system design along with commissioning and early operations data will be presented. 3. Precision PEP-II optics measurement with an SVD-enhanced Least-Square fitting Energy Technology Data Exchange (ETDEWEB) Yan, Y 2004-09-15 A Singular value decomposition (SVD)-enhanced Least-square fitting technique is discussed. By automatic identifying, ordering, and selecting dominant SVD modes of the derivative matrix that responds to the variations of the variables, the converging process of the Least Square fitting is significantly enhanced. Thus the fitting speed can be fast enough for a fairly large system. This technique has been successfully applied to precision PEP-II optics measurement in which we determine all quadrupole strengths (both normal and skew components) and sextupole feed-downs as well as all BPM gains and BPM cross-plane couplings through Least-Square fitting of the phase advances and the Local Green's functions as well as the coupling ellipses among BPMs. The local Green's functions are specified by 4 local transfer matrix components R12, R34, R32, R14. These measurable quantities (the Green's functions, the phase advances and the coupling ellipse tilt angles and axis ratios) are obtained by analyzing turn-by-turn Beam Position Monitor (BPM) data with a high-resolution model-independent analysis (MIA). Once all of the quadrupoles and sextupole feed-downs are determined, we obtain a computer virtual accelerator which matches the real accelerator in linear optics. Thus, beta functions, linear coupling parameters, and interaction point (IP) optics characteristics can be measured and displayed. 4. Optical countermeasures against CLOS weapon systems NARCIS (Netherlands) Toet, A.; Benoist, K.W.; Lingen, J.N.J. van; Schleijpen, H.M.A. 2013-01-01 There are many weapon systems in which a human operator acquires a target, tracks it and designates it. Optical countermeasures against this type of systems deny the operator the possibility to fulfill this visual task. We describe the different effects that result from stimulation of the human 5. 77 FR 21586 - II-VI, Incorporated, Infrared Optics-Saxonburg Division, Saxonburg, PA; Notice of Affirmative... Science.gov (United States) 2012-04-10 ... Employment and Training Administration II-VI, Incorporated, Infrared Optics--Saxonburg Division, Saxonburg... former workers of II-VI, Incorporated, Infrared Optics--Saxonburg Division, Saxonburg, Pennsylvania... employment related to the production of infrared and CO 2 laser optics, and related materials. The initial... 6. 77 FR 36579 - II-VI, Inc., Infrared Optics-Saxonburg Division, Saxonburg, PA; Leased Workers From Adecco, Carol... Science.gov (United States) 2012-06-19 ... Employment and Training Administration II-VI, Inc., Infrared Optics-Saxonburg Division, Saxonburg, PA; Leased...., Infrared Optics-Saxonburg Division, Saxonburg, PA; Notice of Revised Determination on Reconsideration The... period. The determination was applicable to workers and former workers of II-VI, Inc., Infrared Optics... 7. Analysis of advanced optical glass and systems Science.gov (United States) Johnson, R. Barry; Feng, Chen 1991-01-01 Optical lens systems performance utilizing optical materials comprising reluctant glass forming compositions was studied. Such special glasses are being explored by NASA/Marshall Space Flight Center (MSFC) researchers utilizing techniques such as containerless processing in space on the MSFC Acoustic Levitation Furnace and on the High Temperature Acoustic Levitation Furnace in the conceptual design phase for the United States Microgravity Laboratory (USML) series of shuttle flights. The application of high refractive index and low dispersive power glasses in optical lens design was investigated. The potential benefits and the impacts to the optical lens design performance were evaluated. The results of the studies revealed that the use of these extraordinary glasses can result in significant optical performance improvements. Recommendations of proposed optical properties for potential new glasses were also made. Applications of these new glasses are discussed, including the impact of high refractive index and low dispersive power, improvements of the system performance by using glasses which are located outside of traditional glass map, and considerations in establishing glass properties beyond conventional glass map limits. 8. Fiber optical parametric amplifiers in optical communication systems Science.gov (United States) Marhic (†), Michel E; Andrekson, Peter A; Petropoulos, Periklis; Radic, Stojan; Peucheret, Christophe; Jazayerifar, Mahmoud 2015-01-01 The prospects for using fiber optical parametric amplifiers (OPAs) in optical communication systems are reviewed. Phase-insensitive amplifiers (PIAs) and phase-sensitive amplifiers (PSAs) are considered. Low-penalty amplification at/or near 1 Tb/s has been achieved, for both wavelength- and time-division multiplexed formats. High-quality mid-span spectral inversion has been demonstrated at 0.64 Tb/s, avoiding electronic dispersion compensation. All-optical amplitude regeneration of amplitude-modulated signals has been performed, while PSAs have been used to demonstrate phase regeneration of phase-modulated signals. A PSA with 1.1-dB noise figure has been demonstrated, and preliminary wavelength-division multiplexing experiments have been performed with PSAs. 512 Gb/s have been transmitted over 6,000 km by periodic phase conjugation. Simulations indicate that PIAs could reach data rate x reach products in excess of 14,000 Tb/s × km in realistic wavelength-division multiplexed long-haul networks. Technical challenges remaining to be addressed in order for fiber OPAs to become useful for long-haul communication networks are discussed. PMID:25866588 9. Optical Soliton Simulation in Optical Fibers by OptiSystem Science.gov (United States) Gaik Tay, Kim; Huong Kah Ching, Audrey; Loi, Wei Sen; Tiong Ong, Chee 2017-08-01 Fiber optic communication is often known to offer higher frequency transmission of signals with greater bit rate and larger data carrying capacity over a long distance with lower loss and interference as compared to copper wire electrical communication. However, several factors that would affect the performance of an optical fiber transmission are such as group velocity dispersion (GVD), fiber loss and also self-phase modulation (SPM). In this paper, the effects of GVD, SPM, optical soliton formation and fiber loss are simulated using OptiSystem 14. It is found that GVD broaden pulse in temporal domain without modifying its spectrum. Meanwhile, SPM creates chirp in spectrum with its temporal profile maintained. This work concluded that a balance between the GVD and SPM is essential to form solitonthat is able to travel for a long distance without being distorted. It is also found that the decrease in the amplitude of the soliton is dependent on the fiber loss and this decay in the signal increases with the propagation distance. 10. PEP-II RF feedback system simulation Energy Technology Data Exchange (ETDEWEB) Tighe, R. [Stanford Linear Accelerator Center, Menlo Park, CA (United States) 1996-08-01 A model containing the fundamental impedance of the PEP-II cavity along with the longitudinal beam dynamics and RF feedback system components is in use. It is prepared in a format allowing time-domain as well as frequency-domain analysis and full graphics capability. Matlab and Simulink are control system design and analysis programs (widely available) with many built-in tools. The model allows the use of compiled C-code modules for compute intensive portions. We desire to represent as nearly as possible the components of the feedback system including all delays, sample rates and applicable nonlinearities. (author) 11. Optical fiber telecommunications systems and networks CERN Document Server Kaminow, Ivan; Willner, Alan E 2013-01-01 Optical Fiber Telecommunications VI (A&B) is the sixth in a series that has chronicled the progress in the R&D of lightwave communications since the early 1970s. Written by active authorities from academia and industry, this edition brings a fresh look to many essential topics, including devices, subsystems, systems and networks. A central theme is the enabling of high-bandwidth communications in a cost-effective manner for the development of customer applications. These volumes are an ideal reference for R&D engineers and managers, optical systems implementers, university researchers and s 12. All-Optical Regeneration System for Optical Wavelength Division Multiplexed Communication Systems DEFF Research Database (Denmark) 2014-01-01 an input WDM data signal comprising multiple wavelength channels into an input OTDM data signal comprising multiple time multiplexed time channels. The system further comprises an all-optical regenerator unit being configured for regenerating the input OTDM data signal into an output OTDM data signal....... The system additionally comprises an OTDM-to-WDM converter for converting the output OTDM data signal to an output WDM data signal. An input of the all-optical regenerator unit is in optical communication with an output of the WDM-to-OTDM converter, and an output of the all-optical regenerator unit...... is in optical communication with an input of the OTDM-to-WDM converter. The invention further relates to a method for all-optical regeneration of WDM data signals.... 13. Spectroscopic, DNA binding ability, biological activity, DFT calculations and non linear optical properties (NLO) of novel Co(II), Cu(II), Zn(II), Cd(II) and Hg(II) complexes with ONS Schiff base Science.gov (United States) Abdel Aziz, Ayman A.; Elantabli, Fatma M.; Moustafa, H.; El-Medani, Samir M. 2017-08-01 The reaction of Co(II), Cu(II), Zn(II), Cd(II) and Hg(II) with the synthesized N-(2-hydroxy-1-naphthylidene)-2-aminothiophenol Schiff base ligand (H2L) at room temperature resulted in the formation of the five complexes; [Co(HL)2]H2O, 1; [M(HL)2] (M = Cu, Zn and Cd), (2-4) and [Hg(HL)Cl], 5. The ligand and its complexes were characterized based on elemental analyses, IR, 1H NMR, magnetic measurement, molar conductance, and thermal analysis. Coats and Redfern method was used to compute the kinetic and thermodynamic parameters. Antimicrobial activities of H2L and its complexes have been studied. The binding of Co(II), Cu(II) and Zn(II) complexes to calf thymus DNA (CT-DNA) has been investigated using UV-Vis and fluorescence absorption spectra. The results indicated that the ligand and its complexes may bind to DNA by intercalation modes, with a much higher binding affinity of the complexes than that of the ligand. The equilibrium geometries of the studied complexes are investigated theoretically at the B3LYP/LANL2DZ level of theory, and it was found that these geometries are non-linear. The calculated EHOMO and ELUMO energies of the studied complexes can be used to calculate the global properties. The calculated nonlinear optical parameters (NLO); first order hyperpolarizibility (β) of the studied complexes show promising optical properties. 14. Sandia Mark II X-Ray System Energy Technology Data Exchange (ETDEWEB) Morrison, L.W. 1979-11-01 The Sandia Mark II X-Ray System was designed and developed to provide an intense source of mononergetic, ultra-soft x rays with energies between 0.282 and 1.486 keV. The x-ray tube design is similar to one developed by B.L. Henke and incorporates modifications made by Tom Ellsberry. An operations manual section is incorporated to help the experimenter/operator. 15. ALFALFA DISCOVERY OF THE NEARBY GAS-RICH DWARF GALAXY LEO P. II. OPTICAL IMAGING OBSERVATIONS Energy Technology Data Exchange (ETDEWEB) Rhode, Katherine L.; Salzer, John J.; Haurberg, Nathalie C.; Van Sistine, Angela; Young, Michael D. [Department of Astronomy, Indiana University, 727 East Third Street, Bloomington, IN 47405 (United States); Haynes, Martha P.; Giovanelli, Riccardo; Adams, Elizabeth A. K. [Center for Radiophysics and Space Research, Space Sciences Building, Cornell University, Ithaca, NY 14853 (United States); Cannon, John M. [Department of Physics and Astronomy, Macalester College, Saint Paul, MN 55105 (United States); Skillman, Evan D.; McQuinn, Kristen B. W., E-mail: rhode@astro.indiana.edu, E-mail: slaz@astro.indiana.edu, E-mail: riccardo@astro.cornell.edu, E-mail: haynes@astro.cornell.edu, E-mail: betsey@astro.cornell.edu, E-mail: jcannon@macalester.edu, E-mail: skillman@astro.umn.edu, E-mail: kmcquinn@astro.umn.edu [Minnesota Institute for Astrophysics, University of Minnesota, Minneapolis, MN 55455 (United States) 2013-06-15 We present results from ground-based optical imaging of a low-mass dwarf galaxy discovered by the ALFALFA 21 cm H I survey. Broadband (BVR) data obtained with the WIYN 3.5 m telescope at Kitt Peak National Observatory (KPNO) are used to construct color-magnitude diagrams of the galaxy's stellar population down to V{sub o} {approx} 25. We also use narrowband H{alpha} imaging from the KPNO 2.1 m telescope to identify a H II region in the galaxy. We use these data to constrain the distance to the galaxy to be between 1.5 and 2.0 Mpc. This places Leo P within the Local Volume but beyond the Local Group. Its properties are extreme: it is the lowest-mass system known that contains significant amounts of gas and is currently forming stars. 16. Optical Blade Position Tracking System Test Energy Technology Data Exchange (ETDEWEB) Fingersh, L. J. 2006-01-01 The Optical Blade Position Tracking System Test measures the blade deflection along the span of the blade using simple off-the-shelf infrared security cameras along with blade-mounted retro-reflective tape and video image processing hardware and software to obtain these measurements. 17. Astrophysical determination of optical oscillator strengths for Ti II Energy Technology Data Exchange (ETDEWEB) Ryabchikova, T.A. [Institute for Astronomy, Moscow (Russian Federation); Hill, G.M.; Landstreet, J.D. [University of Western Ontario, London, ON (Canada). Dept. of Astronomy; Piskunov, N. [Institute for Astronomy, Moscow (Russian Federation)]|[Helsinki Univ. (Finland). Observatory and Astrophysics Lab.; Sigut, T.A.A. [University of Toronto (Canada). Dept. of Astronomy 1994-12-31 An internally consistent set of relative oscillator strengths for 109 spectral lines of Ti II in the range 3800-5500 A has been derived. These oscillator strengths have been obtained by combining available laboratory data and theoretical calculations with new astrophysical oscillator strengths determined from both published photographic equivalent width lists and spectral synthesis of a number of available Reticon spectra of several sharp-lined stars. The accuracy of our astrophysical oscillator strengths has been tested by determining astrophysical gf-values for a number of lines of Fe II by the same methods. The Ti II data have been normalized to the astrophysical oscillator strengths of Kostyk and Orlova, which also places them essentially on the scale of Danzmann and Kock, one of the most recent and extensive experimental data sets. It appears that the accuracy of the relative gf-values in our final mean list for Ti II is typically about 0.08 dex. The absolute normalization is believed to be secure to within about 0.15 dex. (Author). 18. Evolution Of Map Display Optical Systems Science.gov (United States) Boot, Alan 1983-06-01 It is now over 20 years since Ferranti plc introduced optically projected map displays into operational aircraft navigation systems. Then, as now, it was the function of the display to present an image of a topographical map to a pilot or navigator with his present position clearly identified. Then, as now, the map image was projected from a reduced image stored on colour micro film. Then, as now, the fundamental design problems are the same.In the exposed environment of an aircraft cockpit where brightness levels may vary from those associated with direct sunlight on the one hand, to starlight on the other, how does one design an optical system with sufficient luminance, contrast and resolution where in the daytime sunlight may fall on the display or in the pilot's eyes, and at night time the display luminance must not detract from the pilot's ability to pick up external clues? This paper traces the development of Ferranti plc optically projected map displays from the early V Bomber and the ill-fated TSR2 displays to the Harrier and Concorde displays. It then goes on to the development of combined map and electronic displays (COMED), showing how an earlier design, as fitted to Tornado, has been developed into the current COMED design which is fitted to the F-18 and Jaguar aircraft. In each of the above display systems particular features of optical design interest are identified and their impact on the design as a whole are discussed. The use of prisms both for optical rotation and translation, techniques for the maximisation of luminance, the problems associated with contrast enhancement, particularly with polarising filters in the presence of optically active materials, the use of aerial image combining systems and the impact of the pilot interface on the system parameter are all included.Perhaps the most interesting result in considering the evolution of map displays has not been so much the designer's solutions in overcoming the various design problems but 19. A Bloch equation approach to intensity dependent optical spectra of light harvesting complex II: excitation dependence of light harvesting complex II pump-probe spectra. Science.gov (United States) Richter, Marten; Renger, Thomas; Knorr, Andreas 2008-01-01 On the basis of the recent progress in the resolution of the structure of the antenna light harvesting complex II (LHC II) of the photosystem II, we propose a microscopically motivated theory to predict excitation intensity-dependent spectra. We show that optical Bloch equations provide the means to include all 2( N ) excited states of an oligomer complex of N coupled two-level systems and analyze the effects of Pauli Blocking and exciton-exciton annihilation on pump-probe spectra. We use LHC Bloch equations for 14 Coulomb coupled two-level systems, which describe the S (0) and S (1) level of every chlorophyll molecule. All parameter introduced into the Hamiltonian are based on microscopic structure and a quantum chemical model. The derived Bloch equations describe not only linear absorption but also the intensity dependence of optical spectra in a regime where the interplay of Pauli Blocking effects as well as exciton-exciton annihilation effects are important. As an example, pump-probe spectra are discussed. The observed saturation of the spectra for high intensities can be viewed as a relaxation channel blockade on short time scales due to Pauli blocking. The theoretical investigation is useful for the interpretation of the experimental data, if the experimental conditions exceed the low intensity pump limit and effects like strong Pauli Blocking and exciton-exciton annihilation need to be considered. These effects become important when multiple excitations are generated by the pump pulse in the complex. 20. Optical countermeasures against CLOS weapon systems Science.gov (United States) Toet, Alexander; Benoist, Koen W.; van Lingen, Joost N. J.; Schleijpen, H. Ric M. A. 2013-10-01 There are many weapon systems in which a human operator acquires a target, tracks it and designates it. Optical countermeasures against this type of systems deny the operator the possibility to fulfill this visual task. We describe the different effects that result from stimulation of the human visual system with high intensity (visible) light, and the associated potential operational impact. Of practical use are flash blindness, where an intense flash of light produces a temporary "blind-spot" in (part of) the visual field, flicker distraction, where strong intensity and/or color changes at a discomfortable frequency are produced, and disability glare where a source of light leads to contrast reduction. Hence there are three possibilities to disrupt the visual task of an operator with optical countermeasures such as flares or lasers or a combination of these; namely, by an intense flash of light, by an annoying light flicker or by a glare source. A variety of flares for this purpose is now available or under development: high intensity flash flares, continuous burning flares or strobe flares which have an oscillating intensity. The use of flare arrays seems particularly promising as an optical countermeasure. Lasers are particularly suited to interfere with human vision, because they can easily be varied in intensity, color and size, but they have to be directed at the (human) target, and issues like pointing and eye-safety have to be taken into account. Here we discuss the design issues and the operational impact of optical countermeasures against human operators. 1. Scanning Long-wave Optical Test System: a new ground optical surface slope test system Science.gov (United States) Su, Tianquan; Park, Won Hyun; Parks, Robert E.; Su, Peng; Burge, James H. 2011-09-01 The scanning long-wave optical test system (SLOTS) is under development at the University of Arizona to provide rapid and accurate measurements of aspherical optical surfaces during the grinding stage. It is based on the success of the software configurable optical test system (SCOTS) which uses visible light to measure surface slopes. Working at long wave infrared (LWIR, 7-14 μm), SLOTS measures ground optical surface slopes by viewing the specular reflection of a scanning hot wire. A thermal imaging camera collects data while motorized stages scan the wire through the field. Current experiments show that the system can achieve a high precision at micro-radian level with fairly low cost equipment. The measured surface map is comparable with interferometer for slow optics. This IR system could be applied early in the grinding stage of fabrication of large telescope mirrors to minimize the surface shape error imparted during processing. This advantage combined with the simplicity of the optical system (no null optics, no high power carbon dioxide laser) would improve the efficiency and shorten the processing time. 2. Development of a Novel Fiber Optic Sensor Combined with a Fluorescence Turn-on Probe for Cu (II Detection Directory of Open Access Journals (Sweden) Ma J. 2013-04-01 Full Text Available Existing staining-based methodology for the detection of metal ions is not well suited for real-time or in situ use. This is a significant problem, given that these ions can have a considerable impact on both human health and the environment. Thus, there is growing interest and need for simple, rapid and in-situ monitoring techniques for the purpose of detecting various target analytes (e.g. heavy metals, which is of a significant importance in many fields ranging from environmental monitoring to the study of intracellular processes. Among various sensors developed, optical fiber-optic sensors (FOS, based on fluorescence, are one class of sensors that address this goal [1]. Optical fibers are ideal for environmental sensing applications because of their ability to transmit optical signals to and from the sensing region without the use of free-space optics. In this work, we present, for the first time, a simple FOS incorporating novel fluorescence turn-on mechanism [2] that could detect Cu (II as low as 10−4 M. Traditionally, fluorescence quenching or “turn-off” was used to detect Cu (II [3]. In recent years, fluorescence “turn-on” emerges as a preferable tool. The developed fiber-optic sensor has two fiber leads and one probe head. One fiber lead includes 6 fibers for He-Ne laser excitation light delivery (e-fibers. Another fiber lead has one receiving fiber (r-fiber connected to an Ocean Optics QE65000 scientific grade spectrometer, which is interrogated by a computer via USB connection. The SpectroSuite software is used to observe and to record all spectra. The probe head combines all fibers together to form a coaxial structure with the r-fiber placed in the center. The key component in the proposed fluorescent sensing system is a probe prepared by binding a receptor containing a zwitterionic chromophore (M1, through noncovalent interactions, to the fluorescent polymer (P1 resulting in quenching its emission. The sensing mechanism 3. Interchip link system using an optical wiring method. Science.gov (United States) Cho, In-Kui; Ryu, Jin-Hwa; Jeong, Myung-Yung 2008-08-15 A chip-scale optical link system is presented with a transmitter/receiver and optical wire link. The interchip link system consists of a metal optical bench, a printed circuit board module, a driver/receiver integrated circuit, a vertical cavity surface-emitting laser/photodiode array, and an optical wire link composed of plastic optical fibers (POFs). We have developed a downsized POF and an optical wiring method that allows on-site installation with a simple annealing as optical wiring technologies for achieving high-density optical interchip interconnection within such devices. Successful data transfer measurements are presented. 4. 40 Gb/s optical transmission systems DEFF Research Database (Denmark) Buxens Azcoaga, Alvaro Juan; Nielsen, Steen Krogh 2003-01-01 This thesis investigates state of the art components and subsystems to be used in the next generation of high speed optical transmission systems at 40 Gb/s. The thesis will provide guidelines for the design and implementation of 40 Gb/s systems, investigating topics that could limit transmission......) experimental demonstrations are presented. The first one is a 100 GHz spaced 16 channel WDM system, using 40 Gb/s NRZ modulation over a 200 km link of standard-SMF. The second one is a 100 GHz spaced 32 channel 40 Gb/s WDM system, using CSRZ modulation over a 400 km link of standard-SMF and using Raman... 5. Cornea Optical Topographical Scan System (COTSS) Science.gov (United States) 1986-01-01 The Cornea Optical Topographical Scan System (COTSS) is an instrument designed for use by opthalmologist to aid in performing surgical procedures such as radial keratotomy and to provide quick accurate data to aid in prescribing contact lenses and eyeglasses. A breadboard of the system was built and demonstrated in June of 1984. Additional refinements to the breadboard are needed to meet systems requirements prior to proceeding with prototype development. The present status of the COTSS instrument is given and the areas in which system refinements are required, are defined. 6. Type-II Quantum Dot Nanowire Structures with Large Oscillator Strengths for Optical Quantum Gating Applications DEFF Research Database (Denmark) Taherkhani, Masoomeh; Gregersen, Niels; Willatzen, Morten 2017-01-01 The exciton oscillator strength (OS) in type-II quantum dot (QD) nanowires is calculated by using a fast and efficient method. We propose a new structure in Double-Well QD (DWQD) nanowire that considerably increases OS of type-II QDs which is a key parameter in optical quantum gating in the stimu...... in the stimulated Raman adiabatic passage (STIRAP) process [1] for implementing quantum gates.... 7. Deflection and translation of a ray traversing an optical system, nodal rays and the optical axis Directory of Open Access Journals (Sweden) W. F. Harris 2009-12-01 Full Text Available Expressions are derived for the deflection and transverse translation of a ray as it traverses an arbitrary optical system.  The system may be astigmatic and have relatively decentred refracting elements.  The expressions are in terms of the fundamental properties of the system.  Because nodal rays are rays that undergo no deflection the results lead immediately to expressions that define nodal rays through the system.  An optical axis of an optical system is a nodal ray that is a straight line through the system.  This allows one to write an expression that determines the optical axis of an arbitrary optical system. 8. Boosted X waves in nonlinear optical systems. Science.gov (United States) Arévalo, Edward 2010-01-15 X waves are spatiotemporal optical waves with intriguing superluminal and subluminal characteristics. Here we theoretically show that for a given initial carrier frequency of the system localized waves with genuine superluminal or subluminal group velocity can emerge from initial X waves in nonlinear optical systems with normal group velocity dispersion. Moreover, we show that this temporal behavior depends on the wave detuning from the carrier frequency of the system and not on the particular X-wave biconical form. A spatial counterpart of this behavior is also found when initial X waves are boosted in the plane transverse to the direction of propagation, so a fully spatiotemporal motion of localized waves can be observed. 9. Ultra Small Integrated Optical Fiber Sensing System Directory of Open Access Journals (Sweden) Peter Van Daele 2012-09-01 Full Text Available This paper introduces a revolutionary way to interrogate optical fiber sensors based on fiber Bragg gratings (FBGs and to integrate the necessary driving optoelectronic components with the sensor elements. Low-cost optoelectronic chips are used to interrogate the optical fibers, creating a portable dynamic sensing system as an alternative for the traditionally bulky and expensive fiber sensor interrogation units. The possibility to embed these laser and detector chips is demonstrated resulting in an ultra thin flexible optoelectronic package of only 40 μm, provided with an integrated planar fiber pigtail. The result is a fully embedded flexible sensing system with a thickness of only 1 mm, based on a single Vertical-Cavity Surface-Emitting Laser (VCSEL, fiber sensor and photodetector chip. Temperature, strain and electrodynamic shaking tests have been performed on our system, not limited to static read-out measurements but dynamically reconstructing full spectral information datasets. 10. Fiber laser coupled optical spark delivery system Science.gov (United States) Yalin, Azer [Fort Collins, CO; Willson, Bryan [Fort Collins, CO; Defoort, Morgan [Fort Collins, CO; Joshi, Sachin [Fort Collins, CO; Reynolds, Adam [Fort Collins, CO 2008-03-04 A spark delivery system for generating a spark using a laser beam is provided, and includes a laser light source and a laser delivery assembly. The laser delivery assembly includes a hollow fiber and a launch assembly comprising launch focusing optics to input the laser beam in the hollow fiber. The laser delivery assembly further includes exit focusing optics that demagnify an exit beam of laser light from the hollow fiber, thereby increasing the intensity of the laser beam and creating a spark. Other embodiments use a fiber laser to generate a spark. Embodiments of the present invention may be used to create a spark in an engine. Yet other embodiments include collecting light from the spark or a flame resulting from the spark and conveying the light for diagnostics. Methods of using the spark delivery systems and diagnostic systems are provided. 11. Fiber coupled optical spark delivery system Science.gov (United States) Yalin, Azer; Willson, Bryan; Defoort, Morgan 2008-08-12 A spark delivery system for generating a spark using a laser beam is provided, the spark delivery system including a laser light source and a laser delivery assembly. The laser delivery assembly includes a hollow fiber and a launch assembly comprising launch focusing optics to input the laser beam in the hollow fiber. In addition, the laser delivery assembly includes exit focusing optics that demagnify an exit beam of laser light from the hollow fiber, thereby increasing the intensity of the laser beam and creating a spark. In accordance with embodiments of the present invention, the assembly may be used to create a spark in a combustion engine. In accordance with other embodiments of the present invention, a method of using the spark delivery system is provided. In addition, a method of choosing an appropriate fiber for creating a spark using a laser beam is also presented. 12. Multifunctional optical system-on-a-chip for heterogeneous fiber optic sensor networks Science.gov (United States) Yu, Miao; Pang, Cheng; Gupta, Ashwani 2015-08-01 In this article, we review our recent progress on the development of a multifunctional optical system-on-a-chip platform, which can be used for achieving heterogeneous wireless fiber optical sensor networks. A multifunctional optical sensor platform based on the micro-electromechanical systems (MEMS) technology is developed. The key component of the multifunctional optical sensor platform is a MEMS based tunable Fabry-Pérot (FP) filter, which can be used as a phase modulator or a wavelength tuning device in a multifunctional optical sensing system. Mechanics model of the FP filter and optics model of the multifunctional optical sensing system are developed to facilitate the design of the filter. The MEMS FP filter is implemented in a multifunctional optical sensing system including both Fabry-Perot interferometer based sensors and Fiber Bragg grating sensors. The experimental results indicate that this large dynamic range tunable filter can enable high performance heterogeneous optical sensing for many applications. 13. Evaluation of Laser Stabilization and Imaging Systems for LCLS-II - Final Paper Energy Technology Data Exchange (ETDEWEB) Barry, Matthew [Auburn Univ., AL (United States) 2015-08-20 By combining the top performing commercial laser beam stabilization system with the most ideal optical imaging configuration, the beamline for the Linear Accelerator Coherent Light Source II (LCLS-II) will deliver the highest quality and most stable beam to the cathode. To determine the optimal combination, LCLS-II beamline conditions were replicated and the systems tested with a He-Ne laser. The Guidestar-II and MRC active laser beam stabilization systems were evaluated for their ideal positioning and stability. Both a two and four lens optical imaging configuration was then evaluated for beam imaging quality, magnification properties, and natural stability. In their best performances when tested over fifteen hours, Guidestar-II kept the beam stable over approximately 70-110um while the MRC system kept it stable over approximately 90-100um. During short periods of time, Guidestar-II kept the beam stable between 10-20um, but was more susceptible to drift over time, while the MRC system maintained the beam between 30-50um with less overall drift. The best optical imaging configuration proved to be a four lens system that images to the iris located in the cathode room and from there, imaged to the cathode. The magnification from the iris to the cathode was 2:1, within an acceptable tolerance to the expected 2.1:1 magnification. The two lens configuration was slightly more stable in small periods of time (less than 10 minutes) without the assistance of a stability system, approximately 55um compared to approximately 70um, but the four lens configurations beam image had a significantly flatter intensity distribution compared to the two lens configuration which had a Gaussian distribution. A final test still needs to be run with both stability systems running at the same time through the four lens system. With this data, the optimal laser beam stabilization system can be determined for the beamline of LCLS-II. 14. Zoom optical system using tunable polymer lens Science.gov (United States) Liang, Dan; Wang, Xuan Yin 2016-07-01 This paper demonstrated a zoom optical system with variable magnification based on the tunable polymer lens. The designed system mainly consists of two polymer lenses, voice coil motors, a doublet lens and CMOS chip. The zoom magnification can be adjusted by altering the focal length of the two elastic polymer lenses synergistically through controlling the output displacement of the voice coil motor. A static doublet lens in combination with the polymer lenses stabilize the image plane at the CMOS chip. The optical structure of the zoom system is presented, as well as a detailed description including the lens materials and fabrication process. Images with each zoom magnification are captured, and the Spot diagram and MTF are simulated using Zemax software. A change in magnification from 0.13×to 8.44×is demonstrated within the tiny 0.4 mm variation of the displacement load, and produce a 16.1×full range of magnification experimentally. Simulation analyses show that all the radii of the spot diagram under different magnifications are less than 11.3 um, and the modulation transfer function reaches 107 line pairs per mm. The designed optical system shows the potential for developing stable, integrated, and low-cost zoom systems with large magnification range. 15. Detection of herbicide subclasses by an optical multibiosensor based on an array of photosystem II mutants. Science.gov (United States) Giardi, Maria Teresa; Guzzella, Licia; Euzet, Pierre; Rouillon, Regis; Esposito, Dania 2005-07-15 Massive use of herbicides in agriculture over the last few decades has become a serious environmental problem. The residual concentration of these compounds frequently exceeds the maximum admissible concentration in drinking water for human consumption and is a real environmental risk for the aquatic ecosystem. Herbicides inhibiting photosynthesis via targeting photosystem II function still represent the basic means of weed control. A multibiosensor was constructed for detecting herbicides using as biosensing elements photosynthetic preparations coupled to an optical fluorescence transduction system (Giardi et al. EU patent EP1134585, 01830148.1-2204); this paper is about its application in the detection of herbicide subclasses in river water. Photosynthetic material was immobilized on a silicio septum inside a series of flow cells, close to diodes so as to activate photosystem II (PSII) fluorescence. The principle of the detection was based on the factthat herbicides selectively modify PSII fluorescence activity. The multibiosensor has the original feature of being able to distinguish the subclasses of the photosynthetic herbicides by using specific immobilized biomediators isolated from mutated organisms. This setup resulted in a reusable, portable multibiosensor for the detection of herbicide subclasses with a half-life of 54 h for spinach thylakoids and limit of detection of 3 x 10(-9) M for herbicides present in river water. 16. SAFARI optical system architecture and design concept Science.gov (United States) Pastor, Carmen; Jellema, Willem; Zuluaga-Ramírez, Pablo; Arrazola, David; Fernández-Rodriguez, M.; Belenguer, Tomás.; González Fernández, Luis M.; Audley, Michael D.; Evers, Jaap; Eggens, Martin; Torres Redondo, Josefina; Najarro, Francisco; Roelfsema, Peter 2016-07-01 SpicA FAR infrared Instrument, SAFARI, is one of the instruments planned for the SPICA mission. The SPICA mission is the next great leap forward in space-based far-infrared astronomy and will study the evolution of galaxies, stars and planetary systems. SPICA will utilize a deeply cooled 2.5m-class telescope, provided by European industry, to realize zodiacal background limited performance, and high spatial resolution. The instrument SAFARI is a cryogenic grating-based point source spectrometer working in the wavelength domain 34 to 230 μm, providing spectral resolving power from 300 to at least 2000. The instrument shall provide low and high resolution spectroscopy in four spectral bands. Low Resolution mode is the native instrument mode, while the high Resolution mode is achieved by means of a Martin-Pupplet interferometer. The optical system is all-reflective and consists of three main modules; an input optics module, followed by the Band and Mode Distributing Optics and the grating Modules. The instrument utilizes Nyquist sampled filled linear arrays of very sensitive TES detectors. The work presented in this paper describes the optical design architecture and design concept compatible with the current instrument performance and volume design drivers. 17. The experimental optical burst switching system Science.gov (United States) Li, Xinwan; Chen, Jian-Ping; Wu, Guiling; Wang, Hui; Lu, Jialin; Ye, Ailun 2005-02-01 The first optical burst switching (OBS) system has been demonstrated in China, which includes three edge routers and one core-node. A kind of fast wavelength selective optical switching was used in the system. The core OBS node consists of a kind of wavelength selective optical switch we developed. It consists of two SOA switches and one wavelength selective thin film filter with centre wavelength at one wavelength. There are one input optical fiber and two output fibers, each fiber carries two wavelengths. The Dell PE2650 servers act as the edge OBS routers. The wavelength of each data channel is located in C-band and the bit rate is at 1.25Gbps. The control channel uses bit rate of 100Mbps at wavelength of 1310 nm. A novel effective scheme for Just-In-Time (JIT) protocol was proposed and implemented. OBS services, such as Video on Demand (VOD) and file transfer protocol (FTP), have been demonstrated. Assembling and scheduling methods that are capable to guarantee the QoS (quality of service) of the transported service are studied. 18. Optical Design Using an Expert System Energy Technology Data Exchange (ETDEWEB) Lerner, S A 2003-08-01 We present, as a different perspective on optimization, an expert system for optimization of optical systems that can be used in conjunction with damped least squared methods to find minima for specific design forms. Expert system optimization differs from global optimization in that it preserves the basic structure of the optical system and limits its search for a minima to a relatively small portion of the design space. In general, the high density of local minima obscures the general trend of the merit function in the region of interest for systems with a large number of variables and constraints. Surprisingly, there may be a potential decrease of an order a magnitude in the merit function for a region of solution space. While global optimization is well-suited to identifying design forms of interest, expert system optimization can be used for in-depth optimization of such forms. An expert system based upon such techniques was used to obtain the winning entry for the 2002 IODC lens design problem. The expert system used is discussed along with other design examples. 19. Turbidity Measurement Using An Optical Tomography System Directory of Open Access Journals (Sweden) Sallehuddin Bin Ibrahim 2013-10-01 Full Text Available Turbidity is used to describe water quality and it can be caused by the presence of suspended particles and organic matter such as algae, clay and silt. The measurement of turbidity level of water is vital to domestic water supplies since it is related to public health and water treatment process. This paper presents an investigation on an optical tomography system to estimate the turbidity level in a sample of water. The optical sensors consist of infrared light-emitting diodes (LED as transmitters and photodiodes as the receivers where the projections of the sensors are designed in fan beam mode. The system was tested using a vertical flow pipe. The Independent Component Analysis (ICA method was used to display the concentration profile. Results obtained proved that the technique can provide the concentration profile representing the turbidity level of water. 20. Distributed fiber optic moisture intrusion sensing system Science.gov (United States) Weiss, Jonathan D. 2003-06-24 Method and system for monitoring and identifying moisture intrusion in soil such as is contained in landfills housing radioactive and/or hazardous waste. The invention utilizes the principle that moist or wet soil has a higher thermal conductance than dry soil. The invention employs optical time delay reflectometry in connection with a distributed temperature sensing system together with heating means in order to identify discrete areas within a volume of soil wherein temperature is lower. According to the invention an optical element and, optionally, a heating element may be included in a cable or other similar structure and arranged in a serpentine fashion within a volume of soil to achieve efficient temperature detection across a large area or three dimensional volume of soil. Remediation, moisture countermeasures, or other responsive action may then be coordinated based on the assumption that cooler regions within a soil volume may signal moisture intrusion where those regions are located. 1. Beam Position Monitor System for PEP II Energy Technology Data Exchange (ETDEWEB) Smith, Stephen R.; Aiello, G.Roberto; Hendrickson, Linda J.; Johnson, Ronald G.; Mills, Mark R.; Olsen, Jeff J.; /SLAC 2011-09-12 We describe the beam position monitor system built for PEP-II, the B-factory at SLAC. The system reports beam position for bunches of between 5 x 10{sup 8} and 8 x 10{sup 10} electron charges, either singly or as continuous streams of bunches every 4.2 ns. Resolution at full charge is to be better than 10 microns in a single turn. Higher resolution is available via on-board multi-turn averaging. The position signal is processed in a 20 MHz bandwidth around 952 MHz. This bandwidth, rather broader than that typical of RF position monitors, allows good resolution for low charge single bunches. Additional novel features include stringent control of return losses in order to minimize cross-talk between nearby bunches which may contain very different charges. The digitizing electronics is multiplexed between the two PEP-II storage rings. Design, construction, and installation experience, as well as first results with beam are presented. 2. Fundamentals of physics II electromagnetism, optics, and quantum mechanics CERN Document Server Shankar, R 2016-01-01 R. Shankar, a well-known physicist and contagiously enthusiastic educator, was among the first to offer a course through the innovative Open Yale Course program. His popular online video lectures on introductory physics have been viewed over a million times. In this second book based on his online Yale course, Shankar explains essential concepts, including electromagnetism, optics, and quantum mechanics. The book begins at the simplest level, develops the basics, and reinforces fundamentals, ensuring a solid foundation in the principles and methods of physics. It provides an ideal introduction for college-level students of physics, chemistry, and engineering; for motivated AP Physics students; and for general readers interested in advances in the sciences. 3. Application of optical transmission in control power system Science.gov (United States) 2011-06-01 The advantage of fiber optic data transmission systems is their insensitivity to electromagnetic radiation, in which the optical fiber. Some exemplary applications of fiber optic transmission in power electronics and electric drives which due to their specific job bring significant electromagnetic interference. Electrical signal wires are in an area of very strong electromagnetic field so that the transmission distortion may occur. The paper presents existing and possible applications of optical links using optical fibers, in the control of power systems. 4. Fiber Optic Temperature Sensors for Thermal Protection Systems Project Data.gov (United States) National Aeronautics and Space Administration — Intelligent Fiber Optic Systems Corporation (IFOS) proposes an innovative fiber optic-based, multiplexable, highly ruggedized, integrated sensor system for real-time... 5. Test runs of a Belle II PXD prototype readout system Energy Technology Data Exchange (ETDEWEB) Getzkow, Dennis; Kuehn, Wolfgang; Lange, Soeren; Lautenbach, Klemens [Justus-Liebig-Universitaet Giessen, II. Physikalisches Institut (Germany); Gessler, Thomas [KEK, Tsukuba (Japan); Collaboration: Belle II-Collaboration 2016-07-01 The Belle II PXD readout system (called ONSEN for Online Selection Nodes) uses ATCA (Advanced Telecommunications Architecture) boards with Xilinx Virtex-5 FX70T FPGAs and high speed optical links (6.5 Gbit/s each). The full system consists of 9 carrier boards and 33 daughter cards. The ONSEN system has several interfaces: (a) it receives PXD data from the DHH (Data Handling Hybrid) system, (b) it receives ROI (Regions-of-Interest) data for online data reduction from the HLT (High Level Trigger) system by GbE, and (c) it features data ports to two event builders: EVB1 combines data from all detectors except PXD (in order to generate the ROIs) and EVB2 combines the reduced PXD data with all other data. One of the critical issues is the matching of trigger numbers in the data (received by DHH from the timing distribution system) and trigger numbers in the ROIs (received by the HLT). In order to test the interfaces, in particular for a high HLT rate up to 30 kHz, a prototype system with 3 daughter cards was installed at KEK and tested with DHH, HLT and EVB2. Test results are presented. 6. Digital optical recorder-reproducer system Science.gov (United States) Reddersen, Brad R. (Inventor); Zech, Richard G. (Inventor); Roberts, Howard N. (Inventor) 1980-01-01 A mass archival optical recording and reproduction system includes a recording light source such as a laser beam focussed and directed upon an acousto-optic linear modulator array (or page composer) that receives parallel blocks of data converted from a serial stream of digital data to be stored. The page composer imparts to the laser beam modulation representative of a plurality of parallel channels of data and through focussing optics downstream of the page composer parallel arrays of optical spots are recorded upon a suitable recording medium such as a photographic film floppy disc. The recording medium may be substantially frictionlessly and stably positioned for recording at a record/read station by an air-bearing platen arrangement which is preferably thermodynamically non-throttling so that the recording film may be positioned in the path of the information-carrying light beam in a static or dynamic mode. During readout, the page composer is bypassed and a readout light beam is focussed directly upon the recording medium containing an array of previously recorded digital spots, a sync bit, data positioning bits, and a tracking band. The readout beam which has been directed through the recording medium is then imaged upon a photodetector array, the output of which may be coupled to suitable electronic processing circuitry, such as a digital multiplexer, whereby the parallel spot array is converted back into the original serial data stream. 7. LSST active optics system software architecture Science.gov (United States) Thomas, Sandrine J.; Chandrasekharan, Srinivasan; Lotz, Paul; Xin, Bo; Claver, Charles; Angeli, George; Sebag, Jacques; Dubois-Felsmann, Gregory P. 2016-08-01 The Large Synoptic Survey Telescope (LSST) is an 8-meter class wide-field telescope now under construction on Cerro Pachon, near La Serena, Chile. This ground-based telescope is designed to conduct a decade-long time domain survey of the optical sky. In order to achieve the LSST scientific goals, the telescope requires delivering seeing limited image quality over the 3.5 degree field-of-view. Like many telescopes, LSST will use an Active Optics System (AOS) to correct in near real-time the system aberrations primarily introduced by gravity and temperature gradients. The LSST AOS uses a combination of 4 curvature wavefront sensors (CWS) located on the outside of the LSST field-of-view. The information coming from the 4 CWS is combined to calculate the appropriate corrections to be sent to the 3 different mirrors composing LSST. The AOS software incorporates a wavefront sensor estimation pipeline (WEP) and an active optics control system (AOCS). The WEP estimates the wavefront residual error from the CWS images. The AOCS determines the correction to be sent to the different degrees of freedom every 30 seconds. In this paper, we describe the design and implementation of the AOS. More particularly, we will focus on the software architecture as well as the AOS interactions with the various subsystems within LSST. 8. An exactly solvable system from quantum optics Energy Technology Data Exchange (ETDEWEB) Maciejewski, Andrzej J., E-mail: maciejka@astro.ia.uz.zgora.pl [J. Kepler Institute of Astronomy, University of Zielona Góra, Licealna 9, PL-65-417 Zielona Góra (Poland); Przybylska, Maria, E-mail: M.Przybylska@if.uz.zgora.pl [Institute of Physics, University of Zielona Góra, Licealna 9, 65-417 Zielona Góra (Poland); Stachowiak, Tomasz, E-mail: stachowiak@cft.edu.pl [Center for Theoretical Physics PAS, Al. Lotników 32/46, 02-668 Warsaw (Poland) 2015-07-31 We investigate a generalisation of the Rabi system in the Bargmann–Fock representation. In this representation the eigenproblem of the considered quantum model is described by a system of two linear differential equations with one independent variable. The system has only one irregular singular point at infinity. We show how the quantisation of the model is related to asymptotic behaviour of solutions in a vicinity of this point. The explicit formulae for the spectrum and eigenfunctions of the model follow from an analysis of the Stokes phenomenon. An interpretation of the obtained results in terms of differential Galois group of the system is also given. - Highlights: • New exactly solvable system from quantum optics is found. • Normalisation condition for system in Bargmann representation is used. • Formulae for spectrum and eigenfunctions from analysis of Stokes phenomenon are given. 9. Optical multi-species gas monitoring sensor and system Science.gov (United States) Polzin, Kurt A. (Inventor); Korman, Valentin (Inventor) 2012-01-01 The system includes at least one light source generating light energy having a corresponding wavelength. The system's sensor is based on an optical interferometer that receives light energy from each light source. The interferometer includes a free-space optical path disposed in an environment of interest. The system's sensor includes an optical device disposed in the optical path that causes light energy of a first selected wavelength to continue traversing the optical path whereas light energy of at least one second selected wavelength is directed away from the optical path. The interferometer generates an interference between the light energy of the first selected wavelength so-traversing the optical path with the light energy at the corresponding wavelength incident on the optical interferometer. A first optical detector detects the interference. At least one second detector detects the light energy at the at least one second selected wavelength directed away from the optical path. 10. Optical Systems for Ultra-High-Speed TDM Networking DEFF Research Database (Denmark) Galili, Michael; Hu, Hao; Mulvad, Hans Christian Hansen 2014-01-01 This paper discusses key results in the field of high speed optical networking with particular focus on packet-based systems. Schemes for optical packet labeling, packet switching and packet synchronization will be discussed, along with schemes for optical clock recovery, channel identification...... and detection of ultra-high-speed optical signals.... 11. Optical Systems for Ultra-High-Speed TDM Networking Directory of Open Access Journals (Sweden) Michael Galili 2014-04-01 Full Text Available This paper discusses key results in the field of high speed optical networking with particular focus on packet-based systems. Schemes for optical packet labeling, packet switching and packet synchronization will be discussed, along with schemes for optical clock recovery, channel identification and detection of ultra-high-speed optical signals. 12. Constructions of Optical Queues With a Limited Number of Recirculations--Part II: Optimal Constructions CERN Document Server Huang, Xuan-Chao 2010-01-01 One of the main problems in all-optical packet-switched networks is the lack of optical buffers, and one feasible technology for the constructions of optical buffers is to use optical crossbar Switches and fiber Delay Lines (SDL). In this two-part paper, we consider SDL constructions of optical queues with a limited number of recirculations through the optical switches and the fiber delay lines. Such a problem arises from practical feasibility considerations. In Part I, we have proposed a class of greedy constructions for certain types of optical queues, including linear compressors, linear decompressors, and 2-to-1 FIFO multiplexers, and have shown that every optimal construction among our previous constructions of these types of optical queues under the constraint of a limited number of recirculations must be a greedy construction. In Part II, the present paper, we further show that there are at most two optimal constructions and give a simple algorithm to obtain the optimal construction(s). The main idea i... 13. Optics CERN Document Server Fincham, W H A 2013-01-01 Optics: Ninth Edition Optics: Ninth Edition covers the work necessary for the specialization in such subjects as ophthalmic optics, optical instruments and lens design. The text includes topics such as the propagation and behavior of light; reflection and refraction - their laws and how different media affect them; lenses - thick and thin, cylindrical and subcylindrical; photometry; dispersion and color; interference; and polarization. Also included are topics such as diffraction and holography; the limitation of beams in optical systems and its effects; and lens systems. The book is recommen 14. Triangular platinum(II) metallacycles: syntheses, photophysics, and nonlinear optics. Science.gov (United States) Fan, Yuanpeng; Zhao, Dahui 2015-03-25 Three triangular platinum(II) diimine metallacycles incorporating large cyclic oligo(phenylene-ethynylene) (OPE) bisacetylide ligands are synthesized, and their photophysical properties are studied. Two types of triplet excited states with ligand/metal-to-ligand charge-transfer and acetylide-ligand-centered characteristics respectively, are exhibited by these complexes depending on the size (conjugation length) and electronic features of the cyclic OPE ligands. When the energy levels of the two excited states are close to each other, the lowest triplet state is found to switch between the two in varied solvents, resulting from their relative energy inversion induced by solvent polarity change. Density functional theory and time-dependent density functional theory calculations provide corroborative evidence for such experimental conclusions. More importantly, the designed metallacycles show impressive two-photon absorption (2PA) and two-photon excitation phosphorescing abilities, and the 2PA cross section reaches 1020 GM at 680 nm and 670 GM at 1040 nm by two different metallacycles. Additionally, pronounced reverse saturable absorptions are observed with these metallacycles by virtue of their strong transient triplet-state absorptions. 15. The LCLS-II LLRF System Energy Technology Data Exchange (ETDEWEB) DooLittle, Lawrence [Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States); Huang, G. [Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States); Ratti, A. [Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States); Serrano, C. [Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States); Bachimanchi, Ramakrishna [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Hovater, J. Curt [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Babel, S. [SLAC National Accelerator Lab., Menlo Park, CA (United States); Hong, B. [SLAC National Accelerator Lab., Menlo Park, CA (United States); Van Winkle, D. [SLAC National Accelerator Lab., Menlo Park, CA (United States); Chase, B. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Cullerton, E. [FNAL, Batavia, IL; Varghese, P. [FNAL, Batavia, IL 2015-09-01 The SLAC National Accelerator Laboratory is planning an upgrade (LCLS-II) to the Linear Coherent Light Source with a 4 GeV CW superconducting (SCRF) linac. The SCRF linac consists of 35 ILC style cryomodules (eight cavities each) for a total of 280 cavities. Expected cavity gradients are 16 MV/m with a loaded QL of ~ 4x107. The RF system will have 3.8 kW solid state amplifiers driving single cavities. To ensure optimum field stability a single-source single-cavity control system has been chosen. It consists of a precision four-channel cavity receiver and RF stations (Forward, Reflected and Drive signals). In order to regulate the resonant frequency variations of the cavities due to He pressure, the tuning of each cavity is controlled by a Piezo actuator and a slow stepper motor. In addition the system (LLRF-amplifier-cavity) is being modeled and cavity microphonic testing has started. This paper describes the LLRF system under consideration, including recent modeling and cavity tests. 16. Fiber-optic subscriber system based on passive optical network architecture Energy Technology Data Exchange (ETDEWEB) Kitazawa, M.; Yamazaki, M.; Himi, S. (Hitachi, Ltd., Tokyo (Japan)) 1994-04-01 The subscriber line terminal and optical network unit were developed for the fiber-optic subscriber system based on a passive double star (PDS) optical network architecture which allows narrowband ISDN (integrated services digital network) services and video image transmission. Various LSIs were developed to implement the major system functions such as fast bit-synchronization, time division multiple access control, subscriber multiplexing/demultiplexing and 16-channel multiprocessing. For the optical interface module, the wavelength division multiplexing chip was hermetically sealed together with laser diode and photodiode chips by adopting high-silica based optical-waveguide technology to reduce a system dimension. The PDS fiber-optic subscriber system is useful for construction of economic optical subscriber networks because it allows multiple subscribers to share a single optical transmission line and a central equipment unit through a star coupler. 4 refs., 10 figs., 1 tab. 17. Reproducible Design for the Optical Screening and Sensing of Hg(II Ions Directory of Open Access Journals (Sweden) 2014-10-01 Full Text Available We fabricated silica nanotubes with hexagonally ordered mesopores (6 nm inside a membrane disc with a uniform channel neck size of 200 nm and a longitudinal thickness of 60 μm to design an optical sensor membrane (OSM for the screening and sensing of extremely toxic Hg(II ions. The optical detection and quantitative recognition of Hg(II ions in water were conducted even at trace concentrations without the need for sophisticated instruments. The OSM design was based on the physical interaction of a responsive organic probe with silica pore surfaces followed by strong and selective binding Hg(II–probe interactions under specific sensing conditions, particularly at pH 5. Ultra-trace concentrations of Hg(II ions were easily detected with the naked eye using the OSM. The remarkable ion spectral response of Hg(II ion–OSM ensured the excellent quantification of the OSM for Hg(II ion sensing over a wide range of concentrations with a detection limit of 1.75 × 10−9 M. This result indicated that low concentrations of Hg(II ions can be detected with a high sensitivity. One of the key issues of OSM is the Hg(II ion-selective workability even in the presence of high doses of competitive matrices and species. The OSM design showed significant Hg(II ion-sensing capability despite the number of reuse/recycles using simple decomplexation. Given its high selectivity, fast response, and sensitivity, the OSM could be developed into a specific Hg(II ion-sensing kit in aqueous solutions. 18. The CDF-II silicon tracking system Energy Technology Data Exchange (ETDEWEB) F. Palmonari et al. 2002-01-18 The CDFII silicon tracking system, SVX, for Run II of the Fermilab Tevatron has up to 8 cylindrical layers with average radii spanning from {approx} (1.5 to 28.7) cm, and lengths ranging from {approx} (90 to 200) cm for a total active-area of {approx} 6 m{sup 2} and {approx} 7.2 x 10{sup 5} readout channels. SVX will improve the CDFII acceptance and efficiency for both B and high-Pt physics dependent upon b-tagging. Along with the description of the SVX we report some alignment survey data from the SVX assembly phase and the actual status of the alignment as it results from the offline data analysis. The problems encountered are also reviewed. 19. ADVANTAGES OF DIFFRACTIVE OPTICAL ELEMENTS APPLICATION IN SIMPLE OPTICAL IMAGING SYSTEMS Directory of Open Access Journals (Sweden) N. D. Zoric 2015-01-01 Full Text Available The paper deals with the influence of diffractive optical elements on the optical aberrations. The correction of optical aberrations was investigated in the simple optical systems with one and two lenses (singlet and doublet. The advantages of diffractive optical elements are their ability to generate arbitrary complex wave fronts from a piece of optical material that is essentially flat. The optical systems consisting of the standard surfaces were designed and optimized by using the same starting points. Further, the diffractive and aspheric surfaces were introduced into the developed systems. The resulting hybrid systems were optimized. To compare the complicity of the development of narrow field systems and wide field optical systems, the optimization has been done separately for these two types of the instruments. The optical systems were designed by using special Optical Design Software. Тhe characteristics of designed diffractive surfaces were controlled in Software DIFSYS 2.30. Due to the application of diffractive optical elements the longitudinal chromatic aberration was 5 times reduced for the narrow field systems. The absolute value of Seidel coefficient related to the spherical aberration was reduced in the range of 0.03. Considering that diffractive optical elements have the known disadvantages, like possible parasitic diffraction orders and probable decrease of the transmission, we also developed and analyzed the optical systems with combined aspheric and diffractive surfaces. A combination of the aspheric and diffractive surfaces in the optical disk system of the disk reading lens, gave cutting down of the longitudinal color aberrations almost 15 times on-axis, comparing to the lens consisting of the aspherical and standard surfaces. All of the designed diffractive optical elements possess the parameters within the fabrication limits. 20. Laser action due to Ar II and Ne II ions pumped by an optical breakdown produced by a CO sub 2 laser Energy Technology Data Exchange (ETDEWEB) Apollonov, V.V.; Derzhavin, S.I.; Noraev, D.A.; Sirotkin, A.A. (Institut Obshchei Fiziki, Moscow (USSR)) 1990-09-01 The paper reports laser action in Ar II at wavelengths of 454.5 and 427.8 nm and in Ne II at 332.4 and 334.5 nm under pumping by X-rays emitted from an optical-breakdown plasma produced by a CO2 laser. Alternative mechanisms of active-medium formation under optical breakdown are discussed. 7 refs. 1. Optical monitoring system for a turbine engine Energy Technology Data Exchange (ETDEWEB) Lemieux, Dennis H; Smed, Jan P; Williams, James P; Jonnalagadda, Vinay 2013-05-14 The monitoring system for a gas turbine engine including a viewing tube assembly having an inner end and an outer end. The inner end is located adjacent to a hot gas flow path within the gas turbine engine and the outer end is located adjacent to an outer casing of the gas turbine engine. An aperture wall is located at the inner end of the viewing tube assembly and an optical element is located within the viewing tube assembly adjacent to the inner end and is spaced from the aperture wall to define a cooling and purge chamber therebetween. An aperture is defined in the aperture wall for passage of light from the hot gas flow path to the optical element. Swirl passages are defined in the viewing tube assembly between the aperture wall and the optical element for passage of cooling air from a location outside the viewing tube assembly into the chamber, wherein swirl passages effect a swirling movement of air in a circumferential direction within the chamber. Science.gov (United States) Harper, Warren W [Benton City, WA; Aker, Pamela M [Richland, WA; Pratt, Richard M [Richland, WA 2011-05-10 Free space optical communication methods and systems, according to various aspects are described. The methods and systems are characterized by transmission of data through free space with a digitized optical signal acquired using wavelength modulation, and by discrimination between bit states in the digitized optical signal using a spectroscopic absorption feature of a chemical substance. 3. Dispersion characteristics and compensation in the POLMUX coherent optical communication system Science.gov (United States) Zhu, Guijun; Ruan, Xiukai; Zhou, Zhili 2017-07-01 The trend of high-speed optical communication is to realize large capacity, fast speed and long haul transmission. However, the dispersion in an optical fiber link can cause signals distortion or even inter symbol interference (ISI). With increasing signal rate and transmission distance in the optical communication system, the development of high-speed optical fiber communication is restricted seriously. This paper is focusing on the dispersion characteristics and compensation of the polarization multiplexing (POLMUX) coherent optical communication system: (i) to study on dispersion theoretically, e.g., chromatic dispersion (CD), polarization mode dispersion (PMD) and dispersion tolerance; (ii) to analyze and compare the performance of dispersion with various 16-ary modulation formats; (iii) to compensate dispersion by the classical dispersion compensation methods. 4. An optical imaging system for window chambers in MRI system Science.gov (United States) Lin, Yuxiang; Salek, Mir F. S.; Jennings, Nicki; Gmitro, Arthur F. 2008-02-01 Window chambers are support structures implanted in the dorsal skin fold of a rodent model. Optical imaging of window chambers has been used in many basic cancer and vascular biology studies. We have recently shown that this technique can be extended to MRI by using plastic rather than metal window chambers. Here we describe a system for simultaneous optical and MR imaging of the window chambers. It provides many possibilities for independent cross validation of the measurements of one modality from the other. In the system, a GRIN lens images the tissue to the distal end of a coherent imaging fiber bundle, which relays this image to a camera system located outside the magnet room. Both trans- and epi-illumination are provided to this system. Light sources are located outside the magnet room and the light is delivered through fiber optics. A group of fibers are used to deliver white light from under the window chamber for standard transmission imaging, while another single fiber delivers the laser light from the top to induce fluorescence. An appropriate bandpass emission filter is inserted between the lenses at the camera end for fluorescence imaging. Results of simultaneously optical and MR imaging of tumor and vessel are presented. 5. Theoretical optical spectroscopy of complex systems Energy Technology Data Exchange (ETDEWEB) Conte, A. Mosca, E-mail: adriano.mosca.conte@roma2.infn.it [MIFP, NAST, ETSF,CNR INFM-SMC, Universitá di Roma Tor Vergata, Via della Ricerca Scientifica 1, Roma (Italy); Violante, C., E-mail: claudia.violante@roma2.infn.it [MIFP, NAST, ETSF,CNR INFM-SMC, Universitá di Roma Tor Vergata, Via della Ricerca Scientifica 1, Roma (Italy); Missori, M., E-mail: mauro.missori@isc.cnr.it [Istituto dei Sistemi Complessi, Consiglio Nazionale delle Ricerche, Via Salaria Km 29.300, 00016 Monterotondo Scalo (Rome) (Italy); Bechstedt, F., E-mail: bech@ifto.physik.uni-jena.de [Institut fur Festkorpertheorie und -optik, Friedrich-Schiller-Universitat, Max-Wien-Platz 1, 07743 Jena (Germany); Teodonio, L. [MIFP, NAST, ETSF,CNR INFM-SMC, Universitá di Roma Tor Vergata, Via della Ricerca Scientifica 1, Roma (Italy); Istituto centrale per il restauro e la conservazione del patrimonio archivistico e librario (IC-RCPAL), Italian Minister for Cultural Heritage, Via Milano 76, 00184 Rome (Italy); Ippoliti, E.; Carloni, P. [German Research School for Simulation Sciences, Julich (Germany); Guidoni, L., E-mail: leonardo.guidoni@univaq.it [Università degli Studi di L’Aquila, Dipartimento di Chimica e Materiali, Via Campo di Pile, 67100 L’Aquila (Italy); Pulci, O., E-mail: olivia.pulci@roma2.infn.it [MIFP, NAST, ETSF,CNR INFM-SMC, Universitá di Roma Tor Vergata, Via della Ricerca Scientifica 1, Roma (Italy) 2013-08-15 Highlights: ► We review some theoretical condensed matter ab initio spectroscopic computational techniques. ► We show several applications ranging from 0 to 3 dimensional systems. ► For each system studied, we show which kind of information it is possible to obtain by performing these calculations. -- Abstract: We review here some of the most reliable and efficient computational theoretical ab initio techniques for the prediction of optical and electronic spectroscopic properties and show some important applications to molecules, surfaces, and solids. We investigate the role of the solvent in the optical absorption spectrum of indole molecule. We study the excited-state properties of a photo-active minimal model molecule for the retinal of rhodopsin, responsible for vision mechanism in animals. We then show a study about spectroscopic properties of Si(1 1 1) surface. Finally we simulate a bulk system: paper, that is mainly made of cellulose, a pseudo-crystalline material representing 40% of annual biomass production in the Earth. 6. Type-II quantum-dot-in-nanowire structures with large oscillator strength for optical quantum gate applications DEFF Research Database (Denmark) Taherkhani, Masoomeh; Willatzen, Morten; Mørk, Jesper 2017-01-01 -I systems. To increase the oscillator strength, we propose a double quantum dot structure featuring a strongly localized exciton wave function and a corresponding fourfold relative enhancement of the oscillator strength, paving the way towards efficient optically controlled quantum gate applications......We present a numerical investigation of the exciton energy and oscillator strength in type-II nanowire quantum dots. For a single quantum dot, the poor overlap of the electron part and the weakly confined hole part of the excitonic wave function leads to a low oscillator strength compared to type... 7. Multilayer optical disc system using homodyne detection Science.gov (United States) Kurokawa, Takahiro; Ide, Tatsuro; Tanaka, Yukinobu; Watanabe, Koichi 2014-09-01 A write/read system using high-productivity multilayer optical discs was developed. The recording medium used in the system consists of planar recording layers and a separated guide layer, and is fabricated by web coating and lamination process. The recording layers in the medium are made of one-photon-absorption material, on which data can be recorded with a normal laser diode. The developed system is capable of focusing and tracking on the medium and amplifying readout signals by using phase-diversity homodyne detection. A highly layer-selective focusing method using homodyne detection was also proposed. This method obtains stable focus-error signals with clearly separated S-shaped curves even when layer spacing is quite narrow, causing large interlayer crosstalk. Writing on the medium and reading with the signal amplification effect of homodyne detection was demonstrated. In addition, the effectiveness of the method was experimentally evaluated. 8. The Belle II SVD data readout system Energy Technology Data Exchange (ETDEWEB) Thalmeier, R., E-mail: Richard.Thalmeier@oeaw.ac.at [Institute of High Energy Physics, Austrian Academy of Sciences, 1050 Vienna (Austria); Adamczyk, K. [H. Niewodniczanski Institute of Nuclear Physics, Krakow 31-342 (Poland); Aihara, H. [Department of Physics, University of Tokyo, Tokyo 113-0033 (Japan); Angelini, C. [Dipartimento di Fisica, Universita’ di Pisa, I-56127 Pisa (Italy); INFN Sezione di Pisa, I-56127 Pisa (Italy); Aziz, T.; Babu, V. [Tata Institute of Fundamental Research, Mumbai 400005 (India); Bacher, S. [H. Niewodniczanski Institute of Nuclear Physics, Krakow 31-342 (Poland); Bahinipati, S. [Indian Institute of Technology Bhubaneswar, Satya Nagar (India); Barberio, E.; Baroncelli, Ti.; Baroncelli, To. [School of Physics, University of Melbourne, Melbourne, Victoria 3010 (Australia); Basith, A.K. [Indian Institute of Technology Madras, Chennai 600036 (India); Batignani, G. [Dipartimento di Fisica, Universita’ di Pisa, I-56127 Pisa (Italy); INFN Sezione di Pisa, I-56127 Pisa (Italy); Bauer, A. [Institute of High Energy Physics, Austrian Academy of Sciences, 1050 Vienna (Austria); Behera, P.K. [Indian Institute of Technology Madras, Chennai 600036 (India); Bergauer, T. [Institute of High Energy Physics, Austrian Academy of Sciences, 1050 Vienna (Austria); Bettarini, S. [Dipartimento di Fisica, Universita’ di Pisa, I-56127 Pisa (Italy); INFN Sezione di Pisa, I-56127 Pisa (Italy); Bhuyan, B. [Indian Institute of Technolog y Guwahati, Assam 781039 (India); Bilka, T. [Faculty of Mathematics and Physics, Charles University, 12116 Prague (Czech Republic); Bosi, F. [INFN Sezione di Pisa, I-56127 Pisa (Italy); and others 2017-02-11 The Belle II Experiment at the High Energy Accelerator Research Organization (KEK) in Tsukuba, Japan, will explore the asymmetry between matter and antimatter and search for new physics beyond the standard model. 172 double-sided silicon strip detectors are arranged cylindrically in four layers around the collision point to be part of a system which measures the tracks of the collision products of electrons and positrons. A total of 1748 radiation-hard APV25 chips read out 128 silicon strips each and send the analog signals by time-division multiplexing out of the radiation zone to 48 Flash Analog Digital Converter Modules (FADC). Each of them applies processing to the data; for example, it uses a digital finite impulse response filter to compensate line signal distortions, and it extracts the peak timing and amplitude from a set of several data points for each hit, using a neural network. We present an overview of the SVD data readout system, along with front-end electronics, cabling, power supplies and data processing. 9. Nano-optical conveyor belt, part II: Demonstration of handoff between near-field optical traps. Science.gov (United States) Zheng, Yuxin; Ryan, Jason; Hansen, Paul; Cheng, Yao-Te; Lu, Tsung-Ju; Hesselink, Lambertus 2014-06-11 Optical tweezers have been widely used to manipulate biological and colloidal material, but the diffraction limit of far-field optics makes focused beams unsuitable for manipulating nanoscale objects with dimensions much smaller than the wavelength of light. While plasmonic structures have recently been successful in trapping nanoscale objects with high positioning accuracy, using such structures for manipulation over longer range has remained a significant challenge. In this work, we introduce a conveyor belt design based on a novel plasmonic structure, the resonant C-shaped engraving (CSE). We show how long-range manipulation is made possible by means of handoff between neighboring CSEs, and we present a simple technique for controlling handoff by rotating the polarization of laser illumination. We experimentally demonstrate handoff between a pair of CSEs for polystyrene spheres 200, 390, and 500 nm in diameter. We then extend this technique and demonstrate controlled particle transport down a 4.5 μm long "nano-optical conveyor belt." 10. Systems engineering and analysis of electro-optical and infrared systems CERN Document Server Arrasmith, William Wolfgang 2015-01-01 Introduction to Electro-optic and Infrared (EO/IR) Systems Engineering?Radiation in the Visible and Infrared Parts of the Electromagnetic SpectrumRadiation SourcesThe Effect of the Atmosphere on Optical PropagationBasic OpticsOptical ModulationThe Detection of Optical RadiationNoise in the Optical Detection ProcessTechnical Performance Measures and Metrics of Optical DetectorsModern Detectors and their Measures of PerformanceThe Effects of Cooling on Optical Detector NoiseSignal and Image ProcessingElectro-Optic and Infrared Systems AnalysisLaser Imaging Systems?Spectral Imaging?LIDAR and LADA 11. Bridge SHM system based on fiber optical sensing technology Science.gov (United States) Li, Sheng; Fan, Dian; Fu, Jiang-hua; Huang, Xing; Jiang, De-sheng 2015-09-01 The latest progress of our lab in recent 10 years on the area of bridge structural health monitoring (SHM) based on optical fiber sensing technology is introduced. Firstly, in the part of sensing technology, optical fiber force test-ring, optical fiber vibration sensor, optical fiber smart cable, optical fiber prestressing loss monitoring method and optical fiber continuous curve mode inspection system are developed, which not only rich the sensor types, but also provides new monitoring means that are needed for the bridge health monitoring system. Secondly, in the optical fiber sensing network and computer system platform, the monitoring system architecture model is designed to effectively meet the integration scale and effect requirement of engineering application, especially the bridge expert system proposed integration of sensing information and informatization manual inspection to realize the mode of multi index intelligence and practical monitoring, diagnosis and evaluation. Finally, the Jingyue bridge monitoring system as the representative, the research on the technology of engineering applications are given. 12. Electro-optic intensity chaotic system with an extra optical feedback Science.gov (United States) Hu, Hanping; Shi, Saiying; Xie, Feilong 2017-11-01 Electro-optic (EO) delay oscillators have been widely investigated for secure optical communication. To improve complexity and security, an EO intensity chaotic system with an extra optical feedback is proposed. Simulation results show that the proposed system can improve complexity. Moreover, it can effectively suppress the time delay analysis, and the suppression can be enhanced by increasing the optical feedback strength of our system. In addition, synchronization of the communication system in a chaotic modulation encryption scheme based on the proposed system is discussed. 13. Robust optical oxygen sensors based on polymer-bound NIR-emitting platinum(II)-benzoporphyrins DEFF Research Database (Denmark) Hutter, L.H.; Müller, B.J.; Koren, Klaus 2014-01-01 Several advanced optical oxygen sensor materials are presented. They are based on bright NIR-emitting platinum(II)-benzoporphyrins covalently incorporated into a variety of polymeric matrices. The dye-polymer conjugates are prepared either via Suzuki coupling of the brominated porphyrins to the s......Several advanced optical oxygen sensor materials are presented. They are based on bright NIR-emitting platinum(II)-benzoporphyrins covalently incorporated into a variety of polymeric matrices. The dye-polymer conjugates are prepared either via Suzuki coupling of the brominated porphyrins...... to the styrene backbone or via co-polymerisation of the monomers with monostyryl porphyrin derivative. Importantly, in both strategies a highly stable C-C bond is obtained. The resulted materials benefit from excellent photophysical properties of the benzoporphyrin dyes (high brightness, emission in the NIR part... 14. HabEx space telescope optical system Science.gov (United States) Martin, Stefan; Rud, Mayer; Scowen, Paul; Stern, Daniel; Nissen, Joel; Krist, John 2017-09-01 The HabEx study is defining a concept for a new space telescope with the primary mission of detecting and characterizing planetary systems around nearby stars. The telescope is designed specifically to operate with both a high contrast coronagraph and a starshade, enabling the direct optical detection of exoplanets as close as 70 mas from their star. The telescope will be equipped with cameras for exoplanetary system imaging and with spectrometers capable of characterizing exoplanet atmospheres. Gases such as oxygen, carbon dioxide, water vapor and methane have spectral lines in the visible and near infrared part of the spectrum and may indicate biological activity. In addition to the study of exoplanets, HabEx enables general astrophysics with two dedicated instruments. One instrument is a camera enabling imaging on a 3 arc minute field of view in two bands stretching from the UV to the near infrared. The same instrument can also be operated as a multi-object spectrograph, with resolution of 2000. A second instrument will be a high resolution UV spectrograph operating from 120 nm with up to 60,0000 resolution. We discuss the preliminary designs of the telescope and the optical instruments for the observatory. 15. Optics CERN Document Server Fincham, W H A 2013-01-01 Optics: Eighth Edition covers the work necessary for the specialization in such subjects as ophthalmic optics, optical instruments and lens design. The text includes topics such as the propagation and behavior of light; reflection and refraction - their laws and how different media affect them; lenses - thick and thin, cylindrical and subcylindrical; photometry; dispersion and color; interference; and polarization. Also included are topics such as diffraction and holography; the limitation of beams in optical systems and its effects; and lens systems. The book is recommended for engineering st 16. Indexing system for optical beam steering Science.gov (United States) Sullivan, Mark T.; Cannon, David M.; Debra, Daniel B.; Young, Jeffrey A.; Mansfield, Joseph A.; Carmichael, Roger E.; Lissol, Peter S.; Pryor, G. M.; Miklosy, Les G.; Lee, Jeffrey H. 1990-01-01 This paper describes the design and testing of an indexing system for optical-beam steering. The cryogenic beam-steering mechanism is a 360-degree rotation device capable of discrete, high-precision alignment positions. It uses low-precision components for its rough alignment and kinematic design to meet its stringent repeatability and stability requirements (of about 5 arcsec). The principal advantages of this design include a decoupling of the low-precision, large angular motion from the high-precision alignment, and a power-off alignment position that potentially extends the life or hold time of cryogenic systems. An alternate design, which takes advantage of these attributes while reducing overall motion, is also presented. Preliminary test results show the kinematic mount capable of sub-arc second repeatability. 17. Observing high-frequency optical turbulence properties by the usage of fiber optical turbulence sensing system Science.gov (United States) Huang, Qi-kai; Mei, Hai-ping; Xiao, Shu-mei; Rao, Rui-zhong 2012-10-01 Effects of light propagation in random atmospheric optical turbulence are critical problems for ground based high resolution optical imaging. To get further knowledge of turbulence intensity or structure properties, the concept of fiber optic sensing system is proposed and realized. Different to fine-wire platinum resistance thermometer or laser scintillometer, the system has the ability to make non-contact measurement of optical turbulence up to the frequency of 500Hz during the air gap of 100mm, and has the dynamic range of 10-18 ~ 10-12 . The optic fiber's merit of corrosion resistance is sufficiently demonstrated by one month field test on the seacoast. Some properties of high frequency turbulence power spectrum that have never been observed before in the range of 10cm are reported. In the end, prospects of optical turbulence research by the usage of fiber optic turbulence sensing system are also discussed. 18. Optical studies of charged excitons in II-VI semiconductor quantum wells CERN Document Server Kossacki, P 2003-01-01 A brief review is given of optical studies of doped II-VI quantum wells. The properties of exciton states, neutral as well as positively and negatively charged, are discussed. A wide range of optical measurements is presented: CW as well as picosecond and femtosecond time-resolved absorption, photoluminescence (PL) and PL excitation. The experiments were performed at various carrier concentrations (> 10 sup 1 sup 1 cm sup - sup 2) and temperatures (up to a few tens of kelvins). This review is limited to zero or low magnetic fields, used only to polarize spins of carriers. We discuss the obtained values of various fundamental parameters of the excitonic states: energies, optical transition probabilities and characteristic times of their formation, thermalization and decay. (topical review) 19. The magneto optical system at Risø DEFF Research Database (Denmark) Larsen, Britt Hvolbæk 2001-01-01 The new magneto optical system at Risø is presented in details. Results measured on the system is shown. Image correction and magnetic field to current density conversion is described.......The new magneto optical system at Risø is presented in details. Results measured on the system is shown. Image correction and magnetic field to current density conversion is described.... 20. Optically driven oscillations of ellipsoidal particles. Part II: ray-optics calculations. Science.gov (United States) Loudet, J-C; Mihiretie, B M; Pouligny, B 2014-12-01 We report numerical calculations on the mechanical effects of light on micrometer-sized dielectric ellipsoids immersed in water. We used a simple two-dimensional ray-optics model to compute the radiation pressure forces and torques exerted on the object as a function of position and orientation within the laser beam. Integration of the equations of motion, written in the Stokes limit, yields the particle dynamics that we investigated for different aspect ratios k. Whether the beam is collimated or focused, the results show that above a critical aspect ratio k(C), the ellipsoids cannot be stably trapped on the beam axis; the particle never comes to rest and rather oscillates permanently in a back-and-forth motion involving both translation and rotation in the vicinity of the beam. Such oscillations are a direct evidence of the non-conservative character of optical forces. Conversely, stable trapping can be achieved for k non-linear coupling of the forces and torques and the torque amplitude was identified as the bifurcation control parameter. Interestingly, simulations predict that sustained oscillations can be suppressed through the use of two coaxial counterpropagating beams, which may be of interest whenever a static equilibrium is required as in basic force and torque measurements or technological applications. 1. Pseudo-type-II tuning behavior and mode identification in whispering gallery optical parametric oscillators. Science.gov (United States) Meisenheimer, Sarah-Katharina; Fürst, Josef Urban; Schiller, Annelie; Holderied, Florian; Buse, Karsten; Breunig, Ingo 2016-06-27 Wavelength tuning of conventional mirror-based optical parametric oscillators (OPOs) exhibits parabolically-shaped tuning curves (type-0 and type-I phase matching) or tuning branches that cross each other with a finite slope (type-II phase matching). We predict and experimentally prove that whispering gallery OPOs based on type-0 phase matching show both tuning behaviors, depending on whether the mode numbers of the generated waves coincide or differ. We investigate the wavelength tuning of optical parametric oscillation in a millimeter-sized radially-poled lithium niobate disk pumped at 1 μm wavelength generating signal and idler waves between 1.7 and 2.6 μm wavelength. Our experimental findings excellently coincide with the theoretical predictions. The investigated whispering gallery optical parametric oscillator combines the employment of the highest nonlinear-optical coefficient of the material with a controlled type-II-like wavelength tuning and with the possibility of self-phase locking. 2. Research of STBC in free space optical communications system Science.gov (United States) Deng, Tianping; Peng, Kai; Lu, Yimin 2009-12-01 Free Space Optical (FSO) communication is a communication pattern in which laser is used as the information carrier in atmosphere environment. FSO system has been widely applied in the fields of satellite communication and ground sight distance communication, with a promising application prospect and tremendous market potential. And space-time processing (using multiple antennas) is now recognized as a key to achieving reliable high data rate wireless communications and is being incorporated into the physical layer of many wireless standards. In this letter, error rate performance for space-time block coding (STBC) in FSO communication systems with direct detection operating over strong atmospheric turbulence channels is analyzed. The purpose of the paper is: (i) to determine the channel capacity of FSO system in the presence of atmospheric turbulence, (ii) to see how much this limit can be approached using the best known STBC techniques. In this paper, we use mathematics tool to establish the channel model of FSO system, and then apply the STBC to the FSO system, research the decoding method and simulate the performance using MATLAB tools. The mathematics analysis and simulation results show the great application expectation of STBC in the OOK channel of the FSO system. 3. Optical Imaging and Microscopy Techniques and Advanced Systems CERN Document Server Török, Peter 2007-01-01 This text on contemporary optical systems is intended for optical researchers and engineers, graduate students and optical microscopists in the biological and biomedical sciences. This second edition contains two completely new chapters. In addition most of the chapters from the first edition have been revised and updated. The book consists of three parts: The first discusses high-aperture optical systems, which form the backbone of optical microscopes. An example is a chapter new in the second edition on the emerging field of high numerical aperture diffractive lenses which seems to have particular promise in improving the correction of lenses. In this part particular attention is paid to optical data storage. The second part is on the use of non-linear optical techniques, including nonlinear optical excitation (total internal reflection fluorescence, second and third harmonic generation and two photon microscopy) and non-linear spectroscopy (CARS). The final part of the book presents miscellaneous technique... 4. Optical design of the adaptive optics laser guide star system Energy Technology Data Exchange (ETDEWEB) Bissinger, H. [Lawrence Livermore National Lab., CA (United States) 1994-11-15 The design of an adaptive optics package for the 3 meter Lick telescope is presented. This instrument package includes a 69 actuator deformable mirror and a Hartmann type wavefront sensor operating in the visible wavelength; a quadrant detector for the tip-tile sensor and a tip-tilt mirror to stabilize atmospheric first order tip-tile errors. A high speed computer drives the deformable mirror to achieve near diffraction limited imagery. The different optical components and their individual design constraints are described. motorized stages and diagnostics tools are used to operate and maintain alignment throughout observation time from a remote control room. The expected performance are summarized and actual results of astronomical sources are presented. 5. Multiparameter fiber optic sensing system for monitoring enhanced geothermal systems Energy Technology Data Exchange (ETDEWEB) Challener, William A 2014-12-04 The goal of this project was to design, fabricate and test an optical fiber cable which supports multiple sensing modalities for measurements in the harsh environment of enhanced geothermal systems. To accomplish this task, optical fiber was tested at both high temperatures and strains for mechanical integrity, and in the presence of hydrogen for resistance to darkening. Both single mode (SM) and multimode (MM) commercially available optical fiber were identified and selected for the cable based on the results of these tests. The cable was designed and fabricated using a tube-within-tube construction containing two MM fibers and one SM fiber, and without supporting gel that is not suitable for high temperature environments. Commercial fiber optic sensing instruments using Raman DTS (distributed temperature sensing), Brillouin DTSS (distributed temperature and strain sensing), and Raleigh COTDR (coherent optical time domain reflectometry) were selected for field testing. A microelectromechanical systems (MEMS) pressure sensor was designed, fabricated, packaged, and calibrated for high pressure measurements at high temperatures and spliced to the cable. A fiber Bragg grating (FBG) temperature sensor was also spliced to the cable. A geothermal well was selected and its temperature and pressure were logged. The cable was then deployed in the well in two separate field tests and measurements were made on these different sensing modalities. Raman DTS measurements were found to be accurate to ±5°C, even with some residual hydrogen darkening. Brillouin DTSS measurements were in good agreement with the Raman results. The Rayleigh COTDR instrument was able to detect some acoustic signatures, but was generally disappointing. The FBG sensor was used to determine the effects of hydrogen darkening, but drift over time made it unreliable as a temperature or pressure sensor. The MEMS sensor was found to be highly stable and accurate to better than its 0.1% calibration. 6. Simulation of DKIST solar adaptive optics system Science.gov (United States) Marino, Jose; Carlisle, Elizabeth; Schmidt, Dirk 2016-07-01 Solar adaptive optics (AO) simulations are a valuable tool to guide the design and optimization process of current and future solar AO and multi-conjugate AO (MCAO) systems. Solar AO and MCAO systems rely on extended object cross-correlating Shack-Hartmann wavefront sensors to measure the wavefront. Accurate solar AO simulations require computationally intensive operations, which have until recently presented a prohibitive computational cost. We present an update on the status of a solar AO and MCAO simulation tool being developed at the National Solar Observatory. The simulation tool is a multi-threaded application written in the C++ language that takes advantage of current large multi-core CPU computer systems and fast ethernet connections to provide accurate full simulation of solar AO and MCAO systems. It interfaces with KAOS, a state of the art solar AO control software developed by the Kiepenheuer-Institut fuer Sonnenphysik, that provides reliable AO control. We report on the latest results produced by the solar AO simulation tool. 7. Optical Correlator in Industrial Image Pattern Recognition System Directory of Open Access Journals (Sweden) Dávid Solus 2016-12-01 Full Text Available The aim of this paper is design a system for recognizing industrial image pattern using optical correlator. The proposed system is used to obtain information about the accuracy of the model and the industrial form of images, in this case – pavements. Cambridge optical correlator is used in designed system as comparator. Several experiments have been done by software called “Fourier Optics Experimenter”. Results and conclusion are discussed. 8. 78 FR 71785 - Passenger Train Emergency Systems II Science.gov (United States) 2013-11-29 ... published a final rule on Passenger Train Emergency Systems (PTES) on February 1, 2008. See 73 FR 6370. The... Working Group meetings, FRA initiated the Passenger Train Emergency Systems II (PTES II) rulemaking. In... regulations. For example, the 2008 PTES final rule established requirements that improve passenger emergency... 9. Smart Optical Material Characterization System and Method Science.gov (United States) Choi, Sang Hyouk (Inventor); Park, Yeonjoon (Inventor) 2015-01-01 Disclosed is a system and method for characterizing optical materials, using steps and equipment for generating a coherent laser light, filtering the light to remove high order spatial components, collecting the filtered light and forming a parallel light beam, splitting the parallel beam into a first direction and a second direction wherein the parallel beam travelling in the second direction travels toward the material sample so that the parallel beam passes through the sample, applying various physical quantities to the sample, reflecting the beam travelling in the first direction to produce a first reflected beam, reflecting the beam that passes through the sample to produce a second reflected beam that travels back through the sample, combining the second reflected beam after it travels back though the sample with the first reflected beam, sensing the light beam produced by combining the first and second reflected beams, and processing the sensed beam to determine sample characteristics and properties. 10. Self-amplified optical pattern recognition system Science.gov (United States) Liu, Hua-Kuang (Inventor) 1994-01-01 A self amplifying optical pattern recognizer includes a geometric system configuration similar to that of a Vander Lugt holographic matched filter configuration with a photorefractive crystal specifically oriented with respect to the input beams. An extraordinarily polarized, spherically converging object image beam is formed by laser illumination of an input object image and applied through a photorefractive crystal, such as a barium titanite (BaTiO.sub.3) crystal. A volume or thin-film dif ORIGIN OF THE INVENTION The invention described herein was made in the performance of work under a NASA contract, and is subject to the provisions of Public Law 96-517 (35 USC 202) in which the Contractor has elected to retain title. 11. Geophysical Fluid Flow Cell (GFFC) Optical System Science.gov (United States) 1995-01-01 This drawing shows a cross-section view of the optical system for the Geophysical Fluid Flow Cell (GFFC) experiment that flew on two Spacelab missions. Silicone oil served as the atmosphere around a rotating metal sphere and an electrostatic field pulled the oil inward to mimic gravity's effects during the experiments. The GFFC thus produced flow patterns that simulated conditions inside the atmospheres of Jupiter and the Sun and other stars. The principal investigator was John Hart of the University of Colorado at Boulder. It was managed by NASA's Marshall Space Flight Center. An Acrobat PDF copy of this drawing is available at http://microgravity.nasa.gov/gallery. (Credit: NASA/Marshall Space Flight Center) 12. Advanced modelling of optical coherence tomography systems DEFF Research Database (Denmark) Andersen, Peter E.; Thrane, L.; Yura, H.T. 2004-01-01 Analytical and numerical models for describing and understanding the light propagation in samples imaged by optical coherence tomography (OCT) systems are presented. An analytical model for calculating the OCT signal based on the extended Huygens–Fresnel principle valid both for the single...... and multiple scattering regimes is reviewed. An advanced Monte Carlo model for calculating the OCT signal is also reviewed, and the validity of this model is shown through a mathematical proof based on the extended Huygens–Fresnel principle. Moreover, for the first time the model is verified experimentally....... From the analytical model, an algorithm for enhancing OCT images is developed; the so-called true-reflection algorithm in which the OCT signal may be corrected for the attenuation caused by scattering. For the first time, the algorithm is demonstrated by using the Monte Carlo model as a numerical... 13. Brayton isotope power system. Volume II. System evaluation attributes Energy Technology Data Exchange (ETDEWEB) 1978-03-15 This volume of the Brayton Isotope Power System, Phase II Plan, contains the self-evaluation by AiResearch, GE, and TECO, addressing Section 3 of The Dynamic Systems Evaluation Criteria and Procedures established by the Department of Energy. These evaluation criteria addresses: Component Feasibility; Flight System Design Performance; GDS Test Results; Reliability and Practicality; Safety; Spacecraft Integration; and Cost and Risk. Included in each of these general categories are several attributes, each of which addresses a separate component, feature, or area of interest related to the power system, its development status, degree of preparedness for proceeding into a flight program, and/or the contractors' performance during Phase I. The key elements which indicate the readiness of a radioisotope power system to progress into a flight qualification program are: an advanced state of development of the power conversion system; demonstrated or exhibited potential for space systems standards of reliability; evident capability of meeting system safety requirements; favorable cost/benefit tradeoff considering projected missions and technology advancement potential; and proven feasibility of fabricating and qualifying a flight system and integrating it with a candidate spacecraft and launch vehicle. As a result of considerable government investment in Brayton system component development, the MHW isotope heat source and the BIPS Phase I Ground Demonstration System, the BIPS is a more advanced state of development than any previous radioisotope power system technology. Evidence of this is presented along with a complete review of the attributes, the contractor recommended ratings, and the rationale for the self-evaluation. 14. Optoelectronic devices for optical memory systems Science.gov (United States) Wang, Xu This thesis describes several optoelectronic devices developed for the optical memory system. It consists of three parts, which are liquid crystal beam steering devices, modeling and measurement of optical diffraction from subwavelength structure at focused spot, and CMOS detector design. In liquid crystal beam steering devices, two types of liquid crystal beam deflectors have been successfully demonstrated. The essential working idea of those two deflectors is the same, which is utilizing a blazed grating (either permanently fabricated or virtually built up) to deflect the incident beam and the liquid crystal is used as an electrically controllable birefringent medium. The reflective type deflector exploits the liquid crystal on silicon (LCOS) technology to build up a virtual blazed grating (dynamic grating) and vary this virtual grating's period to achieve multiple steering angles. Seven addressable angles with as high as 93% steering efficiency have been demonstrated. A numerical modeling tool is also developed to analyze liquid crystal molecule director's distribution under nonuniform electrical field (fringe-field-effect). The transmission type deflector is based on a custom-fabricated PMMA blazed grating (fixed grating) and dynamically address multiple angles by stacking several layers of PMMA/LC composite gratings. 16 steering angles are obtained with the contrast ratio of 18. A fabrication-compensation trick is proposed to greatly improve the device's performance. In modeling and measurement of optical diffraction from subwavelength structure at focused spot, we describe a technique for studying scattering from subwavelength features that used a focused ion beam system to generate subwavelength, submicron features and a simple scatterometer to measure the scattering from single subwavelength scatterers. A two-dimensional model that handles arbitrary profiles is described. The model is shown to agree quite well with the experimental measurements. The model 15. Physical Optics Based Computational Imaging Systems Science.gov (United States) Olivas, Stephen Joseph There is an ongoing demand on behalf of the consumer, medical and military industries to make lighter weight, higher resolution, wider field-of-view and extended depth-of-focus cameras. This leads to design trade-offs between performance and cost, be it size, weight, power, or expense. This has brought attention to finding new ways to extend the design space while adhering to cost constraints. Extending the functionality of an imager in order to achieve extraordinary performance is a common theme of computational imaging, a field of study which uses additional hardware along with tailored algorithms to formulate and solve inverse problems in imaging. This dissertation details four specific systems within this emerging field: a Fiber Bundle Relayed Imaging System, an Extended Depth-of-Focus Imaging System, a Platform Motion Blur Image Restoration System, and a Compressive Imaging System. The Fiber Bundle Relayed Imaging System is part of a larger project, where the work presented in this thesis was to use image processing techniques to mitigate problems inherent to fiber bundle image relay and then, form high-resolution wide field-of-view panoramas captured from multiple sensors within a custom state-of-the-art imager. The Extended Depth-of-Focus System goals were to characterize the angular and depth dependence of the PSF of a focal swept imager in order to increase the acceptably focused imaged scene depth. The goal of the Platform Motion Blur Image Restoration System was to build a system that can capture a high signal-to-noise ratio (SNR), long-exposure image which is inherently blurred while at the same time capturing motion data using additional optical sensors in order to deblur the degraded images. Lastly, the objective of the Compressive Imager was to design and build a system functionally similar to the Single Pixel Camera and use it to test new sampling methods for image generation and to characterize it against a traditional camera. These computational 16. Fiber Optic Temperature Sensors for Thermal Protection Systems Project Data.gov (United States) National Aeronautics and Space Administration — In Phase 1, Intelligent Fiber Optic Systems Corporation (IFOS), in collaboration with North Carolina State University, successfully demonstrated a Fiber Bragg... 17. Matrix optical system for plane-point correlation Science.gov (United States) Curatu, Eugen O. 1991-12-01 Matrices of optical components (like LEDs, microlenses, microzone plates, microspatial filters) are used in some arrangements to obtain the optical transformation from an area of a frontal plane to a point and vice versa. By use of these matrices the following arrangement are established: a matrix optical relay-system, a multichannel optoelectronic hybrid computer, an optical system for pattern recognition by spatial multichannel filtering, and uniform illumination system. The established relationships are applied to calculate the positions, the component sizes, and the system performance. Matrix configuration is shown to be advantageous for maximizing the performance for plane-point correlation and, also, for a miniature solution. 18. Electro-Optical Data Acquisition and Tracking System Data.gov (United States) Federal Laboratory Consortium — The Electro-Optical Data Acquisition and Tracking System (EDATS) dynamically tracks and measures target signatures. It consists of an instrumentation van integrated... 19. Optical manipulation with two beam traps in microfluidic polymer systems DEFF Research Database (Denmark) Khoury Arvelo, Maria; Matteucci, Marco; Sørensen, Kristian Tølbøl 2015-01-01 An optical trapping system with two opposing laser beams, also known as the optical stretcher, are naturally constructed inside a microfluidic lab-on-chip system. We present and compare two approaches to combine a simple microfluidic system with either waveguides directly written in the microflui... 20. Contamination effects of oil paint on the space optical system Science.gov (United States) Lu, Chun-lian; Zhou, Yan-ping; Sang, Yi 2008-03-01 Space environment in which the space optical system exposed includes space vacuum, cryogenics, and energetic particles and etc. The contamination effects on the space optical system, so the optical damage occurs. And the image quality of the system will be affected. In this paper, the effects of contamination to optical system were discussed, and the contamination source and its space distribution were analyzed. An experiment was designed to determine the effect value. Numeral fitting method was used to analyze the relationship of the optical damage factor (Transmissivity decay factor) and the contamination degree of the optical system. In the experiment, the contamination degree was expressed by mass thickness with the unit μg/mm2. Oil paint was used as the contamination source. Comparison between previous and present researches was given. 1. Railroad Classification Yard Technology Manual: Volume II : Yard Computer Systems Science.gov (United States) 1981-08-01 This volume (Volume II) of the Railroad Classification Yard Technology Manual documents the railroad classification yard computer systems methodology. The subjects covered are: functional description of process control and inventory computer systems,... 2. Problems of Aero-optics and Adaptive Optical Systems: Analytical Review Directory of Open Access Journals (Sweden) Yu. I. Shanin 2017-01-01 3. SPECIAL ASPECTS OF INITIAL OPTICAL SCHEME SELECTION FOR DESIGN OF NON-IMAGING OPTICAL SYSTEMS Directory of Open Access Journals (Sweden) R. V. Anitropov 2016-01-01 Full Text Available Subject of Research. The research results, structural composition analysis and the parametric synthesis of the projected imaging and non-imaging optical systems were presented. We made an attempt to use the gained experience about imaging systems while designing non-imaging systems, by adapting the composition theory for the calculations of non-imaging systems. Several patterns were revealed, which provide a deeper understanding of the design process of non-imaging optical systems; measures of its optimization were proposed. Method. We investigated the applicability of the theory of composition and synthesis of non-imaging optical systems. The main provisions of the theory of composition are based on the division of all available optical elements in four types depending on their functionality, which corresponds to a modular design. Similar items were identified in non-imaging optical systems and adaptation of composition theory to their design became possible. Main Results. General design patterns of imaging and non-imaging optical systems were studied. Classification of systems, components, as well as technical and generic characteristics of imaging and non-imaging optical systems was determined. Search mechanism of the initial optical system by means of structural and parametric synthesis of non-imaging optical system was formalized. The basic elements were determined included in non-imaging systems and their classification by functionality was done. They were subdivided into basic, corrective, wide angle and high aperture ones. The rules for formation of these elements and their composition were determined: surface reflecting, refracting, spherical and nonspherical elements with total internal reflection. The foundations of composition theory for non-imaging optical systems were laid. The approbation of this method was carried out on the example of the illumination system calculation for surgical room. A 3D model of an illumination optical 4. The at-wavelength metrology facility for UV- and XUV-reflection and diffraction optics at BESSY-II Energy Technology Data Exchange (ETDEWEB) Schäfers, F., E-mail: franz.schaefers@helmholtz-berlin.de; Bischoff, P.; Eggenstein, F.; Erko, A.; Gaupp, A.; Künstner, S.; Mast, M.; Schmidt, J.-S.; Senf, F.; Siewert, F.; Sokolov, A.; Zeschke, Th. [Helmholtz-Zentrum Berlin, Albert-Einstein-Strasse 15, Berlin 12489 (Germany) 2016-01-01 A new optics beamline and a versatile 11-axes UHV-reflectometer for at-wavelength characterization of real life-sized UV- and XUV-reflection gratings and other (nano-) optical elements has been set up and is in operation at BESSY-II. Azimuthal rotation of samples allows for reflectometry and polarimetry measurements in s- and p-polarization. A technology center for the production of high-precision reflection gratings has been established. Within this project a new optics beamline and a versatile reflectometer for at-wavelength characterization of UV- and XUV-reflection gratings and other (nano-) optical elements has been set up at BESSY-II. The Plane Grating Monochromator beamline operated in collimated light (c-PGM) is equipped with an SX700 monochromator, of which the blazed gratings (600 and 1200 lines mm{sup −1}) have been recently exchanged for new ones of improved performance produced in-house. Over the operating range from 10 to 2000 eV this beamline has very high spectral purity achieved by (i) a four-mirror arrangement of different coatings which can be inserted into the beam at different angles and (ii) by absorber filters for high-order suppression. Stray light and scattered radiation is removed efficiently by double sets of in situ exchangeable apertures and slits. By use of in- and off-plane bending-magnet radiation the beamline can be adjusted to either linear or elliptical polarization. One of the main features of a novel 11-axes reflectometer is the possibility to incorporate real life-sized gratings. The samples are adjustable within six degrees of freedom by a newly developed UHV-tripod system carrying a load up to 4 kg, and the reflectivity can be measured between 0 and 90° incidence angle for both s- and p-polarization geometry. This novel powerful metrology facility has gone into operation recently and is now open for external users. First results on optical performance and measurements on multilayer gratings will be presented here. 5. Component configuration control system development at EBR-II Energy Technology Data Exchange (ETDEWEB) Monson, L.R.; Stratton, R.C. 1984-01-01 One ofthe major programs being pursued by the EBR-II Division of Argonne National Laboratory is to improve the reliability of plant control and protection systems. This effort involves looking closely at the present state of the art and needs associated with plant diagnostic, control and protection systems. One of the areas of development at EBR-II involves a component configuration control system (CCCS). This system is a computerized control and planning aid for the nuclear power operator. 6. Characterization of low cost optical imaging system based on optical holography Science.gov (United States) Darwiesh, M.; El Sherif, Ashraf F. 2013-03-01 Optical imaging systems are widely used in different applications including tracking for portable scanners; input pointing devices for laptop computers, cell phones, and cameras; and fingerprint-identification scanners. Also in optical navigation (military target tracking where tracking sensors follow airplanes, missiles, and other targets [1-4]. Since the two main parameters affecting the performance of the optical imaging systems are the optical source and the surface nature. So; The aim of the paper is to study how the optical source affects the performance of the optical imaging systems by exchange the operating surface of the optical imaging system with a standard diffuse object (Gabbor holograms) to study and analyze laser speckle pattern and Circular interference fringes produced by illuminating these standard diffusers using different optical sources [coherent (3mW diode laser, and 10mW He-Ne laser) or partially coherent light (LEDs)]. The Circular interference fringes were used to display the relations between the fringes order and its radii. From these relations we found that the electronic sensor can deliver the same accuracy of laser diodes when replacing it by commercial LEDs. So, we can design a new cheaper, high performance optical imaging system using commercial LED sources. 7. Comparison of APACHE II and SAPS II Scoring Systems in Prediction of Critically Ill Patients' Outcome. Science.gov (United States) Aminiahidashti, Hamed; Bozorgi, Farzad; Montazer, Seyyed Hosein; Baboli, Majid; Firouzian, Abolfazl 2017-01-01 Using physiologic scoring systems for identifying high-risk patients for mortality has been considered recently. This study was designed to evaluate the values of Acute Physiology and Chronic Health Evaluation II (APACHE II) and Simplified Acute Physiologic Score (SAPS II) models in prediction of 1-month mortality of critically ill patients. The present prospective cross sectional study was performed on critically ill patients presented to emergency department during 6 months. Data required for calculation of the scores were gathered and performance of the models in prediction of 1-month mortality were assessed using STATA software 11.0. 82 critically ill patients with the mean age of 53.45 ± 20.37 years were included (65.9% male). Their mortality rate was 48%. Mean SAPS II (p APACHE II (p = 0.0007) scores were significantly higher in dead patients. Area under the ROC curve of SAPS II and APACHE II for prediction of mortality were 0.75 (95% CI: 0.64 - 0.86) and 0.72 (95% CI: 0.60 - 0.83), respectively (p = 0.24). The slope and intercept of SAPS II were 1.02 and 0.04, respectively. In addition, these values were 0.92 and 0.09 for APACHE II, respectively. The findings of the present study showed that APACHE II and SAPS II had similar value in predicting 1-month mortality of patients. Discriminatory powers of the mentioned models were acceptable but their calibration had some amount of lack of fit, which reveals that APACHE II and SAPS II are partially perfect. 8. Sensitivity improvement of optical fiber acoustic probe for all-optical photoacoustic imaging system Science.gov (United States) Seki, Atsushi; Iwai, Katsumasa; Katagiri, Takashi; Matsuura, Yuji 2017-07-01 An acoustic probe based on a Fabry-Perot interferometer composed of a polymer film attached to the end of an optical fiber was designed and fabricated for an endoscopic, all-optical photoacoustic imaging system. The finesse of the interferometer was improved by forming a half-mirror at the end of the fiber and a partial reflection mirror on the outer surface of the polymer film. A photoacoustic imaging system was constructed by combining the fiber-optic acoustic probe with a bundle of hollow optical fibers used for the excitation of the photoacoustic wave, and an image of blood capillaries in a fish ovarian membrane was successfully obtained. 9. Optical-disk-based imaging system to be used as an optical microscope Science.gov (United States) Shima, Takayuki; Fujimaki, Makoto; Awazu, Koichi 2016-07-01 An optical disk surface is scanned spirally by laser light, as in the case of digital versatile discs, and a reflectance image is formed by rearranging the scanned intensity results. A prototype system is developed for imaging with a rotary encoder equipped to precisely control the disk rotation angle. We measured Escherichia coli dispersed on an optical disk sample surface and successfully obtained an image that is identical to that obtained using an optical microscope. The system is advantageous as an optical sensor for detecting sub-micrometer- to micrometer-order substances on a large-area surface. 10. Optical coordinate tracking system using afocal optics for image-guided surgery. Science.gov (United States) Chae, You Seong; Lee, Seung Hyun; Lee, Hyun Ki; Kim, Min Young 2015-02-01 Image-guided surgery using medical robots supports surgeons by providing critical real-time feedback information, such as surgical instrument tracking, patient-specific models, and the use of surgery robots. An image-guided surgery system based on afocal optics was developed to overcome the problems associated with conventional optical tracking systems. An optical tracking system was developed that utilizes afocal optics. Instead of using geometrically specified marker spheres as tracking targets, the proposed system uses a marker with a lens and a micro-engraved data-coded pattern. A position and orientation-tracking algorithm was developed to utilize the observed afocal images of the marker patterns. The marker used in this tracking system can be manufactured in a smaller size than traditional optical tracker markers, and the accuracy of the proposed tracking system has significant potential for improvement due to its focused and highly magnified image. The system was tested in vitro on an optical bench with position and orientation measurement experiments using a commercial optical tracker, Polaris Vicra (NDI Corp.) for comparison. The afocal optical system provided accuracy in position and orientation that was equal or better than a commercial optical tracker system, and provided a high degree of consistency during in vitro testing. The position error was 21μ m, and the orientation error was 0.093°. An afocal optical tracker is feasible and potentially advantageous for surgical navigation, as it is expected to have fewer occlusions and provide greater efficiency for coordinate matching and tracking of patient-specific models, surgical instruments, and surgery robots. This promising new system requires in vivo testing. 11. Present status of the TJ-II remote participation system Energy Technology Data Exchange (ETDEWEB) 2005-11-15 The TJ-II remote participation system (RPS) was designed to extend to Internet the working capabilities provided in the TJ-II local environment, i.e., tracking the TJ-II operation, monitoring/programming data acquisition and control systems, and accessing databases. The TJ-II RPS was based on web and Java technologies because of their open character, security properties and technological maturity. A web server acts as a communication front-end between remote participants and local TJ-II elements. From the server side, web services are provided by means of resources supplied by JSP pages. The client part makes use of web browsers and ad hoc Java applications. The operation requires the use of a distributed authentication and authorization system. This development employs the PAPI System. At present, approximately 1000 digitisation channels can be managed from the TJ-II RPS. Furthermore, processing software based on a 4GL language (LabView) can be downloaded to multiprocessor data acquisition systems. Also, 15 diagnostic control systems, databases and the operation logbook are available from the RPS. The system even allows for the physicist in charge of operation to be in a remote location. Four Spanish universities make use of the TJ-II remote participation system capabilities for joint collaborations: these are the Universidad Politecnica de Madrid (UPM), Universidad Nacional de Educacion a Distancia (UNED), Universidad Complutense de Madrid (UCM) and Universidad Politecnica de Cataluna (UPC) 12. Organo-erbium systems for optical amplification at telecommunications wavelengths. Science.gov (United States) Ye, H Q; Li, Z; Peng, Y; Wang, C C; Li, T Y; Zheng, Y X; Sapelkin, A; Adamopoulos, G; Hernández, I; Wyatt, P B; Gillin, W P 2014-04-01 Modern telecommunications rely on the transmission and manipulation of optical signals. Optical amplification plays a vital part in this technology, as all components in a real telecommunications system produce some loss. The two main issues with present amplifiers, which rely on erbium ions in a glass matrix, are the difficulty in integration onto a single substrate and the need of high pump power densities to produce gain. Here we show a potential organic optical amplifier material that demonstrates population inversion when pumped from above using low-power visible light. This system is integrated into an organic light-emitting diode demonstrating that electrical pumping can be achieved. This opens the possibility of direct electrically driven optical amplifiers and optical circuits. Our results provide an alternative approach to producing low-cost integrated optics that is compatible with existing silicon photonics and a different route to an effective integrated optics technology. 13. Optical wireless communications system and channel modelling with Matlab CERN Document Server Ghassemlooy, Z 2012-01-01 Detailing a systems approach, Optical Wireless Communications: System and Channel Modelling with MATLAB(R), is a self-contained volume that concisely and comprehensively covers the theory and technology of optical wireless communications systems (OWC) in a way that is suitable for undergraduate and graduate-level students, as well as researchers and professional engineers. Incorporating MATLAB(R) throughout, the authors highlight past and current research activities to illustrate optical sources, transmitters, detectors, receivers, and other devices used in optical wireless communications. The 14. Drawing inspiration from biological optical systems Science.gov (United States) Wolpert, H. D. 2009-08-01 Bio-Mimicking/Bio-Inspiration: How can we not be inspired by Nature? Life has evolved on earth over the last 3.5 to 4 billion years. Materials formed during this time were not toxic; they were created at low temperatures and low pressures unlike many of the materials developed today. The natural materials formed are self-assembled, multifunctional, nonlinear, complex, adaptive, self-repairing and biodegradable. The designs that failed are fossils. Those that survived are the success stories. Natural materials are mostly formed from organics, inorganic crystals and amorphous phases. The materials make economic sense by optimizing the design of the structures or systems to meet multiple needs. We constantly "see" many similar strategies in approaches, between man and nature, but we seldom look at the details of natures approaches. The power of image processing, in many of natures creatures, is a detail that is often overlooked. Seldon does the engineer interact with the biologist and learn what nature has to teach us. The variety and complexity of biological materials and the optical systems formed should inspire us. 15. HPC Colony II: FAST_OS II: Operating Systems and Runtime Systems at Extreme Scale Energy Technology Data Exchange (ETDEWEB) Moreira, Jose [IBM, Armonk, NY (United States) 2013-11-13 HPC Colony II has been a 36-month project focused on providing portable performance for leadership class machines—a task made difficult by the emerging variety of more complex computer architectures. The project attempts to move the burden of portable performance to adaptive system software, thereby allowing domain scientists to concentrate on their field rather than the fine details of a new leadership class machine. To accomplish our goals, we focused on adding intelligence into the system software stack. Our revised components include: new techniques to address OS jitter; new techniques to dynamically address load imbalances; new techniques to map resources according to architectural subtleties and application dynamic behavior; new techniques to dramatically improve the performance of checkpoint-restart; and new techniques to address membership service issues at scale. 16. Optical two-beam traps in microfluidic systems DEFF Research Database (Denmark) Berg-Sørensen, Kirstine 2016-01-01 An attractive solution for optical trapping and stretching by means of two counterpropagating laser beams is to embed waveguides or optical fibers in a microfluidic system. The microfluidic system can be constructed in different materials, ranging from soft polymers that may easily be cast in a r...... written waveguides and in an injection molded polymer chip with grooves for optical fibers. (C) 2016 The Japan Society of Applied Physics.... 17. Quantum Optics with Atom-like Systems in Diamond Science.gov (United States) 2013-11-19 29 2.4.1 Optical Rabi oscillations and resonant spin readout . . . . . . 31 2.4.2 Multi-level systems and all-optical spin...external electromagnetic fields. In addition, the atoms are trapped and cooled using ion traps or optical lattices to reduce the effect of atomic motion...the diamond lattice . We do this by making use of the same Λ type system involved in generating spin-photon entanglement. Such a level structure allows 18. The Belle II PXD data acquisition and reduction system Energy Technology Data Exchange (ETDEWEB) Lange, Soeren; Gessler, Thomas; Kuehn, Wolfgang; Muenchow, David; Spruck, Bjoern [Giessen Univ. (Germany). 2. Physikalisches Inst.; Lin, Haichuan; Liu, Zhen-An; Xu, Hao; Zhao, Jingzhou [IHEP Beijing (China); Collaboration: Belle II Collaboration 2013-07-01 The Belle II DEPFET pixel detector (PXD) will deliver high data rates of up to 21.6 Gbytes/s for 3% detector occupancy. Data of this high rate must be buffered for 5 seconds, corresponding to the HLT (High Level Trigger) latency, and then a region-of-interest (ROI) filter is applied to reduce the data rate by a factor of ≥10 by charged track extrapolation from other detectors (SVD, CDC). The PXD readout system is based upon ATCA (Advanced Telecommunications Architecture). The 3rd PCB iteration uses a concept with a xTCA carrier board (with a Virtex-4 FX60 FPGA for ATCA backplane routing) and 4 AMC modules (each with a Xilinx Virtex-5 FX70T FPGA). The FPGA firmware implementation comprises a receiver core for the high speed optical links (≤6.25 Gbps), a buffer management with lookup of ≤270.000 pointers/s, DDR2 memory write (native port interface, ≥1.5 Gbytes/s), Gigabit ethernet (UDP stack in VHDL) and a parallelized ROI selection algorithm. Test results of all the components are presented. 19. Power selective optical filter devices and optical systems using same Science.gov (United States) Koplow, Jeffrey P 2014-10-07 In an embodiment, a power selective optical filter device includes an input polarizer for selectively transmitting an input signal. The device includes a wave-plate structure positioned to receive the input signal, which includes at least one substantially zero-order, zero-wave plate. The zero-order, zero-wave plate is configured to alter a polarization state of the input signal passing in a manner that depends on the power of the input signal. The zero-order, zero-wave plate includes an entry and exit wave plate each having a fast axis, with the fast axes oriented substantially perpendicular to each other. Each entry wave plate is oriented relative to a transmission axis of the input polarizer at a respective angle. An output polarizer is positioned to receive a signal output from the wave-plate structure and selectively transmits the signal based on the polarization state. 20. Optical fiber oxygen sensor based on Pd(II) complex embedded in sol-gel matrix Energy Technology Data Exchange (ETDEWEB) Chu, Cheng-Shane, E-mail: cschu@mail.mcut.edu.tw [Department of Mechanical Engineering, Ming Chi University of Technology, Taishan Dist., New Taipei City, Taiwan (China) 2013-03-15 A simple, low-cost technique for fabrication of high performance optical fiber oxygen sensor is described. An organically modified silicate (ORMOSIL) as a matrix for the fabrication of oxygen sensing film was produced. The technique is based on coating the end of an optical fiber with ormosil composite xerogel film sequestered with luminophore palladium (II) meso-tetrakis(pentafluorophenyl)porphyrin (PdTFPP) prepared by a sol-gel process. The composite xerogel studied is n-propyltrimethoxysilane (n-propyl-TriMOS)/tetraethylorthosilane (TEOS)/n-Octyltriethoxysilane (Octyl-triEOS). Result shows that, expect for PdTFPP-doped n-propyl-TriMOS/TEOS/Octyl-triEOS composite xerogel shows a high sensitivity and linear Stern-Volmer relationship which indicates the homogenous environment of the luminophore. The sensitivity of the optical oxygen sensor is quantified in terms of the ratio I{sub N{sub 2}}/I{sub O{sub 2}}, where I{sub N{sub 2}} and I{sub O{sub 2}} represent the detected phosphorescence intensities in pure nitrogen and pure oxygen environments, respectively. The experimental result reveals that the PdTFPP-doped n-propyl-TriMOS/TEOS/Octyl-triEOS oxygen sensor has sensitivity of I{sub N{sub 2}}/I{sub 100O{sub 2}}=263. - Highlights: Black-Right-Pointing-Pointer A simple technique for fabrication of high performance optical fiber oxygen sensor is described. Black-Right-Pointing-Pointer ORMOSIL was produced to serve as a matrix for the fabrication of oxygen sensing film. Black-Right-Pointing-Pointer The optical fiber oxygen sensor has sensitivity of I{sub N{sub 2}}/I{sub 100O{sub 2}}=263. Black-Right-Pointing-Pointer The stable and reproducible signals were obtained with the fiber-optic oxygen sensor. 1. Development of Optical System for ARGO-M Directory of Open Access Journals (Sweden) Jakyoung Nah 2013-03-01 Full Text Available ARGO-M is a satellite laser ranging (SLR system developed by the Korea Astronomy and Space Science Institute with the consideration of mobility and daytime and nighttime satellite observation. The ARGO-M optical system consists of 40 cm receiving telescope, 10 cm transmitting telescope, and detecting optics. For the development of ARGO-M optical system, the structural analysis was performed with regard to the optics and optomechanics design and the optical components. To ensure the optical performance, the quality was tested at the level of parts using the laser interferometer and ultra-high-precision measuring instruments. The assembly and alignment of ARGO-M optical system were conducted at an auto-collimation facility. As the transmission and reception are separated in the ARGO-M optical system, the pointing alignment between the transmitting telescope and receiving telescope is critical for precise target pointing. Thus, the alignment using the ground target and the radiant point observation of transmitting laser beam was carried out, and the lines of sight for the two telescopes were aligned within the required pointing precision. This paper describes the design, structural analysis, manufacture and assembly of parts, and entire process related with the alignment for the ARGO-M optical system. 2. Optical link card design for the phase II upgrade of TileCal experiment CERN Document Server Carrio, F; Ferrer, A; Gonzalez, V; Higon, E; Marin, C; Moreno, P; Sanchis, E; Solans, C; Valero, A; Valls, J 2011-01-01 This paper presents the design of an optical link card developed in the frame of the R&D activities for the phase 2 upgrade of the TileCal experiment. This board, that is part of the evaluation of different technologies for the final choice in the next years, is designed as a mezzanine that can work independently or be plugged in the optical multiplexer board of the TileCal backend electronics. It includes two SNAP 12 optical connectors able to transmit and receive up to 75 Gb/s and one SFP optical connector for lower speeds and compatibility with existing hardware as the read out driver. All processing is done in a Stratix II GX field-programmable gate array (FPGA). Details are given on the hardware design, including signal and power integrity ana lysis, needed when working with these high data rates and on firmware development to obtain the best performance of the FPGA signal transceivers and for the use of the GBT protocol. 3. Programme Code for Projecting of WDM Fiber Optic Sensor Systems Directory of Open Access Journals (Sweden) R. Probstner 1993-04-01 Full Text Available Wavelength division multiplex (WDM offers a potentially powerful technique for use within optical fibre sensor systems. The paper deals with short description of methodology and a programme code for WDM fiber optic sensor system projecting with use of CAD. 4. An Image Stabilization Optical System Using Deformable Freeform Mirrors Directory of Open Access Journals (Sweden) Qun Hao 2015-01-01 Full Text Available An image stabilization optical system using deformable freeform mirrors is proposed that enables the ray sets to couple dynamically in the object and image space. It aims to correct image blurring and degradation when there is relative movement between the imaging optical axis and the object. In this method, Fermat’s principle and matrix methods are used to describe the optical path of the entire optical system with a shift object plane and a fixed corresponding image plane in the carrier coordinate system. A constant optical path length is determined for each ray set, so the correspondence between the object and the shift free image point is used to calculate the solution to the points on the surface profile of the deformable mirrors (DMs. Off-axis three-mirror anastigmats are used to demonstrate the benefits of optical image stabilization with one- and two-deformable mirrors. 5. System for optical sorting of microscopic objects DEFF Research Database (Denmark) 2014-01-01 in a first reservoir, the one or more force units being suitable for optical momentum transfer. An electromagnetic radiation source (42) yields a radiation beam (31, 32) capable of optically displacing the force transfer units from one position to another within the first reservoir (1R). The force transfer... 6. BeamOptics : a Symbolic Platform for Modeling and the Solution of Beam Optics System Energy Technology Data Exchange (ETDEWEB) Yu-Chiu Chao 2000-11-01 BeamOptics [1] is a Mathematica-based computing platform devoted to the following objectives: (1) Structured representation and manipulation of particle beam optics systems with symbolic capabilities, (2) Analytical and numerical modeling of beam optics system behaviors, (3) Solution to specific beam optical or general accelerator system problems, in algebraic form in certain cases, through customized algorithms. Taking advantage of and conforming to the highly formal and self-contained structure of Mathematica, BeamOptics provides a unique platform for developing accelerator design and analysis programs. The feature of symbolic computation and the ability to manipulate the beam optics system at the programming language level enable the user to solve or optimize his system with considerably more efficiency, rigour and insight than can be easily achieved with passive modeling or numerical simulation methods. BeamOptics is developed with continuous evolution in mind. New features and algorithms from diverse sources can be incorporated without major modification, due to its formal and generic structure. In this report, a survey is given of the basic structure and methodology of BeamOptics, as well as a demonstration of some of its more specialized applications, and possible direction of evolution. 7. Coordination Chemistry of SCS PdII Pincer Systems NARCIS (Netherlands) van Manen, H.J.; Nakashima, Kazuaki; Nakashima, K.; Shinkai, Seiji; Kooijman, Huub; Spek, Anthony L.; van Veggel, F.C.J.M.; Reinhoudt, David 2000-01-01 We have studied the coordination of substituted pyridines, and phosphorus- and sulfur-containing ligands to an SCS PdII pincer system. These ligands coordinate to PdII (trans to the cyclopalladated aryl group) by quantitative substitution of the labile acetonitrile ligand in complex 1. Competition 8. Optical Systems of Biopsy: The Invisible Eye Directory of Open Access Journals (Sweden) Kriti Bagri-Manjrekar 2012-01-01 Full Text Available The exploration of new methods and techniques for the diagnosis of malignant tumours has always attracted the attention of scientists. The development of adjunct tools to facilitate the non invasive screening of high risk lesions in real time has the potential to significantly improve our ability to reduce the dismal morbidity and mortality of oral cancer. Despite easy accessibility of the oral cavity to examination, there is no satisfactory method to adequately screen and detect precancers non-invasively. The current method of oral cancer diagnosis clinically relies heavily on visual examination of the oral cavity. However, discerning potentially malignant and early malignant lesions from common benign inflammatory conditions can be difficult at times. There is a need for an objective method that could provide real- time results and be routinely applied to a large population. Though science is yet to present such a perfect technique, Optical Biopsy Systems developed using knowledge of light and tissue interaction, can provide a plausible option. 9. Raman amplification in optical communication systems DEFF Research Database (Denmark) Kjær, Rasmus 2008-01-01 the length of future access networks, and inter-channel transient crosstalk and a new all-optical method to remedy these unwanted effects. Finally, a numerical amplifier model is developed to better understand the dynamics, gain and noise limitations of the amplifier. I denne afhandling undersøges...... støjtal under 4,5 dB og en samlet udgangseffekt på 22 dBm. Med henblik på at forlænge rækkevidden af fremtidige access-netværk foreslås en ny arkitektur for såkaldte langdistance passive optiske netværk (PON). Dette system evalueres både teoretisk og eksperimentelt. Distribueret Raman-forstærkning bruges...... modtagerfølsomheden p.g.a. den bidirektionelle trafik. De dynamiske egenskaber af PON-forbindelsen testes med forstyrrende trafik og som konklusion vurderes systemet at være en lovende kandidat til fremtidige langdistance access PONs med dynamisk trafik. Krydstale mellem to eller flere kanaler i mættede... 10. The at-wavelength metrology facility for UV- and XUV-reflection and diffraction optics at BESSY-II Science.gov (United States) Schäfers, F.; Bischoff, P.; Eggenstein, F.; Erko, A.; Gaupp, A.; Künstner, S.; Mast, M.; Schmidt, J.-S.; Senf, F.; Siewert, F.; Sokolov, A.; Zeschke, Th. 2016-01-01 A technology center for the production of high-precision reflection gratings has been established. Within this project a new optics beamline and a versatile reflectometer for at-wavelength characterization of UV- and XUV-reflection gratings and other (nano-) optical elements has been set up at BESSY-II. The Plane Grating Monochromator beamline operated in collimated light (c-PGM) is equipped with an SX700 monochromator, of which the blazed gratings (600 and 1200 lines mm−1) have been recently exchanged for new ones of improved performance produced in-house. Over the operating range from 10 to 2000 eV this beamline has very high spectral purity achieved by (i) a four-mirror arrangement of different coatings which can be inserted into the beam at different angles and (ii) by absorber filters for high-order suppression. Stray light and scattered radiation is removed efficiently by double sets of in situ exchangeable apertures and slits. By use of in- and off-plane bending-magnet radiation the beamline can be adjusted to either linear or elliptical polarization. One of the main features of a novel 11-axes reflectometer is the possibility to incorporate real life-sized gratings. The samples are adjustable within six degrees of freedom by a newly developed UHV-tripod system carrying a load up to 4 kg, and the reflectivity can be measured between 0 and 90° incidence angle for both s- and p-polarization geometry. This novel powerful metrology facility has gone into operation recently and is now open for external users. First results on optical performance and measurements on multilayer gratings will be presented here. PMID:26698047 11. The at-wavelength metrology facility for UV- and XUV-reflection and diffraction optics at BESSY-II. Science.gov (United States) Schäfers, F; Bischoff, P; Eggenstein, F; Erko, A; Gaupp, A; Künstner, S; Mast, M; Schmidt, J-S; Senf, F; Siewert, F; Sokolov, A; Zeschke, Th 2016-01-01 A technology center for the production of high-precision reflection gratings has been established. Within this project a new optics beamline and a versatile reflectometer for at-wavelength characterization of UV- and XUV-reflection gratings and other (nano-) optical elements has been set up at BESSY-II. The Plane Grating Monochromator beamline operated in collimated light (c-PGM) is equipped with an SX700 monochromator, of which the blazed gratings (600 and 1200 lines mm(-1)) have been recently exchanged for new ones of improved performance produced in-house. Over the operating range from 10 to 2000 eV this beamline has very high spectral purity achieved by (i) a four-mirror arrangement of different coatings which can be inserted into the beam at different angles and (ii) by absorber filters for high-order suppression. Stray light and scattered radiation is removed efficiently by double sets of in situ exchangeable apertures and slits. By use of in- and off-plane bending-magnet radiation the beamline can be adjusted to either linear or elliptical polarization. One of the main features of a novel 11-axes reflectometer is the possibility to incorporate real life-sized gratings. The samples are adjustable within six degrees of freedom by a newly developed UHV-tripod system carrying a load up to 4 kg, and the reflectivity can be measured between 0 and 90° incidence angle for both s- and p-polarization geometry. This novel powerful metrology facility has gone into operation recently and is now open for external users. First results on optical performance and measurements on multilayer gratings will be presented here. 12. Optical systems fabricated by printing-based assembly Energy Technology Data Exchange (ETDEWEB) Rogers, John; Nuzzo, Ralph; Meitl, Matthew; Menard, Etienne; Baca, Alfred; Motala, Michael; Ahn, Jong-Hyun; Park, Sang-Il; Yu, Chang-Jae; Ko, Heung Cho; Stoykovich, Mark; Yoon, Jongseung 2017-03-21 Provided are optical devices and systems fabricated, at least in part, via printing-based assembly and integration of device components. In specific embodiments the present invention provides light emitting systems, light collecting systems, light sensing systems and photovoltaic systems comprising printable semiconductor elements, including large area, high performance macroelectronic devices. Optical systems of the present invention comprise semiconductor elements assembled, organized and/or integrated with other device components via printing techniques that exhibit performance characteristics and functionality comparable to single crystalline semiconductor based devices fabricated using conventional high temperature processing methods. Optical systems of the present invention have device geometries and configurations, such as form factors, component densities, and component positions, accessed by printing that provide a range of useful device functionalities. Optical systems of the present invention include devices and device arrays exhibiting a range of useful physical and mechanical properties including flexibility, shapeability, conformability and stretchablity. 13. A novel collinear optical setup for holographic data storage system Science.gov (United States) Horimai, Hideyoshi; Li, Jun 2004-09-01 In this paper, A novel collinear optical setup for holographic data storage system is presented. Simulated/experimental results are given. Combined with sub-page based data format, the system is simple and robust. 14. Control System Performance of a Woofer-Tweeter Adaptive Optics System Science.gov (United States) Hampton, P.; Bradley, C.; Agathoklis, P.; Conan, R. In the next generation of large optical telescope (LOT), it will be essential to use adaptive optics (AO) systems to achieve optimal image quality. Furthermore, in order to meet the broad set of user requirements, several new AO configurations are currently being investigated for use on proposed LOTs. One configuration currently under investigation at the University of Victoria is the Woofer-Tweeter system. The Adaptive Optics Laboratory, at the University of Victoria, has recently completed the development of a test bench for this Woofer-Tweeter concept. This project is part of the Thirty Meter Telescope (TMT) development program that will see completion within the next decade. The actuator density and maximum actuator stroke requirements of deformable mirrors, necessary for the LOTs, increase significantly due to the enormous collecting area of the LOT, operating site seeing, and scientific requirements. It would be cost prohibitive to try and develop a single deformable mirror that satisfies the actuator density and actuator stroke requirements. Fortunately, the large stroke requirement is for the compensation of low spatial frequency distortion. This allows a system to be designed with two deformable mirrors (DMs): (i) a high stroke, low actuator density DM termed the Woofer, and (ii) a low stroke, high actuator density DM termed the Tweeter. The initial simulations and experimental results have shown that the controller developed in this work can appropriately split the correction between the two deformable mirrors. The Woofer, correcting for the low-spatial-low-temporal frequency disturbances, is an electro-mechanical device, whereas the Tweeter, correcting for the remaining disturbance, is a MEMS device. This paper focuses on the development and need for the woofer-tweeter test bench; bench design and operating specifications; the deformable mirror technology employed; closed loop control of the bench system when a hot air turbulence generator is introduced 15. High frame-rate en face optical coherence tomography system using KTN optical beam deflector Science.gov (United States) Ohmi, Masato; Shinya, Yusuke; Imai, Tadayuki; Toyoda, Seiji; Kobayashi, Junya; Sakamoto, Tadashi 2017-02-01 We developed high frame-rate en face optical coherence tomography (OCT) system using KTa1-xNbxO3 (KTN) optical beam deflector. In the imaging system, the fast scanning was performed at 200 kHz by the KTN optical beam deflector, while the slow scanning was performed at 800 Hz by the galvanometer mirror. As a preliminary experiment, we succeeded in obtaining en face OCT images of human fingerprint with a frame rate of 800 fps. This is the highest frame-rate obtained using time-domain (TD) en face OCT imaging. The 3D-OCT image of sweat gland was also obtained by our imaging system. 16. Global coordinates and exact aberration calculations applied to physical optics modeling of complex optical systems Energy Technology Data Exchange (ETDEWEB) Lawrence, G.; Barnard, C.; Viswanathan, V.K. 1986-01-01 Historically, wave optics computer codes have been paraxial in nature. Folded systems could be modeled by ''unfolding'' the optical system. Calculation of optical aberrations is, in general, left for the analyst to do with off-line codes. While such paraxial codes were adequate for the simpler systems being studied 10 years ago, current problems such as phased arrays, ring resonators, coupled resonators, and grazing incidents optics require a major advance in analytical capability. This paper describes extension of the physical optics codes GLAD and GLAD V to include a global coordinate system and exact ray aberration calculations. The global coordinate system allows components to be positioned and rotated arbitrarily. Exact aberrations are calculated for components in aligned or misaligned configurations by using ray tracing to compute optical path differences and diffraction propagation. Optical path lengths between components and beam rotations in complex mirror systems are calculated accurately so that coherent interactions in phased arrays and coupled devices may be treated correctly. 17. Handbook of Optical Systems, Volume 2, Physical Image Formation Science.gov (United States) Singer, Wolfgang; Totzeck, Michael; Gross, Herbert 2005-10-01 Herbert Gross, born in 1955, joined Carl Zeiss in 1982 after finishing his physics degree as specialist for optical design. Since 1995 he has been working as head of the department of optical design, while also teaching as a lecturer in Aalen and Lausanne. The new handbook is an intuitive, didactically elegant approach to the subject of optical systems and is not competed by any other work on the market. The selected board of authors, all reputed industrial experts, guarantee the timeliness of the well coordinated, coherent chapters. The second volume presents a more rigorous physical description of the image formation in optical systems on the basis of first principles. Starting with wave equation and the theory of diffraction, readers are introduced in detail to the Fourier theory of optics, since this is a necessary assumption for an understanding of the finite resolution of optical systems, the basic optical quality criteria, the imaging in three dimensions, the influence of the illumination and the coherence and polarization properties of the light source. In particular, the connection between the geometrical and the wave optical models are explained and readers are able to understand the well-known simulation algorithms used in the calculation of the exact properties of modern optical systems. 18. Accuracy requirements of optical linear algebra processors in adaptive optics imaging systems Science.gov (United States) Downie, John D. 1990-01-01 A ground-based adaptive optics imaging telescope system attempts to improve image quality by detecting and correcting for atmospherically induced wavefront aberrations. The required control computations during each cycle will take a finite amount of time. Longer time delays result in larger values of residual wavefront error variance since the atmosphere continues to change during that time. Thus an optical processor may be well-suited for this task. This paper presents a study of the accuracy requirements in a general optical processor that will make it competitive with, or superior to, a conventional digital computer for the adaptive optics application. An optimization of the adaptive optics correction algorithm with respect to an optical processor's degree of accuracy is also briefly discussed. 19. Geometric optics theory and design of astronomical optical systems using Mathematica CERN Document Server Romano, Antonio 2016-01-01 This text, now in its second edition, presents the mathematical background needed to design many optical combinations that are used in astronomical telescopes and cameras. It uses a novel approach to third-order aberration theory based on Fermat’s principle and the use of particular optical paths (called stigmatic paths) instead of rays, allowing for easier derivation of third-order formulae. Each optical combination analyzed is accompanied by a downloadable Mathematica® notebook that automates its third-order design, eliminating the need for lengthy calculations. The essential aspects of an optical system with an axis of rotational symmetry are introduced first, along with a development of Gaussian optics from Fermat’s principal. A simpler approach to third-order monochromatic aberrations based on both Fermat’s principle and stigmatic paths is then described, followed by a new chapter on fifth-order aberrations and their classification. Several specific optical devices are discussed and analyzed, incl... 20. Optical Tomography System: Charge-coupled Device Linear Image Sensors Directory of Open Access Journals (Sweden) M. Idroas 2010-09-01 Full Text Available This paper discussed an optical tomography system based on charge-coupled device (CCD linear image sensors. The developed system consists of a lighting system, a measurement section and a data acquisition system. Four CCD linear image sensors are configured around a flow pipe with an octagonal-shaped measurement section, for a four projections system. The four CCD linear image sensors consisting of 2048 pixels with a pixel size of 14 micron by 14 micron are used to produce a high-resolution system. A simple optical model is mapped into the system’s sensitivity matrix to relate the optical attenuation due to variations of optical density within the measurement section. A reconstructed tomographic image is produced based on the model using MATLAB software. The designed instrumentation system is calibrated and tested through different particle size measurements from different projections. 1. An Optical Receiver Post Processing System for the Integrated Radio and Optical Communications Software Defined Radio Test Bed Science.gov (United States) Nappier, Jennifer M.; Tokars, Roger P.; Wroblewski, Adam C. 2016-01-01 The Integrated Radio and Optical Communications (iROC) project at the National Aeronautics and Space Administrations (NASA) Glenn Research Center is investigating the feasibility of a hybrid radio frequency (RF) and optical communication system for future deep space missions. As a part of this investigation, a test bed for a radio frequency (RF) and optical software defined radio (SDR) has been built. Receivers and modems for the NASA deep space optical waveform are not commercially available so a custom ground optical receiver system has been built. This paper documents the ground optical receiver, which is used in order to test the RF and optical SDR in a free space optical communications link. 2. An Optical Receiver Post-Processing System for the Integrated Radio and Optical Communications Software Defined Radio Test Bed Science.gov (United States) Nappier, Jennifer M.; Tokars, Roger P.; Wroblewski, Adam C. 2016-01-01 The Integrated Radio and Optical Communications (iROC) project at the National Aeronautics and Space Administration's (NASA) Glenn Research Center is investigating the feasibility of a hybrid radio frequency (RF) and optical communication system for future deep space missions. As a part of this investigation, a test bed for a radio frequency (RF) and optical software defined radio (SDR) has been built. Receivers and modems for the NASA deep space optical waveform are not commercially available so a custom ground optical receiver system has been built. This paper documents the ground optical receiver, which is used in order to test the RF and optical SDR in a free space optical communications link. 3. Turnable Semiconductor Laser Spectroscopy in Hollow Optical Waveguides, Phase II SBIR Energy Technology Data Exchange (ETDEWEB) Gregory J. Fetzer, Ph.D. 2001-12-24 In this study a novel optical trace gas sensor based on a perforated hollow waveguide (PHW) was proposed. The sensor has been given the acronym ESHOW for Environmental Sensor using Hollow Optical Waveguides. Realizations of the sensor have demonstrated rapid response time (<2s), low minimum detection limits (typically around 3 x 10-5 absorbance). Operation of the PHW technology has been demonstrated in the near-infrared (NIR) and mid0infrared (MIR) regions of the spectrum. Simulation of sensor performance provided in depth understanding of the signals and signal processing required to provide high sensitivity yet retain rapid response to gas changes. A dedicated sensor electronics and software foundation were developed during the course of the Phase II effort. Commercial applications of the sensor are ambient air and continuous emissions monitoring, industrial process control and hazardous waste site monitoring. There are numerous other applications for such a sensor including medical diagnosis and treatment, breath analysis for legal purposes, water quality assessment, combustion diagnostics, and chemical process control. The successful completion of Phase II resulted in additional funding of instrument development by the Nations Institute of Heath through a Phase I SBIR grant and a strategic teaming relationship with a commercial manufacture of medical instrumentation. The purpose of the NIH grant and teaming relationship is to further develop the sensor to monitor NO in exhaled breath for the purposes of asthma diagnosis. 4. Application of Optical Forces in Microphotonic Systems Science.gov (United States) 2013-05-01 light enhancement of radiation pressure in an omnidirectional-reflector waveguide," Applied Physics Letters 85, 1466-1468 (2004). [71]. A. Mizrahi , and...optical waveguides," Opt. Lett. 30, 3042- 3044 (2005). 59 42 [3]. A. Mizrahi , and L. Schachter, "Mirror manipulation by attractive and...Optical and Plasma Physics 46, 157-164 (2008). [10]. A. Mizrahi , and L. Schachter, "Electromagnetic forces on the dielectric layers of the planar 5. The aero optics effect on near space laser communication optical system Science.gov (United States) Hu, Yuan; Fu, Yuegang; Jiang, Huilin 2013-08-01 With the developing of the space laser communication link, the performance index including higher transfer speed, extending transfer distance, and environmental adaptability, all ask the system accuracy and indexes improving. Special the developing near space platform, its environmental is extremes, the near space drone and other airplane flight speed is very quickly from the subsonic to supersonic. The aero optics effect caused by high speed will generate a thin turbulent air layer. It affects the performance of laser communication optical system by laser light vibration, deviation and so on, further more affects the performance of laser communication system working performance, even can't communication. Therefore, for achieving optical system indexes, we need do more research in optical system near space aero optics environmental adaptability. In this paper, near space link environmental characteristic are researched. And on the base of the aero optics theory, computer simulating method is applied to analyze the relationship among the altitude, the flight speed and the image dispersion. The result shows that, the aero optics effect cannot be ignored when the terminal is in low altitude or is moving with supersonic speed. The effect must be taken into considered from overall design. The result will provide the basis of research design. 6. The Muon system of the run II D0 detector Energy Technology Data Exchange (ETDEWEB) Abazov, V.M.; Acharya, B.S.; Alexeev, G.D.; Alkhazov, G.; Anosov, V.A.; Baldin, B.; Banerjee, S.; Bardon, O.; Bartlett, J.F.; Baturitsky, M.A.; Beutel, D.; Bezzubov,; Bodyagin, V.; Butler, J.M.; Cease, H.; Chi, E.; Denisov, D.; Denisov, S.P.; Diehl, H.T.; Doulas, S.; Dugad, S.R.; /Beijing, Inst. High Energy Phys. /Charles U. /Prague, Tech. 2005-03-01 The authors describe the design, construction and performance of the upgraded D0 muon system for Run II of the Fermilab Tevatron collider. Significant improvements have been made to the major subsystems of the D0 muon detector: trigger scintillation counters, tracking detectors, and electronics. The Run II central muon detector has a new scintillation counter system inside the iron toroid and an improved scintillation counter system outside the iron toroid. In the forward region, new scintillation counter and tracking systems have been installed. Extensive shielding has been added in the forward region. A large fraction of the muon system electronics is also new. 7. Optical design of an in vivo laparoscopic lighting system. Science.gov (United States) Liu, Xiaolong; Abdolmalaki, Reza Yazdanpanah; Mancini, Gregory J; Tan, Jindong 2017-12-01 This paper proposes an in vivo laparoscopic lighting system design to address the illumination issues, namely poor lighting uniformity and low optical efficiency, existing in the state-of-the-art in vivo laparoscopic cameras. The transformable design of the laparoscopic lighting system is capable of carrying purposefully designed freeform optical lenses for achieving lighting performance with high illuminance uniformity and high optical efficiency in a desired target region. To design freeform optical lenses for extended light sources such as LEDs with Lambertian light intensity distributions, we present an effective and complete freeform optical design method. The procedures include (1) ray map computation by numerically solving a standard Monge-Ampere equation; (2) initial freeform optical surface construction by using Snell's law and a lens volume restriction; (3) correction of surface normal vectors due to accumulated errors from the initially constructed surfaces; and (4) feedback modification of the solution to deal with degraded illuminance uniformity caused by the extended sizes of the LEDs. We employed an optical design software package to evaluate the performance of our laparoscopic lighting system design. The simulation results show that our design achieves greater than 95% illuminance uniformity and greater than 89% optical efficiency (considering Fresnel losses) for illuminating the target surgical region. (2017) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). 8. Alternative pathways for angiotensin II generation in the cardiovascular system Directory of Open Access Journals (Sweden) C. Becari 2011-09-01 Full Text Available The classical renin-angiotensin system (RAS consists of enzymes and peptides that regulate blood pressure and electrolyte and fluid homeostasis. Angiotensin II (Ang II is one of the most important and extensively studied components of the RAS. The beneficial effects of angiotensin converting enzyme (ACE inhibitors in the treatment of hypertension and heart failure, among other diseases, are well known. However, it has been reported that patients chronically treated with effective doses of these inhibitors do not show suppression of Ang II formation, suggesting the involvement of pathways alternative to ACE in the generation of Ang II. Moreover, the finding that the concentration of Ang II is preserved in the kidney, heart and lungs of mice with an ACE deletion indicates the important role of alternative pathways under basal conditions to maintain the levels of Ang II. Our group has characterized the serine protease elastase-2 as an alternative pathway for Ang II generation from Ang I in rats. A role for elastase-2 in the cardiovascular system was suggested by studies performed in heart and conductance and resistance vessels of normotensive and spontaneously hypertensive rats. This mini-review will highlight the pharmacological aspects of the RAS, emphasizing the role of elastase-2, an alternative pathway for Ang II generation. 9. An optical fiber interferometric system for non-contact measurement of atmospheric optical turbulence Science.gov (United States) Mei, Haiping; Rao, Ruizhong 2006-02-01 Optical turbulence degrades the quality of laser beam propagation and the quality of the image of optical system, limiting the spatial resolution that can be obtained. A novel single-air-gap fiber optical interferometric system useful for non-contact measurement of the fine structure of optical turbulence is presented. The main idea of this system is based on the application of a specially constructed optical fiber Mach-Zehnder interferometer to measure the phase fluctuations effected by the random fluctuations of refractive index in the turbulent atmosphere. The light source is a long coherence length infrared laser operating at the wavelength of 1.31μm and the optical path exposed to the atmosphere can be adjusted to a most suitable value according to the operational environment. Theoretical estimation illustrates that the system can measure the minimal atmospheric refractive index fluctuation up to 10 -10 during a 2cm propagation path. It is easy to have a calibration of the system and the result shows that the voltage refractive index sensitivity is about 2.1x10 -6 V -1. The system is integrated and well fixed in a burly airproof box with only the sensing arm exposed to the air. It follows that the system is suitable for the measurement of atmospheric turbulence over land and ocean surfaces. 10. The analysis of optical wave beams propagation in lens systems Science.gov (United States) Kazakov, I.; Mosentsov, S.; Moskaletz, O. 2016-08-01 In this paper some aspects of the formation and propagation of optical wave beams in lens systems were considered. As an example, the two-lens optical information processing system was considered. Analysis of the two-lens optical circuit has been made with a systems approach perspective. As part of the radio-optical analogies had been applied certain provisions of the theory of dynamical systems to the spatial optical system. The lens system is represented as a simple series-connected optical elements with known spatial impulse response. General impulse response of such a system has been received, as well as consider some special cases of the impulse response. The question of the relationship between the parameters and the size of the input aperture lenses for undistorted transmission of the optical signal has been considered. Analysis of the energy loss resulting from the finite aperture of the lens. It's based on an assessment of the fraction of radiation that propagates beyond the lens. Analysis showed that the energy losses depend explicitly on the following parameters: radiation wavelength, distance between input aperture and lens, and ratio of the input aperture and lens aperture. With the computer help simulation the dependence of losses was shown on the above parameters 11. ICDAS II CRITERIA (INTERNATIONAL CARIES DETECTION AND ASSESSMENT SYSTEM Directory of Open Access Journals (Sweden) Benin DİKMEN 2015-10-01 Full Text Available The International Caries Detection and Assessment System (ICDAS is a clinical scoring system which allows detection and assessment of caries activity. ICDAS was developed for use in clinical research, clinical practice and for epidemiological purposes. A recent review of caries detection criteria systems found that there were inconsistencies among the research criteria for caries measuring systems. There is a need to an uniform system which allows comparison of data collected in different researches. ICDAS allows detection of caries process at every stage and characterization of the caries activity status of lesion. Later, the criteria were modified and ICDAS II created. The aim of this review is to inform about the ICDAS II and make a comparison between ICDAS II criteria and other caries detection systems. 12. Information Optics and Photonics Algorithms, Systems, and Applications CERN Document Server Javidi, Bahram 2010-01-01 This book addresses applications, recent advances, and emerging areas in fields with applications in information optics and photonics systems. The objective of this book is to illustrate and discuss novel approaches, analytical techniques, models, and technologies that enhance sensing, measurement, processing, interpretation, and visualization of information using free space optics and photonics. The material in this book concentrates on integration of diverse fields for cross-disciplinary applications including bio-photonics, digitally enhanced sensing and imaging systems, multi-dimensional optical imaging and image processing, bio-inspired imaging, 3D visualization, 3D displays, imaging on the nano-scale, quantum optics, super resolution imaging, photonics for biological applications, and holographic information systems. As a result, this book is a useful resource for researchers, engineers, and graduate students who work in the diverse fields comprising information optics and photonics. 13. Ultra-high Frequency Linear Fiber Optic Systems CERN Document Server Lau, Kam 2011-01-01 This book provides an in-depth treatment of both linear fiber-optic systems and their key enabling devices. It presents a concise but rigorous treatment of the theory and practice of analog (linear) fiber-optics links and systems that constitute the foundation of Hybrid Fiber Coax infrastructure in present-day CATV distribution and cable modem Internet access. Emerging applications in remote fiber-optic feed for free-space millimeter wave enterprise campus networks are also described. Issues such as dispersion and interferometric noise are treated quantitatively, and means for mitigating them are explained. This broad but concise text will thus be invaluable not only to students of fiber-optics communication but also to practicing engineers. To the second edition of this book important new aspects of linear fiber-optic transmission technologies are added, such as high level system architectural issues, algorithms for deriving the optimal frequency assignment, directly modulated or externally modulated laser t... 14. The optical system of the proposed Chinese 12-m optical/infrared telescope Science.gov (United States) Su, Ding-qiang; Liang, Ming; Yuan, Xiangyan; Bai, Hua; Cui, Xiangqun 2017-08-01 The lack of a large-aperture optical/infrared telescope has seriously affected the development of astronomy in China. In 2016, the authors published their concept study and suggestions for a 12-m telescope optical system. This article presents the authors' further research and some new results. Considering that this telescope should be a general-purpose telescope for a wide range of scientific goals and could be used for frontier scientific research in the future, the authors studied and designed a variety of 12-m telescope optical systems for comparison and final decision-making. In general, we still adopt our previous configuration, but the Nasmyth and prime-focus corrector systems have been greatly improved. In this article, the adaptive optics is given special attention. Ground-layer adaptive optics (GLAO) is adopted. It has a 14-arcmin field of view. The secondary mirror is used as the adaptive optical deformable mirror. Obviously, not all the optical systems in this telescope configuration will be used or constructed at the same stage. Some will be for the future and some are meant for research rather than for construction. 15. High-stability 48-core bendable and movable optical cable for FAST telescope optical transmission system Science.gov (United States) Liu, Hongfei; Pan, Gaofeng; Lin, Zhong; Liu, Cheng; Zhu, Wenbai; Nan, Rendong; Li, Chunsheng; Gao, Guanjun; Luo, Wenyong; Jin, Chengjin; Song, Jinyou 2017-11-01 The construction of FAST telescope was completed in Guizhou province of China in September 2016, and a kind of novel high-stability 48-core bendable and movable optical cable was developed and applied in analog data optical transmission system of FAST. Novel structure and selective material of this optical cable ensure high stability of optical power in the process of cables round-trip motion when telescope is tracking a radio source. The 105 times bend and stretch accelerated experiment for this optical cable was implemented, and real-time optical and RF signal power fluctuation were measured. The physical structure of optical cables after 105 times round-trip motion is in good condition; the real-time optical power attenuation fluctuation is smaller than 0.044 dB; the real-time RF power fluctuation is smaller than 0.12 dB. The optical cable developed in this letter meets the requirement of FAST and has been applied in FAST telescope. 16. Optical Assembly and Characterization System for Nano-Photonics Research Science.gov (United States) 2016-03-01 Unlimited Final Report: Optical Assembly and Characterization System for Nano -Photonics Research The views, opinions and/or findings contained in this...reviewed journals: Final Report: Optical Assembly and Characterization System for Nano -Photonics Research Report Title With this equipment funding support...Assembly and Characterization System for Nano -Photonics Research PI: Prof. Weidong Zhou, University of Texas at Arlington (UTA) 500 S. Cooper St 17. PEP-II RF System Operation and Performance Energy Technology Data Exchange (ETDEWEB) McIntosh, P. 2005-01-18 The Low Energy Ring (LER) and High Energy Ring (HER) RF systems have operated now on PEP-II since July 1998 and have assisted in breaking all design luminosity records back in June 2001. Luminosity on PEP-II has steadily increased since then as a consequence of larger e+ and e- beam currents being accumulated. This has meant that the RF systems have inevitably been driven harder, not only to achieve these higher stored beam currents, but also to reliably keep the beams circulating whilst at the same time minimizing the number of aborts due to RF system faults. This paper details the current PEP-II RF system configurations for both rings, as well as future upgrade plans spanning the next 3-5 years. Limitations of the current RF system configurations are presented, highlighting improvement projects which will target specific areas within the RF systems to ensure that adequate operating overheads are maintained and reliable operation is assured. The Low Energy Ring (LER) and High Energy Ring (HER) RF systems have operated now on PEP-II since July 1998 and have assisted in breaking all design luminosity records back in June 2001. Luminosity on PEP-II has steadily increased since then as a consequence of larger e+ and e- beam currents being accumulated. This has meant that the RF systems have inevitably been driven harder, not only to achieve these higher stored beam currents, but also to reliably keep the beams circulating whilst at the same time minimizing the number of aborts due to RF system faults. This paper details the current PEP-II RF system configurations for both rings, as well as future upgrade plans spanning the next 3-5 years. Limitations of the current RF system configurations are presented, highlighting improvement projects which will target specific areas within the RF systems to ensure that adequate operating overheads are maintained and reliable operation is assured. 18. Variable configuration fiber optic laser doppler vibrometer system Science.gov (United States) Posada-Roman, Julio E.; Jackson, David A.; Garcia-Souto, Jose A. 2016-06-01 A multichannel heterodyne fiber optic vibrometer is demonstrated which can be operated at ranges in excess of 50 m. The system is designed to measure periodic signals, impacts, rotation, 3D strain, and vibration mapping. The displacement resolution of each channel exceeds 1 nm. The outputs from all channels are simultaneous, and the number of channels can be increased by using optical switches. 19. MERTIS: geometrical calibration of thermal infrared optical system by applying diffractive optical elements Science.gov (United States) Bauer, M.; Baumbach, D.; Buder, M.; Börner, A.; Grießbach, D.; Peter, G.; Santier, E.; Säuberlich, T.; Schischmanow, A.; Schrader, S.; Walter, I. 2015-09-01 Geometrical sensor calibration is essential for space applications based on high accuracy optical measurements, in this case for the thermal infrared push-broom imaging spectrometer MERTIS. The goal is the determination of the interior sensor orientation. A conventional method is to measure the line of sight for a subset of pixels by single pixel illumination with collimated light. To adjust angles, which define the line of sight of a pixel, a manipulator construction is used. A new method for geometrical sensor calibration is using Diffractive Optical Elements (DOE) in connection with laser beam equipment. Diffractive optical elements (DOE) are optical microstructures, which are used to split an incoming laser beam with a dedicated wavelength into a number of beams with well-known propagation directions. As the virtual sources of the diffracted beams are points at infinity, the resulting image is invariant against translation. This particular characteristic allows a complete geometrical sensor calibration with only one taken image avoiding complex adjustment procedures, resulting in a significant reduction of calibration effort. We present a new method for geometrical calibration of a thermal infrared optical system, including an thermal infrared test optics and the MERTIS spectrometer bolometer detector. The fundamentals of this new approach for geometrical infrared optical systems calibration by applying diffractive optical elements and the test equipment are shown. 20. An optical CDMA system based on chaotic sequences Science.gov (United States) Liu, Xiao-lei; En, De; Wang, Li-guo 2014-03-01 In this paper, a coherent asynchronous optical code division multiple access (OCDMA) system is proposed, whose encoder/decoder is an all-optical generator. This all-optical generator can generate analog and bipolar chaotic sequences satisfying the logistic maps. The formula of bit error rate (BER) is derived, and the relationship of BER and the number of simultaneous transmissions is analyzed. Due to the good property of correlation, this coherent OCDMA system based on these bipolar chaotic sequences can support a large number of simultaneous users, which shows that these chaotic sequences are suitable for asynchronous OCDMA system. 1. Optical fiber sensors measurement system and special fibers improvement Science.gov (United States) Jelinek, Michal; Hrabina, Jan; Hola, Miroslava; Hucl, Vaclav; Cizek, Martin; Rerucha, Simon; Lazar, Josef; Mikel, Bretislav 2017-06-01 We present method for the improvement of the measurement accuracy in the optical frequency spectra measurements based on tunable optical filters. The optical filter was used during the design and realization of the measurement system for the inspection of the fiber Bragg gratings. The system incorporates a reference block for the compensation of environmental influences, an interferometric verification subsystem and a PC - based control software implemented in LabView. The preliminary experimental verification of the measurement principle and the measurement system functionality were carried out on a testing rig with a specially prepared concrete console in the UJV Řež. The presented system is the laboratory version of the special nuclear power plant containment shape deformation measurement system which was installed in the power plant Temelin during last year. On the base of this research we started with preparation other optical fiber sensors to nuclear power plants measurement. These sensors will be based on the microstructured and polarization maintaining optical fibers. We started with development of new methods and techniques of the splicing and shaping optical fibers. We are able to made optical tapers from ultra-short called adiabatic with length around 400 um up to long tapers with length up to 6 millimeters. We developed new techniques of splicing standard Single Mode (SM) and Multimode (MM) optical fibers and splicing of optical fibers with different diameters in the wavelength range from 532 to 1550 nm. Together with development these techniques we prepared other techniques to splicing and shaping special optical fibers like as Polarization-Maintaining (PM) or hollow core Photonic Crystal Fiber (PCF) and theirs cross splicing methods with focus to minimalize backreflection and attenuation. The splicing special optical fibers especially PCF fibers with standard telecommunication and other SM fibers can be done by our developed techniques. Adjustment 2. Optical mesosensors for monitoring and removal of ultra-trace concentration of Zn(II) and Cu(II) ions from water. Science.gov (United States) El-Safty, Sherif A; Shenashen, M A; Ismael, M; Khairy, M; Awual, Md R 2012-11-21 Optical captor design is necessary for the controlled development of a technique for detecting and removing heavy and toxic metals from drinking water with high flexibility and low capital cost. We designed chemical mesocaptors for optical separation/extraction and monitoring/detection of Cu(II) and Zn(II) ions from water even at trace concentration levels without a preconcentration process. The mesoporous aluminosilica carriers with three-dimensional (3D) structures, high aluminum content, natural surfaces, and active acid sites strongly induced H-bonding and dispersive interactions with organic moieties, thereby leading to the formation of stable captors without chromophore leaching during the removal assays of Cu(II) and Zn(II) ions. Using such a tailored mesocaptor design, the direct immobilization of these hydrophobic ligands (4,5-diamino-6-hydroxy-2-mercaptopyrimidine and diphenylthiocarbazone) into ordered pore-based aluminasilica monoliths enabled the easy generation and transduction of optical colour signals as a response to metal-to-ligand binding events, even at ultra-trace concentrations (~10(-9) mol dm(-3)) of Cu(II) and Zn(II) ions in drinking water, without the need for sophisticated instruments. Theoretical models have been developed to provide insights into the effect of active site surfaces on the enhancement of the optical removal process in terms of long-term stability, reversibility, and selectivity, hence allowing us to understand the role of mesoscopic geometry and nanoscale pore orientation of mesocaptors better. Generally, this ion-capture model enables the development of a simple and effective technique for effective wastewater treatment and management. 3. Optical beam transport system at FEL-SUT CERN Document Server Nomaru, K; Yokoyama, M; Oda, F; Nakayama, A; Koike, H; Kuroda, H 2000-01-01 Kawasaki Heavy Industries Ltd. has installed an FEL beam transport system at the IR FEL Research Center of the Science University of Tokyo (FEL-SUT). This system transports the FEL output beam from the FEL machine room to the optical diagnostic room through a vacuum tube. The in-vacuum multi-mirror synchronized system operated from the FEL control room enables the operator to control the multiple mirrors simultaneously on or off axis of the FEL beam and to distribute the FEL output to one of the laboratories. The essential component of the transport system is the passive control optics that is composed of an elliptical and parabolic mirror couple. Once the control optics is aligned, a parallel FEL beam with a good pointing stability is obtained without any active operation to tune the optical system for different wavelengths. 4. Hillside Administration Reporting Timetabling System. HARTS II. Science.gov (United States) Goddard, William P. This second version of a computerized school records system utilizing punch cards was developed by high school students under faculty guidance. Not only is it inexpensive to use, but it is tailored to the needs of a particular school and, therefore, responsive to its needs. Detailed step by step instructions are provided for utilizing the system.… 5. The Bacillus cereus spoIIS programmed cell death system Directory of Open Access Journals (Sweden) Jana eMelnicakova 2015-08-01 Full Text Available Programmed cell death in bacteria is generally associated with two¬ component toxin antitoxin systems. The SpoIIS toxin-antitoxin system, consisting of a membrane bound SpoIISA toxin and a small, cytosolic antitoxin SpoIISB, was originally identified in Bacillus subtilis. In this work we describe the Bacillus cereus SpoIIS system which is a three-component system, harbouring an additional gene spoIISC. Its protein product serves as an antitoxin, and similarly as SpoIISB, is able to bind SpoIISA and abolish its toxic effect. Our results indicate that SpoIISC seems to be present not only in B. cereus but also in other Bacilli containing a SpoIIS toxin antitoxin system. In addition, we show that B. cereus SpoIISA can form higher oligomers and we discuss the possible role of this multimerization for the protein’s toxic function. 6. Optical response of correlated electron systems Science.gov (United States) Maslov, Dmitrii L.; Chubukov, Andrey V. 2017-02-01 Recent progress in experimental techniques has made it possible to extract detailed information on dynamics of carriers in a correlated electron material from its optical conductivity, σ (Ω,T) . This review consists of three parts, addressing the following three aspects of optical response: (1) the role of momentum relaxation; (2) Ω /T scaling of the optical conductivity of a Fermi-liquid metal, and (3) the optical conductivity of non-Fermi-liquid metals. In the first part (section 2), we analyze the interplay between the contributions to the conductivity from normal and umklapp electron-electron scattering. As a concrete example, we consider a two-band metal and show that although its optical conductivity is finite it does not obey the Drude formula. In the second part (sections 3 and 4), we re-visit the Gurzhi formula for the optical scattering rate, 1/τ (Ω,T)\\propto {{ Ω }2}+4{π2}{{T}2} , and show that a factor of 4{π2} is the manifestation of the ‘first-Matsubara-frequency rule’ for boson response, which states that 1/τ (Ω,T) must vanish upon analytic continuation to the first boson Matsubara frequency. However, recent experiments show that the coefficient b in the Gurzhi-like form, 1/τ (Ω,T)\\propto {{ Ω }2}+b{π2}{{T}2} , differs significantly from b  =  4 in most of the cases. We suggest that the deviations from Gurzhi scaling may be due to the presence of elastic but energy-dependent scattering, which decreases the value of b below 4, with b  =  1 corresponding to purely elastic scattering. In the third part (section 5), we consider the optical conductivity of metals near quantum phase transitions to nematic and spin-density-wave states. In the last case, we focus on ‘composite’ scattering processes, which give rise to a non-Fermi-liquid behavior of the optical conductivity at T  =  0: {σ\\prime}(Ω )\\propto {{ Ω }-1/3} at low frequencies and {σ\\prime}(Ω )\\propto {{ Ω }-1} at higher frequencies. We also 7. The Thorvald II Agricultural Robotic System Directory of Open Access Journals (Sweden) 2017-09-01 Full Text Available This paper presents a novel and modular approach to agricultural robots. Food production is highly diverse in several aspects. Even farms that grow the same crops may differ in topology, infrastructure, production method, and so on. Modular robots help us adapt to this diversity, as they can quickly be configured for various farm environments. The robots presented in this paper are hardware modular in the sense that they can be reconfigured to obtain the necessary physical properties to operate in different production systems—such as tunnels, greenhouses and open fields—and their mechanical properties can be adapted to adjust for track width, power requirements, ground clearance, load capacity, and so on. The robot’s software is generalizing to work with the great variation of robot designs that can be realized by assembling hardware modules in different configurations. The paper presents several novel ideas for agricultural robotics, as well as extensive field trials of several different versions of the Thorvald II platform. 8. Optical detection/collection of toxic Cd(II) ions using cubic Ia3d aluminosilica mesocage sensors. Science.gov (United States) El-Safty, Sherif A; Shenashen, Mohamed A; Khairy, Mohamed 2012-08-30 Optical sensors for selective removal and detection of extremely toxic ions such as cadmium (Cd(II)) in aquatic samples were successfully fabricated via simple strategy. Aluminosilica-based network platforms are used as selective mesopore shape and size carriers in order to fabricate optical sensors through the direct functionalization of α, β, γ, and δ-tetrakis(1-methylpyridinium-4-yl)porphine ρ-toluenesulfonate (TMPyP) moieties without any prior surface modification using silane or thiol agents. In turn, the key advantage of a heretical three-dimensional (3D) cubic Ia3d mesocage is the facile access of target ions such as ion transports and the high affinity responses of TMPyP receptor-Cd(II) analyte binding events, which result in the easy generation and transduction of optical signals even at the trace level of the Cd(II) ion. The optical sensor design-based aluminosilica cages enable the sensitive detection and selective removal of Cd(II) ions even at ultra-trace concentrations of 10(-10)mol/dm(3) with rapid response time (in minutes). This rational strategy is crucial to the development of optical mesocollectors (i.e., probe surface-mounted naked-eye ion-sensor strips) with highly selective Cd(II) ion removal from aqueous water. These new classes of optical mesocollectors exhibit long-term stability and reusability of deleterious Cd(II) ions, which makes them efficient for various analytical applications. Copyright © 2012 Elsevier B.V. All rights reserved. 9. Electronic construction collaboration system -- phase II. Science.gov (United States) 2010-06-01 During the first year of research, work was completed to identify Iowa DOT needs for web-based project management system (WPMS) : and evaluate how commercially available solutions could meet these needs. Researchers also worked to pilot test custom d... 10. Novel optical probe for quantum Hall system Light is chopped using a MEMS optical switch triggered with lock-in amplifier. 3. Results and discussion. The B-SPV and E-SPV spectra measured at zero magnetic field are compared in figure 3. The edge spectrum rises abruptly at an energy characteristic of the band edge of the QW with no excitonic-like feature and is flat ... 11. Rheo-optics and food systems NARCIS (Netherlands) Linden, van der E.; Sagis, L.M.C.; Venema, P. 2003-01-01 Recent developments in two- and three-dimensional rheo-optics in foods and non-foods are being reviewed. These techniques are shown to be very valuable to construct constitutive equations that describe the non-linear rheological behaviour in 2 and 3 dimensions and couple this behaviour to the 12. Porphyrin-functionalized porous polysulfone membrane towards an optical sensor membrane for sorption and detection of cadmium(II) Energy Technology Data Exchange (ETDEWEB) Zhao, Lizhi, E-mail: zhaolizhi_phd@163.com [State Key Laboratory of Separation Membranes and Membrane Processes, School of Materials Science and Engineering, Tianjin Polytechnic University, Tianjin 300387 (China); Li, Min; Liu, Manman; Zhang, Yuecong [State Key Laboratory of Separation Membranes and Membrane Processes, School of Materials Science and Engineering, Tianjin Polytechnic University, Tianjin 300387 (China); Wu, Chenglin [School of Pharmaceutical and Chemical Engineering, Taizhou University, Taizhou 317000, Zhejiang Province (China); Zhang, Yuzhong, E-mail: zhangyz2004cn@163.com [State Key Laboratory of Separation Membranes and Membrane Processes, School of Materials Science and Engineering, Tianjin Polytechnic University, Tianjin 300387 (China) 2016-01-15 Highlights: • An optical sensor membrane is prepared by TMPyP and PNaSS-grafted PSF membrane. • The optical sensor membrane shows enhanced sorption for cadmium(II). • Visual and spectrophotometric detection can be achieved. • The functional membrane exhibits good stability and reusability. - Abstract: In this study, an optical sensor membrane was prepared for sorption and detection of cadmium(II) (Cd(II)) in aqueous solution. A polyanion, poly(sodium 4-styrenesulfonate) (PNaSS), was grafted onto the chloromethylated polysulfone (CMPSF) microporous membrane via surface-initiated ATRP. 5,10,15,20-tetrakis(4-N-methylpyridyl) porphyrin p-toluenesulfonate (TMPyP) was immobilized onto the PNaSS-grafted polysulfone (PSF-PNaSS) membrane through electrostatic interaction. The TMPyP-functionalized membrane exhibited an enhanced sorption for, and distinct color and spectral response to cadmium(II) (Cd(II)) in aqueous solution. Larger immobilization capacity of TMPyP on the membrane led to stronger sorption for Cd(II), and smaller one made the optical sensor have a faster (in minutes) and more sensitive response to the ion. The detection limit study indicated that the functional membrane with proper amount of TMPyP (<0.5 mg/g) could still have color and spectral response to Cd(II) solutions at an extreme low concentration (10{sup −4} mg/L). The optical sensor membrane exhibited good stability and reusability which made it efficient for various sorptive removal and detection applications. 13. Optical Fiber Embedded in Epoxy Glass Unidirectional Fiber Composite System. Science.gov (United States) Severin, Irina; El Abdi, Rochdi; Corvec, Guillaume; Caramihai, Mihai 2013-12-20 We aimed to embed silica optical fibers in composites (epoxy vinyl ester matrix reinforced with E-glass unidirectional fibers in mass fraction of 60%) in order to further monitor the robustness of civil engineering structures (such as bridges). A simple system was implemented using two different silica optical fibers (F1-double coating of 172 µm diameter and F2-single coating of 101.8 µm diameter respectively). The optical fibers were dynamically tensile tested and Weibull plots were traced. Interfacial adhesion stress was determined using pull-out test and stress values were correlated to fracture mechanisms based on SEM observations. In the case of the optical fiber (OF) (F1)/resin system and OF (F1)/composite system, poor adhesion was reported that may be correlated to interface fracture at silica core level. Relevant applicable results were determined for OF (F2)/composite system. 14. Automated Mulitple Object Optical Tracking and Recognition System Project Data.gov (United States) National Aeronautics and Space Administration — OPTRA proposes to develop an optical tracking system that is capable of recognizing and tracking up to 50 different objects within an approximately 2 degree x 3... 15. FTTA System Demo Using Optical Fiber-Coupled Active Antennas Directory of Open Access Journals (Sweden) Niels Neumann 2014-08-01 Full Text Available The convergence of optical and wireless systems such as Radio-over-Fiber (RoF networks is the key to coping with the increasing bandwidth demands due to the increasing popularity of video and other high data rate applications. A high level of integration of optical technologies enables simple base stations with a fiber-to-the-antenna (FTTA approach. In this paper, we present a complete full-duplex RoF–FTTA system consisting of integrated active fiber-coupled optical receiving and transmitting antennas that are directly connected to a standard single mode fiber optical link. Data rates up to 1 Gbit/s could be shown without advanced modulation formats on a 1.5 GHz carrier frequency. The antennas as well as the whole system are explained and the results of the system experiments are discussed. 16. Optical Imaging Sensors and Systems for Homeland Security Applications CERN Document Server Javidi, Bahram 2006-01-01 Optical and photonic systems and devices have significant potential for homeland security. Optical Imaging Sensors and Systems for Homeland Security Applications presents original and significant technical contributions from leaders of industry, government, and academia in the field of optical and photonic sensors, systems and devices for detection, identification, prevention, sensing, security, verification and anti-counterfeiting. The chapters have recent and technically significant results, ample illustrations, figures, and key references. This book is intended for engineers and scientists in the relevant fields, graduate students, industry managers, university professors, government managers, and policy makers. Advanced Sciences and Technologies for Security Applications focuses on research monographs in the areas of -Recognition and identification (including optical imaging, biometrics, authentication, verification, and smart surveillance systems) -Biological and chemical threat detection (including bios... 17. Student project of optical system analysis API-library development Science.gov (United States) Ivanova, Tatiana; Zhukova, Tatiana; Dantcaranov, Ruslan; Romanova, Maria; Zhadin, Alexander; Ivanov, Vyacheslav; Kalinkina, Olga 2017-08-01 In the paper API-library software developed by students of Applied and Computer Optics Department (ITMO University) for optical system design is presented. The library performs paraxial and real ray tracing, calculates 3d order (Seidel) aberration and real ray aberration of axis and non-axis beams (wave, lateral, longitudinal, coma, distortion etc.) and finally, approximate wave aberration by Zernike polynomials. Real aperture can be calculated by considering of real rays tracing failure on each surface. So far we assume optical system is centered, with spherical or 2d order aspherical surfaces. Optical glasses can be set directly by refraction index or by dispersion coefficients. The library can be used for education or research purposes in optical system design area. It provides ready to use software functions for optical system simulation and analysis that developer can simply plug into their software development for different purposes, for example for some specific synthesis tasks or investigation of new optimization modes. In the paper we present an example of using the library for development of cemented doublet synthesis software based on Slusarev's methodology. The library is used in optical system optimization recipes course for deep studying of optimization model and its application for optical system design. Development of such software is an excellent experience for students and help to understanding optical image modeling and quality analysis. This development is organized as student group joint project. We try to organize it as a group in real research and development project, so each student has his own role in the project and then use whole library functionality in his own master or bachelor thesis. Working in such group gives students useful experience and opportunity to work as research and development engineer of scientific software in the future. 18. Design of a multimodal fibers optic system for small animal optical imaging. Science.gov (United States) Spinelli, Antonello E; Pagliazzi, Marco; Boschi, Federico 2015-02-01 Small animals optical imaging systems are widely used in pre-clinical research to image in vivo the bio-distribution of light emitting probes using fluorescence or bioluminescence modalities. In this work we presented a set of simulated results of a novel small animal optical imaging module based on a fibers optics matrix, coupled with a position sensitive detector, devoted to acquire bioluminescence and Cerenkov images. Simulations were performed using GEANT 4 code with the GAMOS architecture using the tissue optics plugin. Results showed that it is possible to image a 30 × 30 mm region of interest using a fiber optics array containing 100 optical fibers without compromising the quality of the reconstruction. The number of fibers necessary to cover an adequate portion of a small animal is thus quite modest. This design allows integrating the module with magnetic resonance (MR) in order to acquire optical and MR images at the same time. A detailed model of the mouse anatomy, obtained by segmentation of 3D MRI images, will improve the quality of optical 3D reconstruction. Copyright © 2014 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved. 19. Optical properties of II-VI semiconductor nanoclusters for use as phosphors Science.gov (United States) Wilcoxon, Jess P.; Newcomer, Paula 2002-11-01 The optical properties of both II-VI (direct gap) and type IV (indirect gap) nanosize semiconductors are significantly affected not only by their size, but by the nature of the chemical interface of the cluster with the embedding medium. This affects the light conversion efficiency and can alter the shape and position (i.e. the color) of the photoluminescence (PL). As the goal of our work is to embed nanoclusters into either organic or inorganic matrices for use as near UV, LED-excited phosphor thin films, understanding and controlling this interface is very important for preserving the high Q.E. of nanoclusters known for dilute solution conditions. We describe a room temperature synthesis of semiconductor nanoclusters which employs inexpensive, less toxic ionic precursors (metal salts), and simple coordinating solvents (e.g. tetrahydrofuran). This allows us to add passivating agents, ions, metal or semiconductor coatings to identical, highly dispersed bare clusters, post-synthesis. We can also increase the cluster size by heterogeneous growth on the seed nanoclusters. One of the most interesting observations for our II-VI nanomaterials is that both the absorbance excitonic features and the photoluminescence (PL) energy and intensity depend on the nature of the surface as well as the average size. In CdS, for example, the presence of electron traps (i.e Cd(II) sites) decreases the exciton absorbance peak amplitude but increases the PL nearly two-fold. Hole traps (i.e. S(II)) have the opposite effect. In the coordinating solvents used for the synthesis, the PL yield for d~2 nm, blue emitting CdSe clusters increases dramatically with sample age as the multiple absorbance features sharpen. Liquid chromatographic (LC) separation of the nanoclusters from other chemicals and different sized clusters is used to investigate the intrinsic optical properties of the purified clusters and identify which clusters are contributing most strongly to the PL. Both LC and dynamic 20. Optical Power Transfer System for Powering a Remote Mobility System for Multiple Missions Science.gov (United States) Stone, William C. (Inventor); Hogan, Bartholomew P. (Inventor) 2016-01-01 An optical power transfer system for powering a remote mobility system for multiple missions comprising a high power source and a chilling station connected to a laser source. The laser source transmits a high optical energy to a beam switch assembly via an optical fiber. The beam switch assembly is optically connected to actively cooled fiber spoolers. Docking stations are adapted for securing the fiber spoolers until alternatively ready for use by a remote mobility system. The remote mobility system is optically connected to the fiber spoolers and has a receiving port adapted for securing the fiber spoolers thereon. The fiber spooler transmits the optical energy to a power conversion system which converts the optical energy received to another usable form of energy. More than one power source may be used where the remote mobility system transfers from one source to another while maintaining an operational radius to each source. 1. Mode-Locked Semiconductor Lasers for Optical Communication Systems DEFF Research Database (Denmark) Yvind, Kresten; Larsson, David; Oxenløwe, Leif Katsuo 2005-01-01 We present investigations on 10 and 40 GHz monolithic mode-locked lasers for applications in optical communications systems. New all-active lasers with one to three quantum wells have been designed, fabricated and characterized.......We present investigations on 10 and 40 GHz monolithic mode-locked lasers for applications in optical communications systems. New all-active lasers with one to three quantum wells have been designed, fabricated and characterized.... 2. Laser Transmitters for the optical link systems used in CMS CERN Multimedia Maximilien Brice 2005-01-01 In the CMS experiment of the now new flagship LHC optical links will be used for the tracker readout system. One part of this components will be semiconductor laser (~50.000 !!!), named correctly: 1310 nm InGaAsP (DCPBH-MQW) edge-emitting laser. They are foreseen as transmitter in the Tx Hybrid part of the optical link system. 3. Optical production systems using neural networks and symbolic substitution Science.gov (United States) Botha, Elizabeth; Casasent, David; Barnard, Etienne 1988-01-01 Two optical implementations of production systems are advanced. The production systems operate on a knowledge base where facts and rules are encoded as formulas in propositional calculus. The first implementation is a binary neural network. An analog neural network is used to include reasoning with uncertainties. The second implementation uses a new optical symbolic substitution correlator. This implementation is useful when a set of similar situations has to be handled in parallel on one processor. 4. Spontaneity and Equilibrium II: Multireaction Systems Science.gov (United States) Raff, Lionel M. 2014-01-01 The thermodynamic criteria for spontaneity and equilibrium in multireaction systems are developed and discussed. When N reactions are occurring simultaneously, it is shown that G and A will depend upon N independent reaction coordinates, ?a (a = 1,2, ..., N), in addition to T and p for G or T and V for A. The general criteria for spontaneity and… 5. Assessment of Integrated Information System (IIS) in organization ... African Journals Online (AJOL) The assessment of Integrated Information System (IIS) in organisation is an important initiative to enable the Information System (IS) managers, as well as top management to understand the success status of their investment in IS integration efforts. However, without a proper assessment, an organisation will not know its IIS ... 6. Optical Fiber System For The Optical Monitoring Of Membrane Potentials Of Excitable Tissues. Science.gov (United States) Nassif, G.; Fillette, F.; Kado, R. T. 1988-06-01 Spectrometry of potential-sensitive dyes permits now the monitoring of membrane potential (MP) variations on excitable tissues. It is also possible to monitor qualitatively the electromechanical activity (EMA) on contractile tissues using this same method. We improved two new optical fiber systems for measuring fluorescence. System 1 used two 200 um optical fibers permitting fluorescence excitation of the potential sensitive dye (PSD) (1st Fiber) and the measurement of its fluorescence (2nd Fiber). System 2 used two 200 um optical fibers for a differencial measurement of excitation reflexion and fluorescence through two different optical filters. The two fibers are assembled in a tube to constitute the "Optrode". In System 1, a single photodiode placed behind a 665 nm high pass optical filter (HPOF) is preamplified by a LH-0022 op. amplifier. In system 2, two photodiodes are placed respectively behind a 645 nm and a 665 nm HPOFs and differencialy preampli-fied by a AD-521 instrumentation amplifier. Analogic signals are lowpass filtered with a LT-1062 digital filter. PSDs WW781 and RH 237 were used respectively on myocardial and nervous tissues. Fluorescence excitation of PSD WW 781 was performed with a 5 mWatts Helium-Neon LASER (He-Ne) beam focused into the first opti-cal fiber of System 1 or illuminating directly the stained tissue. Fluorescence excitation of PSD RH 237 was performed with a 100 Watts filament lamp through a 400-500 nm blue bandpass filter (BBPF). These two systems are now used to study MP variations on myocardial and nervous tissues. We are now able to study the effects of different drugs on the EMA of heart muscle. This technique is usable "in Vitro" and "in Vivo". This optical fiber method easy to improve for a low cost permits now to perform studies on excitable tissues in a non traumatic way with foreseeable applications to pharmacological investigations on experimental models. 7. Microscopic Modeling of Intersubband Optical Processes in Type II Semiconductor Quantum Wells: Linear Absorption Science.gov (United States) Li, Jian-Zhong; Kolokolov, Kanstantin I.; Ning, Cun-Zheng 2003-01-01 Linear absorption spectra arising from intersubband transitions in semiconductor quantum well heterostructures are analyzed using quantum kinetic theory by treating correlations to the first order within Hartree-Fock approximation. The resulting intersubband semiconductor Bloch equations take into account extrinsic dephasing contributions, carrier-longitudinal optical phonon interaction and carrier-interface roughness interaction which is considered with Ando s theory. As input for resonance lineshape calculation, a spurious-states-free 8-band kp Hamiltonian is used, in conjunction with the envelop function approximation, to compute self-consistently the energy subband structure of electrons in type II InAs/AlSb single quantum well structures. We demonstrate the interplay of nonparabolicity and many-body effects in the mid-infrared frequency range for such heterostructures. 8. Lighting the way: photonics leaders II (PL2) optics and photonics teacher professional development Science.gov (United States) Gilchrist, Pamela O.; Hilliard-Clark, Joyce; Bowles, Tuere; Carpenter, Eric 2014-07-01 A sample group of nineteen teachers completed the second phase of the Photonics Leaders II Optics and Photonics professional development program. Participants took a basic Physics content knowledge test that was designed by a Professor of Physics. The test was completed before the teachers participated in the program and at the end of the program to gather data for statistical inquiry. Statistical studies on pre-test and post-test data indicated significant gains in physics content knowledge over time, and that instructors teaching at the middle school level or only teaching one subject area scored significantly lower during the pretest. Reports from previous participants are summarized to disseminate the percentage of teachers who have incorporated at least one workshop activity and the kind of activity performed. The concerns and limitations reported by previous participants are reviewed as well. 9. Combined Electrical, Optical and Nuclear Investigations of Impurities and Defects in II-VI Semiconductors CERN Multimedia 2002-01-01 % IS325 \\\\ \\\\ To achieve well controlled bipolar conductivity in II-VI semiconductors represents a fundamental problem in semiconductor physics. The doping problems are controversely discussed, either in terms of self compensation or of compensation and passivation by unintentionally introduced impurities. \\\\ \\\\It is the goal of our experiments at the new ISOLDE facility, to shed new light on these problems and to look for ways to circumvent it. For this aim the investigation of impurities and native defects and the interaction between each other shall be investigated. The use of radioactive ion beams opens the access to controlled site selective doping of only one sublattice via nuclear transmutation. The compensating and passivating mechanisms will be studied by combining nuclear, electrical and optical methods like Perturbed Angular Correlation~(PAC), Hall Effect~(HE), Deep Level Transient Spectroscopy~(DLTS), Photoluminescence Spectroscopy~(PL) and electron paramagnetic resonance (EPR). \\\\ \\\\We intend to ... 10. Fiber optic signal collection system for primary flight control applications Science.gov (United States) Harper, Sandy L. 1994-10-01 The FOPMN is a fiber-optic signal collection system for primary flight control applications. An avionics bay protected electro-optic interface unit transmits light down fiber optic cable to an optical sensor housed in the harsh environment of a hydraulic actuator. The interface unit also receives the sensor's reflected pattern and calculates independent positions from the multiplexed signals. This paper discusses the FOPMN method for fiber-optically sensing and multiplexing two channels of position of a TEF actuator's main ram cylinder. Currently installed in NASA Dryden's SRA F/A-18, the FOPMN has accumulated approximately 15 hours of flight time. A performance comparison is made between the FOPMN positions and the flight control computer's feedback mechanism (the actuator LVDTs). Included is a discussion of some of the lessons learned as a result of testing the FOPMN in the lab and in flight. The FOPMN is well on its way to proving itself as a robust fiber optic system with the ability to multiplex numerous optical sensors for primary flight control. The success of the FOPMN leads to the second phase of the project--optical loop closure. Our goal for this phase is to have four FOPMN sensor channels on the main ram and/or the main control valve of the actuator to serve as the quad redundant feedback mechanism for flight control. 11. A new approach for the verification of optical systems Science.gov (United States) Siddique, Umair; Aravantinos, Vincent; Tahar, Sofiène 2013-09-01 Optical systems are increasingly used in microsystems, telecommunication, aerospace and laser industry. Due to the complexity and sensitivity of optical systems, their verification poses many challenges to engineers. Tra­ditionally, the analysis of such systems has been carried out by paper-and-pencil based proofs and numerical computations. However, these techniques cannot provide perfectly accurate results due to the risk of human error and inherent approximations of numerical algorithms. In order to overcome these limitations, we propose to use theorem proving (i.e., a computer-based technique that allows to express mathematical expressions and reason about them by taking into account all the details of mathematical reasoning) as an alternative to computational and numerical approaches to improve optical system analysis in a comprehensive framework. In particular, this paper provides a higher-order logic (a language used to express mathematical theories) formalization of ray optics in the HOL Light theorem prover. Based on the multivariate analysis library of HOL Light, we formalize the notion of light ray and optical system (by defining medium interfaces, mirrors, lenses, etc.), i.e., we express these notions mathematically in the software. This allows us to derive general theorems about the behavior of light in such optical systems. In order to demonstrate the practical effectiveness, we present the stability analysis of a Fabry-Perot resonator. 12. Measurements of Intensive Aerosol Optical Properties During TexAQS II Science.gov (United States) Atkinson, D. B.; Radney, J. G.; Wright, M. E. 2007-12-01 Time-resolved measurements of the bulk extensive aerosol optical properties - particle extinction coefficient (bext) and particle scattering coefficient (bscat) - and particle number concentrations were made as part of the six-week TRAMP experiment during the TexAQS II (2006) study. These measurements were done at a nominal surface site (the roof of an 18 story building) on the University of Houston campus near downtown Houston, Texas. Our ground-based tandem cavity ring-down transmissometer/nephelometer instrument (CRDT/N) provided the aerosol optical property measurements. A commercial Condensation Particle Counter (TSI 3007) was used to measure the number concentrations during part of the study period. The optical data was used to construct the intensive aerosol optical properties single scattering albedo ω0 at 532 nm and the Angstrom exponent for extinction between 532 nm and 1064 nm. Recent validation studies of size- selected laboratory generated aerosols are presented to illustrate the soundness of this approach using our instrument. The Angstrom exponent is compared to values from other instruments operating in the area and is found to be a characteristic of the regional air mass under some conditions. Size distributions measured during the study were used to create a new empirical adjustment to scattering measured by the Radiance Research nephelometer, resulting in improved results for particle absorption coefficient and single scattering albedo. The study average value of ω0(532 nm) = 0.78 is lower than expected from comparable field studies and even lower values are experienced during the study. Possible causes of this discrepancy are examined and the utility of using the current version of the CRDT/N instrument to measure the key radiative property ω0 is assessed. Observed episodes of rapid increases in particle number concentration with little corresponding growth in the optical properties can presumably be used to signal the occurrence of particle 13. Design of apochromatic telescopic optical system based on PWC method Science.gov (United States) Wei, Xiao-Xiao; Xu, Feng; Yu, Jian-jun 2011-11-01 According to technical requirements,an apochromatic telescopic optical system which is characterized by the focal length 1.5m~2m,with the visible region spectrum band is needed.The successful design consists in using normal glasses based on PWC method to get a completely secondary-spectrum-removed and cramped construction telescopic optical system in this paper.Firstly, by the aberration theory and ZEMAX optical design software,a catadioptric achromatic optical system composed of a reflecting mirror and a couple of air-spaced negative doublets are designed.The initial configuration of optical system is calculated by PWC method and the proper couple of air-spaced negative doublets are chosen by the theory of achromatic.By comparing with various optical systems with different couple of air-spaced negative doublets, it is found that the combination of the heavy crown glass ZK7 and heavy flint glass ZF3 is effective to correct chromatic aberration and its secondary spectrum value is minimality.It can conclude that fact with the abbe number of ZK7 is twice than ZF3 make the good result and give guiding to the design of the same type of optical system. Secondary,according to the theory of secondary spectrum,the secondary spectrum value to be corrected is analyzed and the compensation can be changed by adjust the magnification power.The comparison is presented and the final telescopic optical system with cramped construction is achieved, and the goal of apochromatism and other aberration is reached. 14. On split Lie triple systems II In the present paper we extend these results to arbitrary split Lie triple systems with no restrictions on their 0-root spaces. Author Affiliations. Antonio J Calderón Martín1 M Forero Piulestán1. Departamento de Matemáticas, Universidad de Cádiz, 11510 Puerto Real, Cádiz, Spain. Dates. Manuscript received: 24 June 2009 ... 15. NIAC Phase II Orbiting Rainbows: Future Space Imaging with Granular Systems Science.gov (United States) Quadrelli, Marco B.; Basinger, Scott; Arumugam, Darmindra; Swartzlander, Grover 2017-01-01 allow for unprecedented high resolution to discern continents and important features of other planets, hyperspectral imaging, adaptive systems, spectroscopy imaging through limb, and stable optical systems from Lagrange-points. Furthermore, future micro-miniaturization might hold promise of a further extension of our dust aperture concept to other more exciting smart dust concepts with other associated capabilities. Our objective in Phase II was to experimentally and numerically investigate how to optically manipulate and maintain the shape of an orbiting cloud of dust-like matter so that it can function as an adaptable ultra-lightweight surface. Our solution is based on the aperture being an engineered granular medium, instead of a conventional monolithic aperture. This allows building of apertures at a reduced cost, enables extremely fault-tolerant apertures that cannot otherwise be made, and directly enables classes of missions for exoplanet detection based on Fourier spectroscopy with tight angular resolution and innovative radar systems for remote sensing. In this task, we have examined the advanced feasibility of a crosscutting concept that contributes new technological approaches for space imaging systems, autonomous systems, and space applications of optical manipulation. The proposed investigation has matured the concept that we started in Phase I to TRL 3, identifying technology gaps and candidate system architectures for the space-borne cloud as an aperture. 16. Desing of a Laser Guide Star System for the Keck II Telescope Energy Technology Data Exchange (ETDEWEB) Friedman, H.W.; Erbert, G.V.; Kuklo, T.; Thompson, G.R.; Wong, N.J.; Gavel, D.T.; Salmon, J.T.; Feldman, M. 1997-09-11 A laser guide star system similar to that deployed at the Lick Observatory has been designed for the Keck II 10 m telescope on Mauna Kea, Hawaii. The subaperature size on the primary is comparable to that at Lick, and at the same observational wavelength centered about the K band, so that the average power requirements of the laser system are also comparable, at about 20 W. One major difference is that the seeing at Mauna Kea is about a factor of two better than at Lick so that the spot diameter requirements are smaller and this can give rise to reduced back scatter resulting from saturation effects in the sodium layer. To reduce the peak flux in the sodium layer and obtain a smaller spot diameter, the output beam diameter has been increased along with the repetition rate of the laser. As with the Lick laser system, a dye laser is pumped by a series of frequency doubled YAG lasers which are remotely located and coupled to the dye laser on the telescope by optical fibers. The laser system has a full set of beam control optics as well as launch telescope and safety systems. A computer system couples the laser system to the User Interface and Supervisory Control system of the main telescope. The laser system is due to be shipped to Keck during the fall of 1997 where it will be integrated with the telescope at Mauna Kea. The Adaptive Optics and Optics Bench systems will be integrated first and be ready for integration with the laser in the summer of 1998. 1 ref., 8 figs. 17. Optical Symbolic Processor for Expert System Execution Science.gov (United States) 1989-12-31 primitive funcions are contained in leaf nodes. A leaf node is a node whose LC contains a value and whose RC is empty. The & node described earlier is used...device was designed using a semi- vectorial finite-difference analysis technique, enabling the range of experimental parameters to be investigated to be...scheme for evaluating propagation constants and mode profiles of optical waveguides. It is semi- vectorial in that solutions to the 1 eigenvalue equation 18. Ultralight Weight Optical Systems Using Nano-Layered Synthesized Materials Science.gov (United States) Clark, Natalie; Breckinridge, James 2014-01-01 Optical imaging is important for many NASA science missions. Even though complex optical systems have advanced, the optics, based on conventional glass and mirrors, require components that are thick, heavy and expensive. As the need for higher performance expands, glass and mirrors are fast approaching the point where they will be too large, heavy and costly for spacecraft, especially small satellite systems. NASA Langley Research Center is developing a wide range of novel nano-layered synthesized materials that enable the development and fabrication of ultralight weight optical device systems that enable many NASA missions to collect science data imagery using small satellites. In addition to significantly reducing weight, the nano-layered synthesized materials offer advantages in performance, size, and cost. 19. Electronic and optical properties of single excitons and biexcitons in type-II quantum dot nanocrystals Energy Technology Data Exchange (ETDEWEB) Koc, Fatih, E-mail: fatih.koc@msn.com [Department of Physics, Faculty of Sciences, Selcuk University, 42075 Konya (Turkey); Sahin, Mehmet, E-mail: mehmet.sahin@agu.edu.tr, E-mail: mehsahin@gmail.com [Department of Physics, Faculty of Sciences, Selcuk University, 42075 Konya (Turkey); Department of Material Science and Nanotechnology Engineering, Abdullah Gül University, Kayseri (Turkey) 2014-05-21 In this study, a detailed investigation of the electronic and optical properties (i.e., binding energies, absorption wavelength, overlap of the electron-hole wave functions, recombination oscillator strength, etc.) of an exciton and a biexciton in CdTe/CdSe core/shell type-II quantum dot heterostructures has been carried out in the frame of the single band effective mass approximation. In order to determine the electronic properties, we have self-consistently solved the Poisson-Schrödinger equations in the Hartree approximation. We have considered all probable Coulomb interaction effects on both energy levels and also on the corresponding wave functions for both single exciton and biexciton. In addition, we have taken into account the quantum mechanical exchange-correlation effects in the local density approximation between same kinds of particles for biexciton. Also, we have examined the effect of the ligands and dielectric mismatch on the electronic and optical properties. We have used a different approximation proposed by Sahin and Koc [Appl. Phys. Lett. 102, 183103 (2013)] for the recombination oscillator strength of the biexciton for bound and unbound cases. The results obtained have been presented comparatively as a function of the shell thicknesses and probable physical reasons in behind of the results have been discussed in a detail. 20. Iterative Decoding for an Optical CDMA based Laser communication System Energy Technology Data Exchange (ETDEWEB) Kim, Jin Young; Kim, Eun Cheol [Kwangwoon Univ., Seoul (Korea, Republic of); Cha, Jae Sang [Seoul National Univ. of Technology, Seoul (Korea, Republic of) 2008-11-15 An optical CDMA(code division multiple access)based Laser communication system has attracted much attention since it requires minimal optical Laser signal processing and it is virtually delay free, while from the theoretical point of view, its performance depends on the auto and cross correlation properties of employed sequences. Various kinds of channel coding schemes for optical CDMA based Laser communication systems have been proposed and analyzed to compensate nonideal channel and receiver conditions in impaired photon channels. In this paper, we propose and analyze an iterative decoding of optical CDMA based Laser communication signals for both shot noise limited and thermal noise limited systems. It is assumed that optical channel is an intensity modulated (IM)channel and direct detection scheme is employed to detect the received optical signal. The performance is evaluated in terms of bit error probability and throughput. It is demonstrated that the BER and throughput performance is substantially improved with interleaver length for a fixed code rate and with alphabet size of PPM (pulse position modulation). Also, the BER and throughput performance is significantly enhanced with the number of iterations for decoding process. The results in this paper can be applied to the optical CDMA based Laser communication network with multiple access applications. 1. Antenna-load interactions at optical frequencies: impedance matching to quantum systems. Science.gov (United States) Olmon, R L; Raschke, M B 2012-11-09 resonance, (ii) subsequent transformation of that mode into a nanoscale spatial localization, and (iii) near-field coupling via an enhanced local density of states to a quantum load. These three steps define the goal of efficient transformation of incident radiation into a quantum excitation in an impedance-matched fashion. We review the physical basis of the light-matter interaction at the transition from the RF to optical regime, discuss the extension of antenna theory as needed for the design of impedance-matched optical antenna-load coupled systems, and provide several examples of the state of the art in design strategies and suggest future extensions. We furthermore suggest new performance metrics based on the combination of electric vector field, field enhancement and capture cross section measurement to aid in comparison between different antenna designs and optimization of optical antenna performance within the physical parameter space. 2. 450 d of Type II SN 2013ej in optical and near-infrared Science.gov (United States) Yuan, Fang; Jerkstrand, A.; Valenti, S.; Sollerman, J.; Seitenzahl, I. R.; Pastorello, A.; Schulze, S.; Chen, T.-W.; Childress, M. J.; Fraser, M.; Fremling, C.; Kotak, R.; Ruiter, A. J.; Schmidt, B. P.; Smartt, S. J.; Taddia, F.; Terreran, G.; Tucker, B. E.; Barbarino, C.; Benetti, S.; Elias-Rosa, N.; Gal-Yam, A.; Howell, D. A.; Inserra, C.; Kankare, E.; Lee, M. Y.; Li, K. L.; Maguire, K.; Margheim, S.; Mehner, A.; Ochner, P.; Sullivan, M.; Tomasella, L.; Young, D. R. 2016-09-01 We present optical and near-infrared photometric and spectroscopic observations of SN 2013ej, in galaxy M74, from 1 to 450 d after the explosion. SN 2013ej is a hydrogen-rich supernova, classified as a Type IIL due to its relatively fast decline following the initial peak. It has a relatively high peak luminosity (absolute magnitude MV = -17.6) but a small 56Ni production of ˜0.023 M⊙. Its photospheric evolution is similar to other Type II SNe, with shallow absorption in the Hα profile typical for a Type IIL. During transition to the radioactive decay tail at ˜100 d, we find the SN to grow bluer in B - V colour, in contrast to some other Type II supernovae. At late times, the bolometric light curve declined faster than expected from 56Co decay and we observed unusually broad and asymmetric nebular emission lines. Based on comparison of nebular emission lines most sensitive to the progenitor core mass, we find our observations are best matched to synthesized spectral models with a MZAMS = 12-15 M⊙ progenitor. The derived mass range is similar to but not higher than the mass estimated for Type IIP progenitors. This is against the idea that Type IIL are from more massive stars. Observations are consistent with the SN having a progenitor with a relatively low-mass envelope. 3. Design data brochure for a pyramidal optics solar system Energy Technology Data Exchange (ETDEWEB) 1980-09-01 This Design Data Brochure provides information on a Pyramidal Optics Solar System for solar heating and domestic hot water. The system is made up of the collecting, storage, and distribution subsystems. Contained in the brochure are such items as system description, available accessories, installation arrangements, physical data, piping and wiring diagrams, and guide specifications. 4. Circular dichroism and optical absorption spectra of mononuclear and trinuclear chiral Cu(II) amino-alcohol coordinated compounds: A combined theoretical and experimental study Science.gov (United States) Valencia, Israel; Ávila-Torres, Yenny; Barba-Behrens, Norah; Garzón, Ignacio L. 2015-04-01 Studies on the physicochemical properties of biomimetic compounds of multicopper oxidases are fundamental to understand their reaction mechanisms and catalytic behavior. In this work, electronic, optical, and chiroptical properties of copper(II) complexes with amino-alcohol chiral ligands are theoretically studied by means of time-dependent density functional theory. The calculated absorption and circular dichroism spectra are compared with experimental measurements of these spectra for an uncoordinated pseudoephedrine derivative, as well as for the corresponding mononuclear and trinuclear copper(II)-coordinated complexes. This comparison is useful to gain insights into their electronic structure, optical absorption and optical activity. The optical absorption and circular dichroism bands of the pseudoephedrine derivative are located in the UV-region. They are mainly due to transitions originated from n to π anti-bonding orbitals of the alcohol and amino groups, as well as from π bonding to π anti-bonding orbitals of carboxyl and phenyl groups. In the case of the mononuclear and trinuclear compounds, additional signals in the visible spectral region are present. In both systems, the origin of these bands is due to charge transfer from ligand to metal and d-d transitions. 5. Sensory systems II senses other than vision CERN Document Server Wolfe, Jeremy M 1988-01-01 This series of books, "Readings from the Encyclopedia of Neuroscience." consists of collections of subject-clustered articles taken from the Encyclopedia of Neuroscience. The Encyclopedia of Neuroscience is a reference source and compendium of more than 700 articles written by world authorities and covering all of neuroscience. We define neuroscience broadly as including all those fields that have as a primary goal the under­ standing of how the brain and nervous system work to mediate/control behavior, including the mental behavior of humans. Those interested in specific aspects of the neurosciences, particular subject areas or specialties, can of course browse through the alphabetically arranged articles of the En­ cyclopedia or use its index to find the topics they wish to read. However. for those readers-students, specialists, or others-who will find it useful to have collections of subject-clustered articles from the Encyclopedia, we issue this series of "Readings" in paperback. Students in neuroscienc... 6. QKD system with fast active optical path length compensation Science.gov (United States) Park, Byung Kwon; Lee, Min Soo; Woo, Min Ki; Kim, Yong-Su; Han, Sang-Wook; Moon, Sung 2017-06-01 We develop a quantum key distribution (QKD) system with fast active optical path length compensation. A rapid and reliable active optical path length compensation scheme is proposed and applied to a plug-and-play QKD system. The system monitors changes in key rates and controls it is own operation automatically. The system achieves its optimal performance within three seconds of operation, which includes a sifted key rate of 5.5 kbps and a quantum bit error rate of less than 2% after an abrupt temperature variation along the 25 km quantum channel. The system also operates well over a 24 h period while completing more than 60 active optical path length compensations. 7. OPSys: optical payload systems facility for testing space coronagraphs Science.gov (United States) Fineschi, S.; Crescenzio, G.; Massone, G.; Capobianco, G.; Zangrilli, L.; Antonucci, E.; Anselmi, F. 2011-10-01 The Turin Astronomical Observatory, Italy, has implemented in ALTEC, Turin, a new Optical Payload Systems (OPSys) facility for testing of contamination sensitive optical space flight instrumentation. The facility is specially tailored for tests on solar instruments like coronagraphs. OPSys comprises an ISO 7 clean room for instrument assembly and a relatively large (4.4 m3) optical test and calibration vacuum chamber: the Space Optics Calibration Chamber (SPOCC). SPOCC consists of a test section with a vacuum-compatible motorized optical bench, and of a pipeline section with a sun simulator at the opposite end of the optical bench hosting the instrumentation under tests. The solar simulator is an off-axis parabolic mirror collimating the light from the source with the solar angular divergence. After vacuum conditioning, the chamber will operate at an ultimate pressure of 10-6 mbar. This work describes the SPOCC's vacuum system and optical design, and the post-flight stray-light tests to be carried out on the Sounding-rocket Experiment (SCORE). This sub-orbital solar coronagraph is the prototype of the METIS coronagraph for the ESA Solar Orbital mission whose closest perihelion is one-third of the Sun-Earth distance. The plans are outlined for testing METIS in the SPOCC simulating the observing conditions from the Solar Orbiter perihelion. 8. Complex fluids: probing mechanical properties of biological systems with optical tweezers. Science.gov (United States) Ou-Yang, H Daniel; Wei, Ming-Tzo 2010-01-01 The mechanical properties of cells are crucial for cell sensing and reaction to mechanical environments. This review describes the basic principles of optical tweezers and their use as force sensors for studying the mechanical properties of biological systems. It covers experiments of four groups of biological systems arranged by increasing complexity: (a) packaging DNA into viral capsids by bacteriophage portal motors and the dynamical stiffness of DNA upon protein binding, (b) actin-coated giant vesicles and the myosin-II embedded actin polymer network, (c) suspension cells, and (d) adhesion cells. These examples demonstrate how optical tweezers have been used to improve the understanding of the mechanical properties of biological systems at subcellular and molecular levels. 9. The Impact Of Optical Storage Technology On Image Processing Systems Science.gov (United States) Garges, Daniel T.; Durbin, Gerald T. 1984-09-01 The recent announcement of commercially available high density optical storage devices will have a profound impact on the information processing industry. Just as the initial introduction of random access storage created entirely new processing strategies, optical technology will allow dramatic changes in the storage, retrieval, and dissemination of engineering drawings and other pictorial or text-based documents. Storage Technology Corporation has assumed a leading role in this arena with the introduction of the 7600 Optical Storage Subsystem, and the formation of StorageTek Integrated Systems, a subsidiary chartered to incorporate this new technology into deliverable total systems. This paper explores the impact of optical storage technology from the perspective of a leading-edge manufacturer and integrator. 10. Ultra-high Frequency Linear Fiber Optic Systems CERN Document Server Lau, Kam Y 2009-01-01 Designed for a one-semester course on fiber-optics systems and communication links, this book provides a concise but rigorous treatment of the theory and practice of analog (linear) fiber-optics links and systems that constitute the foundation of Hybrid Fiber Coax infrastructure in present-day CATV distribution and cable modem Internet access. Emerging applications in remote fiber-optic feed for free-space millimeter wave enterprise campus networks are also described. Issues such as dispersion and interferometric noise are treated quantitatively, and means for mitigating them are explained. This broad but concise text will thus be invaluable not only to students of fiber-optics communication but also to practicing engineers. 11. A compact fiber optic eye diagnostic system Science.gov (United States) Ansari, Rafat R.; Suh, Kwang I.; Dubin, Stephen; Della Vecchia, Michael A. 1996-03-01 A new fiber optic probe developed for determining transport properties of sub-micron particles in fluids experiments in a microgravity environment has been applied to study different parts of an eye. The probe positioned in front of an eye, delivers a low power (˜few μW) light from a laser diode into the eye and guides the light which is back scattered by different components (aqueous humor, lens, and vitreous humor) of the eye through a receiving optical fiber to a photo detector. The probe provides rapid determination of macromolecular diffusivities and their respective size distributions in the eye lens and the gel-like material in the vitreous humor. In a clinical setting, the probe can be mounted on a standard slit-lamp apparatus simply using a Hruby lens holder. The capability of detecting cataracts, both nuclear and cortical, in their early stages of formation, in a non invasive and quantitative fashion, has the potential in patient monitoring and in developing and testing new drugs or diet therapies to dissolve'' or slow down the cataract formation before the surgery becomes necessary. The ability to detect biochemical and macromolecular changes in the vitreous structure can be very useful in identifying certain diseases of the posterior chamber and their complications, e.g., posterior vitreous detachment and diabetic retinopathy. 12. Kilowatt isotope power system phase II plan. Volume II: flight System Conceptual Design (FSCD) Energy Technology Data Exchange (ETDEWEB) 1978-03-01 The Kilowatt Isotope Power System (KIPS) Flight System Conceptual Design (FSCD) is described. Included are a background, a description of the flight system conceptual design, configuration of components, flight system performance, Ground Demonstration System test results, and advanced development tests. 13. Development of distributed fiber optic sensor system for electric power systems (III) Energy Technology Data Exchange (ETDEWEB) Kim, Y. S.; Kim, Y. H.; Oh, S. K.; Kim, I. S.; Park, H. S.; Roh, J. D. [Korea Electrotechnology Research Institute, Changwon (Korea, Republic of) 1996-12-01 In this research, the key technologies on the distributed fiber optic sensor system are developed. A high power pulse laser driver is designed. The optical coupler for discriminating backscattered optical signal and splicing sensing optical fiber, and signal precessing electronics for receiving backscattered optical signal are also developed. By using the parallel processing algorithms and EPLD technique, a high speed signal averaged which is used for improving the signal to noise ratio is developed. A communication software used for sending received optical sensing signal to personal computer and a graphic software used for displaying the measurement result on personal computer monitor are developed. Test of distributed fiber optic sensor system was made along 1.1 km optical fiber. Experimental results show that optical output power of the backscattering light wave exponentially decays as the time axis approaches to the fiber end point. Since the output power of backscattered signal is very weak, the high speed signal averaged which is used for improving the signal to noise ratio and parallel processing algorithm which is used for reducing the measuring time of distributed fiber optic sensor system should be adapted. In the near future, developed technologies will be applied to new application system about distributed fiber optic sensor system. (author). refs., figs., tabs. 14. All-optical delay technique for supporting multiple antennas in a hybrid optical - wireless transmission system DEFF Research Database (Denmark) Prince, Kamau; Chiuchiarelli, A; Presi, M 2008-01-01 We introduce a novel continuously-variable optical delay technique to support beam-forming wireless communications systems using antenna arrays. We demonstrate delay with 64-QAM modulated signals at a rate of 15 Msymbol/sec with 2.5 GHz carrier frequency.......We introduce a novel continuously-variable optical delay technique to support beam-forming wireless communications systems using antenna arrays. We demonstrate delay with 64-QAM modulated signals at a rate of 15 Msymbol/sec with 2.5 GHz carrier frequency.... 15. Electro-optic side-chain polyimide system with large optical nonlinearity and high thermal stability Science.gov (United States) Sotoyama, Wataru; Tatsuura, Satoshi; Yoshimura, Tetsuzo 1994-04-01 We report electro-optic (EO) efficiency and thermal stability of a poled polyimide system with nonlinear optical dyes as side chains. The side-chain polyimide system is synthesized from a dianhydride containing azobenzene dye and a diamine. The dye in the polymer is chemically stable for temperatures below 250 °C. The polymer can be poled simultaneously with or after imidization of the polyamic acid. Our sample poled after imidization shows a large EO coefficient (r33=10.8 pm/V at λ=1.3 μm) and long-term thermal stability at 120 °C. 16. Optical response in a laser-driven quantum pseudodot system Energy Technology Data Exchange (ETDEWEB) Kilic, D. Gul [Physics Department, Graduate School of Natural and Applied Sciences, Dokuz Eylül University, 35390 Izmir (Turkey); Sakiroglu, S., E-mail: serpil.sakiroglu@deu.edu.tr [Physics Department, Faculty of Science, Dokuz Eylül University, 35390 Izmir (Turkey); Ungan, F.; Yesilgul, U. [Department of Optical Engineering, Faculty of Technology, Cumhuriyet University, 58140 Sivas (Turkey); Kasapoglu, E. [Physics Department, Faculty of Science, Cumhuriyet University, 58140 Sivas (Turkey); Sari, H. [Department of Primary Education, Faculty of Education, Cumhuriyet University, 58140 Sivas (Turkey); Sokmen, I. [Physics Department, Faculty of Science, Dokuz Eylül University, 35390 Izmir (Turkey) 2017-03-15 We investigate theoretically the intense laser-induced optical absorption coefficients and refractive index changes in a two-dimensional quantum pseudodot system under an uniform magnetic field. The effects of non-resonant, monochromatic intense laser field upon the system are treated within the framework of high-frequency Floquet approach in which the system is supposed to be governed by a laser-dressed potential. Linear and nonlinear absorption coefficients and relative changes in the refractive index are obtained by means of the compact-density matrix approach and iterative method. The results of numerical calculations for a typical GaAs quantum dot reveal that the optical response depends strongly on the magnitude of external magnetic field and characteristic parameters of the confinement potential. Moreover, we have demonstrated that the intense laser field modifies the confinement and thereby causes remarkable changes in the linear and nonlinear optical properties of the system. 17. Laboratory evaluation of Fecker and Loral optical IR PWI systems Science.gov (United States) Gorstein, M.; Hallock, J. N.; Houten, M.; Mcwilliams, I. G. 1971-01-01 A previous flight test of two electro-optical pilot warning indicators, using a flashing xenon strobe and silicon detectors as cooperative elements, pointed out several design deficiencies. The present laboratory evaluation program corrected these faults and calibrated the sensitivity of both systems in azimuth elevation and range. The laboratory tests were performed on an optical bench and consisted of three basic components: (1) a xenon strobe lamp whose output is monitored at the indicator detector to give pulse to pulse information on energy content at the receiver; (2) a strobe light attenuating optical system which is calibrated photometrically to provide simulated range; and (3) a positioning table on which the indicator system under study is mounted and which provides spatial location coordinates for all data points. The test results for both systems are tabulated. 18. Open Architecture Standards and Information Systems (OASIS II ... International Development Research Centre (IDRC) Digital Library (Canada) Open Architecture Standards and Information Systems (OASIS II) - Developing Capacity, Sharing Knowledge and Good Principles Across eHealth in Africa. Health care across much of the African continent is hampered by meager resources and a growing burden of disease, with HIV/AIDS, tuberculosis (TB) and malaria ... 19. Ruthenium (II)-bipyridyl with extended -system: Improved thermo ... Ruthenium(II)- bipyridyl with extended -system: Improved thermo-stable sensitizer for efficient and long-term durable dye sensitized solar cells ... The Density Functional Theory (DFT) and Time-Dependent DFT excited state calculations of the new sensitizer show that the first three HOMOs have t2g character with sizeable ... 20. Microstructured optical fibers for gas sensing systems Energy Technology Data Exchange (ETDEWEB) Challener, William Albert; Choudhury, Niloy; Palit, Sabarni 2017-10-17 1. Optical transfection using an endoscope-like system Science.gov (United States) Ma, Nan; Gunn-Moore, Frank; Dholakia, Kishan 2011-02-01 Optical transfection is a powerful method for targeted delivery of therapeutic agents to biological cells. A tightly focused pulsed laser beam may transiently change the permeability of a cell membrane to facilitate the delivery of foreign genetic material into cells. We report the first realization of an endoscope-like integrated system for optical transfection. An imaging fiber (coherent optical fiber bundle) with ~6000 cores (pixels) embedded in a fiber cladding of ~300 μm in diameter, produces an image circle (area) of ~270 μm diam. This imaging fiber, with an ordered axicon lens array chemically etched at its exit face, is used for the delivery of a femtosecond laser to the cell membrane for optical transfection along with subcellular resolution imaging. A microcapillary-based microfluidic system for localized drug delivery was also combined in this miniature, flexible system. Using this novel system, a plasmid transfection efficiency up to ~72% was obtained for CHO-K1 cells. This endoscope-like system opens a range of exciting applications, in particular, in the targeted in vivo optical microsurgery area. 2. Energy feedback system for the PLS-II linac Science.gov (United States) Kim, Changbum; Kim, Mungyung; Hwang, Ilmoon; Choi, Jae-Young; Shin, Seunghwan; Kim, Sung Chul; Park, Chongdo 2017-12-01 The upgraded Pohang-Light-Source (PLS-II) was opened to the public in 2012. Among many improvements of the PLS-II, a top-up operation was one of the highlights of them, and the stability of the electron beam was improved significantly. For the top-up operation, a stable injection from the linac to the storage ring was critically important, so that an energy feedback system was introduced to reduce the energy jitter of the linac electron beam. The result of the feedback system was successful and the measured energy jitter was less than ±0.1% (rms). In this work, the details of the energy feedback system are presented. It includes the setup for the energy feedback system, measurement results in the optimization process, and the future work for a better performance. 3. Digital Signal Processing for Optical Coherent Communication Systems DEFF Research Database (Denmark) Zhang, Xu wavelength division multiplex (U-DWDM) optical coherent systems based on 10-Gbaud QPSK. We report U-DWDM 1.2-Tb/s QPSK coherent system achieving spectral efficiency of 4.0-bit/s/Hz. In the experimental demonstration, digital decision feed back equalizer (DFE) algorithms and a finite impulse response (FIR... 4. Optical Disk for Digital Storage and Retrieval Systems. Science.gov (United States) Rose, Denis A. 1983-01-01 Availability of low-cost digital optical disks will revolutionize storage and retrieval systems over next decade. Three major factors will effect this change: availability of disks and controllers at low-cost and in plentiful supply; availability of low-cost and better output means for system users; and more flexible, less expensive communication… 5. TOD characterization of the gatekeeper electro optical security system NARCIS (Netherlands) Gosselink, G.A.B.; Anbeek, H.; Bijl, P.; Hogervorst, M.A. 2013-01-01 The Triangle Orientation Discrimination (TOD) test method was applied to characterize thermal and visual range performance of the Gatekeeper Electro Optical Security System. Gatekeeper developed by Thales Nederland BV, is currently in use with the Royal Netherlands Navy. The system houses uncooled 6. A simple and efficient optical character recognition system for basic ... Optical Character Recognition (OCR) systems have been effectively developed for the recognition of printed characters of non-Indian languages. Efforts are on the way for the development of efficient OCR systems for Indian languages, especially for Kannada, a popular South Indian language. We present in this paper an ... 7. Kansas Communication and Instruction System through Fiber-Optic Transmission. Science.gov (United States) Kansas State Dept. of Education, Topeka. Schools and communities will restructure as they move into the next decade. The success of this restructuring will be dependent upon access to and sharing of quality teaching and information through an expanded communication system. One of the major two-way interactive technologies is the fiber-optic cable: a delivery system that will provide… 8. CFO compensation method using optical feedback path for coherent optical OFDM system Science.gov (United States) Moon, Sang-Rok; Hwang, In-Ki; Kang, Hun-Sik; Chang, Sun Hyok; Lee, Seung-Woo; Lee, Joon Ki 2017-07-01 We investigate feasibility of carrier frequency offset (CFO) compensation method using optical feedback path for coherent optical orthogonal frequency division multiplexing (CO-OFDM) system. Recently proposed CFO compensation algorithms provide wide CFO estimation range in electrical domain. However, their practical compensation range is limited by sampling rate of an analog-to-digital converter (ADC). This limitation has not drawn attention, since the ADC sampling rate was high enough comparing to the data bandwidth and CFO in the wireless OFDM system. For CO-OFDM, the limitation is becoming visible because of increased data bandwidth, laser instability (i.e. large CFO) and insufficient ADC sampling rate owing to high cost. To solve the problem and extend practical CFO compensation range, we propose a CFO compensation method having optical feedback path. By adding simple wavelength control for local oscillator, the practical CFO compensation range can be extended to the sampling frequency range. The feasibility of the proposed method is experimentally investigated. 9. Electrochemical, linear optical, and nonlinear optical properties and interpretation by density functional theory calculations of (4-N,N-dimethylaminostyryl)-pyridinium pendant group associated with polypyridinic ligands and respective multifunctional metal complexes (Ru(II) or Zn(II)). Science.gov (United States) Dumur, Frédéric; Mayer, Cédric R; Hoang-Thi, Khuyen; Ledoux-Rak, Isabelle; Miomandre, Fabien; Clavier, Gilles; Dumas, Eddy; Méallet-Renault, Rachel; Frigoli, Michel; Zyss, Joseph; Sécheresse, Francis 2009-09-07 The synthesis, linear optical and nonlinear optical properties, as well as the electrochemical behavior of a series of pro-ligands containing the 4-(4-N,N-dimethylaminostyryl)-1-methyl pyridinium (DASP(+)) group as a push-pull moiety covalently linked to terpyridine or bipyridine as chelating ligands are reported in this full paper. The corresponding multifunctional Ru(II) and Zn(II) complexes were prepared and investigated. The structural, electronic, and optical properties of the pro-ligands and the ruthenium complexes were investigated using density functional theory (DFT) and time-dependent (TD) DFT calculations. A fairly good agreement was observed between the experimental and the calculated electronic spectra of the pro-ligands and their corresponding ruthenium complexes. A quenching of luminescence was evidenced in all ruthenium complexes compared with the free pro-ligands but even the terpyridine-functionalized metal complexes exhibited detectable luminescence at room temperature. Second order nonlinear optical (NLO) measurements were performed by Harmonic Light Scattering and the contribution of the DASP(+) moieties (and their relative ordering) and the metal-polypyridyl core need to be considered to explain the nonlinear optical properties of the metal complexes. 10. Closed loop high precision position control system with optical scale Science.gov (United States) Ge, Cheng-liang; Liao, Yuan; He, Zhong-wu; Luo, Zhong-xiang; Huang, Zhi-wei; Wan, Min; Hu, Xiao-yang; Fan, Guo-bin; Liang, Zheng 2008-03-01 With the developments of science of art, there are more and more demands on the high resolution control of position of object to be controlled, such as lathe, product line, elements in the optical resonant cavity, telescope, and so on. As one device with high resolution, the optical scale has more and more utility within the industrial and civil applications. With one optical scale and small DC servo motor, one closed loop high resolution position control system is constructed. This apparatus is used to control the position of the elements of optical system. The optical scale is attached on the object or reference guide way. The object position is sampled by a readhead of non-contact optical encoder. Control system processes the position information and control the position of object through the motion control of servo DC motor. The DC motor is controlled by one controller which is connected to an industrial computer. And the micro frictionless slide table does support the smooth motion of object to be controlled. The control algorithm of system is PID (Proportional-Integral-Differential) methods. The PID control methods have well ROBUST. The needed data to control are position, velocity and acceleration of the object. These three parameters correspond to the PID characters respectively. After the accomplishments of hardware, GUI (Graphical user interface), that is, the software of control system is also programmed. The whole system is assembled by specialized worker. Through calibration experiments, the coefficients of PID are obtained respectively. And then the precision of position control of the system is about 0.1μm. 11. Quantitative analysis of eyes and other optical systems in linear optics. Science.gov (United States) Harris, William F; Evans, Tanya; van Gool, Radboud D 2017-05-01 To show that 14-dimensional spaces of augmented point P and angle Q characteristics, matrices obtained from the ray transference, are suitable for quantitative analysis although only the latter define an inner-product space and only on it can one define distances and angles. The paper examines the nature of the spaces and their relationships to other spaces including symmetric dioptric power space. The paper makes use of linear optics, a three-dimensional generalization of Gaussian optics. Symmetric 2 × 2 dioptric power matrices F define a three-dimensional inner-product space which provides a sound basis for quantitative analysis (calculation of changes, arithmetic means, etc.) of refractive errors and thin systems. For general systems the optical character is defined by the dimensionally-heterogeneous 4 × 4 symplectic matrix S, the transference, or if explicit allowance is made for heterocentricity, the 5 × 5 augmented symplectic matrix T. Ordinary quantitative analysis cannot be performed on them because matrices of neither of these types constitute vector spaces. Suitable transformations have been proposed but because the transforms are dimensionally heterogeneous the spaces are not naturally inner-product spaces. The paper obtains 14-dimensional spaces of augmented point P and angle Q characteristics. The 14-dimensional space defined by the augmented angle characteristics Q is dimensionally homogenous and an inner-product space. A 10-dimensional subspace of the space of augmented point characteristics P is also an inner-product space. The spaces are suitable for quantitative analysis of the optical character of eyes and many other systems. Distances and angles can be defined in the inner-product spaces. The optical systems may have multiple separated astigmatic and decentred refracting elements. © 2017 The Authors Ophthalmic & Physiological Optics © 2017 The College of Optometrists. 12. The structural and optical properties of GaSb/InGaAs type-II quantum dots grown on InP (100) substrate. Science.gov (United States) Shuhui, Zhang; Lu, Wang; Zhenwu, Shi; Yanxiang, Cui; Haitao, Tian; Huaiju, Gao; Haiqiang, Jia; Wenxin, Wang; Hong, Chen; Liancheng, Zhao 2012-01-25 We have investigated the structural and optical properties of type-II GaSb/InGaAs quantum dots [QDs] grown on InP (100) substrate by molecular beam epitaxy. Rectangular-shaped GaSb QDs were well developed and no nanodash-like structures which could be easily found in the InAs/InP QD system were formed. Low-temperature photoluminescence spectra show there are two peaks centered at 0.75eV and 0.76ev. The low-energy peak blueshifted with increasing excitation power is identified as the indirect transition from the InGaAs conduction band to the GaSb hole level (type-II), and the high-energy peak is identified as the direct transition (type-I) of GaSb QDs. This material system shows a promising application on quantum-dot infrared detectors and quantum-dot field-effect transistor. 13. Optical diagnostics based on elastic scattering: An update of clinical demonstrations with the Optical Biopsy System Energy Technology Data Exchange (ETDEWEB) Bigio, I.J.; Boyer, J.; Johnson, T.M.; Lacey, J.; Mourant, J.R. [Los Alamos National Lab., NM (United States); Conn, R. [Lovelace Medical Center, Albuquerque, NM (United States); Bohorfoush, A. [Wisconsin Medical School, Milwaukee, WI (United States) 1994-10-01 The Los Alamos National Laboratory has continued the development of the Optical Biopsy System (OBS) for noninvasive, real-time in situ diagnosis of tissue pathologies. Our clinical studies have expanded since the last Biomedical Optics Europe conference (Budapest, September 1993), and we report here on the latest results of clinical tests in gastrointestinal tract. The OBS invokes a unique approach to optical diagnosis of tissue pathologies based on the elastic scattering properties, over a wide range of wavelengths, of the tissue. The use of elastic scattering as the key to optical tissue diagnostics in the OBS is based on the fact that many tissue pathologies, including a majority of cancer forms, manifest significant architectural changes at the cellular and sub-cellular level. Since the cellular components that cause elastic scattering have dimensions typically on the order of visible to near-IR wavelengths, the elastic (Mie) scattering properties will be wavelength dependent. Thus, morphology and size changes can be expected to cause significant changes in an optical signature that is derived from the wavelength-dependence of elastic scattering. The OBS employs a small fiberoptic probe that is amenable to use with any endoscope or catheter, or to direct surface examination. The probe is designed to be used in optical contact with the tissue under examination and has separate illuminating and collecting fibers. Thus, the light that is collected and transmitted to the analyzing spectrometer must first scatter through a small volume of the tissue before entering the collection fiber(s). Consequently, the system is also sensitive to the optical absorption spectrum of the tissue, over an effective operating range of <300 to 950 nm, and such absorption adds valuable complexity to the scattering spectral signature. 14. Fiber laser master oscillators for optical synchronization systems Energy Technology Data Exchange (ETDEWEB) Winter, A. 2008-04-15 New X-ray free electron lasers (e.g. the European XFEL) require a new generation of synchronization system to achieve a stability of the FEL pulse, such that pump-probe experiments can fully utilize the ultra-short pulse duration (50 fs). An optical synchronization system has been developed based on the distribution of sub-ps optical pulses in length-stabilized fiber links. The synchronization information is contained in the precise repetition frequency of the optical pulses. In this thesis, the design and characterization of the laser serving as laser master oscillator is presented. An erbium-doped mode-locked fiber laser was chosen. Amplitude and phase noise were measured and record-low values of 0.03 % and 10 fs for the frequency range of 1 kHz to the Nyquist frequency were obtained. Furthermore, an initial proof-of-principle experiment for the optical synchronization system was performed in an accelerator environment. In this experiment, the fiber laser wase phase-locked to a microwave reference oscillator and a 500 meter long fiber link was stabilized to 12 fs rms over a range of 0.1 Hz to 20 kHz. RF signals were obtained from a photodetector without significant degradation at the end of the link. Furthermore, the laser master oscillator for FLASH was designed and is presently in fabrication and the initial infrastructure for the optical synchronization system was setup. (orig.) 15. The pupils and optical systems of gecko eyes. Science.gov (United States) Roth, Lina S V; Lundström, Linda; Kelber, Almut; Kröger, Ronald H H; Unsbo, Peter 2009-03-30 The nocturnal helmet gecko, Tarentola chazaliae, discriminates colors in dim moonlight when humans are color blind. The sensitivity of the helmet gecko eye has been calculated to be 350 times higher than human cone vision at the color vision threshold. The optics and the large cones of the gecko are important reasons why they can use color vision at low light intensities. Using photorefractometry and an adapted laboratory Hartmann-Shack wavefront sensor of high resolution, we also show that the optical system of the helmet gecko has distinct concentric zones of different refractive powers, a so-called multifocal optical system. The intraspecific variation is large but in most of the individuals studied the zones differed by 15 diopters. This is of the same magnitude as needed to focus light of the wavelength range to which gecko photoreceptors are most sensitive. We compare the optical system of the helmet gecko to that of the diurnal day gecko, Phelsuma madagascariensis grandis. The optical system of the day gecko shows no signs of distinct concentric zones and is thereby monofocal. 16. Chinese National Optical Education Small Private Online Course system Science.gov (United States) Zhang, XiaoJie; Lin, YuanFang; Liu, Xu; Liu, XiangDong; Cen, ZhaoFeng; Li, XiaoTong; Zheng, XiaoDong; Wang, XiaoPing 2017-08-01 In order to realize the sharing of high quality course resources and promote the deep integration of Internet+' higher education and talent training, a new on-line to off-line specialized courses teaching mode was explored in Chinese colleges and universities, which emphasized different teaching places, being organized asynchronously and localized. The latest progress of the Chinese National Optical Education Small Private On-line Course (CNOESPOC) system set up by Zhejiang University and other colleges and universities having disciplines in the field of optics and photonics under the guidance of the Chinese National Steering Committee of Optics and Photonics (CNSCOP) was introduced in this paper. The On-line to Off-line (O2O) optical education teaching resource sharing practice offers a new good example for higher education in China under the background of Internet +. 17. Design of optical systems for large space telescopes Science.gov (United States) Malamed, Evgeny R.; Sokolsky, M. N. 1995-09-01 On the basis of long-term experience of LOMO PLC in creating large optical systems for ground and space telescopes, with diameter of primary mirror from 1 to 6 meters, the following issues should be considered: principles of constructing optical systems for space telescopes and selecting their optimum design in respect of dimensions/mass and performance criteria; ensuring the fulfillment of image quality requirements in the process of manufacturing optical systems for controlling ground telescope elements in operating conditions; providing automatic adjustment of telescope secondary mirror, automatic focusing, interferometric control of image quality by means of stellar interferometer with radial shift and internal control with Gartman's test. Description of space telescope equipped with primary mirror of diameter 1.5 m, manufactured in LOMO PLC, is given. 18. Integrated intravascular optical coherence tomography ultrasound imaging system Science.gov (United States) Yin, Jiechen; Yang, Hao-Chung; Li, Xiang; Zhang, Jun; Zhou, Qifa; Hu, Changhong; Shung, K. Kirk; Chen, Zhongping 2010-01-01 We report on a dual-modality optical coherence tomography (OCT) ultrasound (US) system for intravascular imaging. To the best of our knowledge, we have developed the first integrated OCT-US probe that combines OCT optical components with an US transducer. The OCT optical components mainly consist of a single-mode fiber, a gradient index lens for light-beam focusing, and a right-angled prism for reflecting light into biological tissue. A 40-MHz piezoelectric transducer (PZT-5H) side-viewing US transducer was fabricated to obtain the US image. These components were integrated into a single probe, enabling both OCT and US imaging at the same time. In vitro OCT and ultrasound images of a rabbit aorta were obtained using this dual-modality imaging system. This study demonstrates the feasibility of an OCT-US system for intravascular imaging, which is expected to have a prominent impact on early detection and characterization of atherosclerosis. 19. Texture information processing system with binary optical wavelet element Science.gov (United States) Feng, Wenyi; Yan, Yingbai; Jin, Guofan; Wang, Wenlu; Wu, Minxian 1996-12-01 The implementation strategy of optical wavelet transform for texture information processing is discussed in this paper. An opto-electronic hybrid system is constructed for texture segmentation, which is based on the multi-channel filtering framework in the early stages of human visual theory. First, a traditional optical system with a Damman grating as the beam splitter and a bank of Gabor wavelets as the channel filters is set up for feature extraction, and several clustering algorithms are then used for feature integration. Furthermore, a novel binary optical element with the functions of splitting, filtering and imaging is designed and fabricated to simplify the traditional system. The experimental results and the primary applications are also provided. 20. Performance assessment of the SOFA, APACHE II scoring system, and SAPS II in intensive care unit organophosphate poisoned patients. Science.gov (United States) Kim, Yong Hwan; Yeo, Jung Hoon; Kang, Mun Ju; Lee, Jun Ho; Cho, Kwang Won; Hwang, SeongYoun; Hong, Chong Kun; Lee, Young Hwan; Kim, Yang Weon 2013-12-01 This study assessed the ability of the Sequential Organ Failure Assessment (SOFA) and Acute Physiology, Chronic Health Evaluation (APACHE) II scoring systems, as well as the Simplified Acute Physiology Score (SAPS) II method to predict group mortality in intensive care unit (ICU) patients who were poisoned with organophosphate. The medical records of 149 organophosphate poisoned patients admitted to the ICU from September 2006 to December 2012 were retrospectively examined. The SOFA, APACHE II, and SAPS II were calculated based on initial laboratory data in the Emergency Department, and during the first 24 hr of ICU admission. The probability of death was calculated for each patient based on the SOFA score, APACHE II score, and SAPS II equations. The ability to predict group mortality by the SOFA score, APACHE II score, and SAPS II method was assessed using two by two decision matrices and receiver operating characteristic (ROC) curve analysis. A total of 131 patients (mean age, 61 yr) were enrolled. The sensitivities, specificities, and accuracies were 86.2%, 82.4%, and 83.2% for the SOFA score, respectively; 65.5%, 68.6%, and 67.9% for the APACHE II scoring system, respectively; and 86.2%, 77.5%, and 79.4% for the SAPS II, respectively. The areas under the curve in the ROC curve analysis for the SOFA score, APACHE II scoring system, and SAPS II were 0.896, 0.716, and 0.852, respectively. In conclusion, the SOFA, APACHE II, and SAPS II have different capability to discriminate and estimate early in-hospital mortality of organophosphate poisoned patients. The SOFA score is more useful in predicting mortality, and easier and simpler than the APACHE II and SAPS II. 1. Camera, handlens, and microscope optical system for imaging and coupled optical spectroscopy Science.gov (United States) Mungas, Greg S. (Inventor); Boynton, John (Inventor); Sepulveda, Cesar A. (Inventor); Nunes de Sepulveda, legal representative, Alicia (Inventor); Gursel, Yekta (Inventor) 2012-01-01 An optical system comprising two lens cells, each lens cell comprising multiple lens elements, to provide imaging over a very wide image distance and within a wide range of magnification by changing the distance between the two lens cells. An embodiment also provides scannable laser spectroscopic measurements within the field-of-view of the instrument. 2. Doppler reflectometer system in the stellarator TJ-II. Science.gov (United States) Happel, T; Estrada, T; Blanco, E; Tribaldos, V; Cappa, A; Bustos, A 2009-07-01 A Doppler reflectometer system has recently been installed in the stellarator TJ-II. The system is optimized for the Q-band (33-50 GHz) and the high-curvature plasmas produced in TJ-II. The launch angle of the microwave beam can be controlled by a steerable mirror to obtain angles between +/-20 degrees enabling the measurement of perpendicular wave numbers in the range of 3-15 cm(-1). The available angular range allows for comparisons between positive and negative values and additionally for calibration of the system. Localization and k(perpendicular)-estimation is done via the three-dimensional ray/beam-tracing code TRUBA. First measured spectra and radial profiles of the perpendicular velocity of plasma density fluctuations are presented. 3. The optical system for the Big European Bubble Chamber CERN Document Server Harigel, G G 1977-01-01 The optical system for the new giant bubble chamber, built for the European Organization for Nuclear Research (CERN), consists of four sets of fisheye windows, each equipped with a wide-angle lens which has an aperture angle of 108 degrees , while the fifth set has a periscope for visual observation of the chamber interior. Each of the fisheye sets is assembled from three hemispherical windows. The largest hemisphere is made from Schott BK7 glass and is exposed to the temperature of liquid hydrogen. The entire optical system has been operated successfully for the past 4 years. (13 refs). 4. Battle Staff Training System II: Computer-Based Instruction Supporting the Force XXI Training Program National Research Council Canada - National Science Library Wampler, Richard 1998-01-01 This report documents the methodology and lessons learned in the development of the Innovative Tools and Techniques for Brigade and Below Staff Training II - Battle Staff Training System II (ITTBBST-BSTS II... 5. On the temperature dependence of the optical spectral weight in correlated electron systems Energy Technology Data Exchange (ETDEWEB) Millis, A.J. [AT& T Bell Laboratories, Murray Hill, NJ (United States) 1994-12-31 A temperature dependence of the low frequency optical spectral weight has recently been observed in several strongly correlated insulating or nearly insulating systems including FeSi, Ce{sub 3}Bi{sub 4}Pt{sub 3} and V{sub 2}O{sub 3{minus}y}, and Bi{sub 2}Te{sub 3}. This temperature dependence is at first sight surprising because one is accustomed to thinking of optical spectral weigth in terms of the f-sum rule, which in its most general form states that the integral of any of the diagonal components of the optical conductivity {sigma}{sub ii}({omega}) is {pi}ne{sup 2}/2m. This is not very useful in most condensed matter physics contexts because the quantity n is the total number of electrons, including e.g. those in core levels, and because one must extend the integral to energies greater than the binding energy of the 1s shell to exhaust the sum rule. In condensed matter problems one typically focuses on a small number of bands close to the chemical potential. One may then ask what is the restricted sum rule governing optical transitions involving only these bands. Of course, the answer to this question is useful only if optical transitions involving bands retained in the model can be separated from those involving bands not retained. 6. Bulk crystal growth and nonlinear optical characterization of semiorganic single crystal: Cadmium (II) dibromide L - Proline monohydrate Energy Technology Data Exchange (ETDEWEB) Balakrishnan, T., E-mail: balacrystalgrowth@gmail.com [Crystal Growth Laboratory, PG & Research Department of Physics, Periyar EVR College (Autonomous), Tiruchirappalli, 620 023, Tamil Nadu (India); Sathiskumar, S. [Crystal Growth Laboratory, PG & Research Department of Physics, Periyar EVR College (Autonomous), Tiruchirappalli, 620 023, Tamil Nadu (India); Ramamurthi, K. [Crystal Growth and Thin Film Laboratory, Department of Physics and Nanotechnology, SRM University, Kattankulathur, 603 203, Kancheepuram, Tamil Nadu (India); Thamotharan, S. [Department of Bioinformatics, School of Chemical and Biotechnology, SASTRA University, Thanjavur, 613 401 (India) 2017-01-15 Single crystal of a novel metal organic nonlinear optical (NLO) cadmium (II) dibromide L - proline monohydrate (CBLPM) of size 7 × 7 × 5 mm{sup 3} was grown from slow evaporation technique. Single crystal X – ray diffraction analysis reveals that the crystal belongs to orthorhombic system with lattice parameters a = 10.1891 (8) Å, b = 13.4961 (11) Å, c = 7.4491 (5) Å and space group P2{sub 1}2{sub 1}2{sub 1}. The powder X – ray diffraction pattern of CBLPM was recorded and the X – ray diffraction peaks were indexed. The various functional groups of CBLPM were identified by the FT – IR and FT – Raman spectral analyses. The optical transmittance window and lower cut off wavelength of CBLPM were identified from UV – Vis – NIR studies. The mechanical strength of the grown crystal was estimated using Vickers microhardness test. Dielectric constant and dielectric loss measurements were carried out at different temperatures in the frequency range of 50 Hz - 2 MHz. The photoluminescence spectrum was recorded in the wavelength range 200–400 nm and the estimated optical band gap was ∼4.1 eV. Etching studies were carried out for different etching time. Thermal stability of CBLPM was determined using thermogravimetric analysis. Laser induced damage threshold study was carried out for the grown crystal using Nd:YAG laser. Size dependent second harmonic generation efficiency of the grown crystal was determined by Kurtz and Perry powder technique with different particle size using Nd:YAG laser with wavelength 1064 nm. Second harmonic generation efficiency of the powdered CBLPM crystal was ∼2.3 times that of potassium dihydrogen orthophosphate. - Highlights: • CBLPM crystal belongs to orthorhombic crystal system with space group P2{sub 1}2{sub 1}2{sub 1.} • Transmittance of CBLPM is ∼80% in the 650–1100 nm range. • Powder SHG efficiency of CBLPM increases with increase in particle size. • SHG efficiency of 0.57 μm size powdered CBLPM is ∼2 7. Automated optical sensing system for biochemical assays Science.gov (United States) Oroszlan, Peter; Duveneck, Gert L.; Ehrat, Markus; Widmer, H. M. 1994-03-01 In this paper, we present a new system called FOBIA that was developed and optimized with respect to automated operation of repetitive assay cycles with regenerable bioaffinity sensors. The reliability and precision of the new system is demonstrated by an application in a competitive assay for the detection of the triazine herbicide Atrazine. Using one sensor in more than 300 repetitive cycles, a signal precision better than 5% was achieved. 8. ATHENA: system studies and optics accommodation Science.gov (United States) Ayre, M.; Bavdaz, M.; Ferreira, I.; Wille, E.; Fransen, S.; Stefanescu, A.; Linder, M. 2016-07-01 ATHENA is currently in Phase A, with a view to adoption upon a successful Mission Adoption Review in 2019/2020. After a brief presentation of the reference spacecraft (SC) design, this paper will focus on the functional and environmental requirements, the thermo-mechanical design and the Assembly, Integration, Verification & Test (AIVT) considerations related to housing the Silicon Pore Optics (SPO) Mirror Modules (MM) in the very large Mirror Assembly Module (MAM). Initially functional requirements on the MM accommodation are presented, with the Effective Area and Half Energy Width (HEW) requirements leading to a MAM comprising (depending on final mirror size selected) between 700-1000 MMs, co-aligned with exquisite accuracy to provide a common focus. A preliminary HEW budget allocated across the main error-contributors is presented, and this is then used as a reference to derive subsequent requirements and engineering considerations, including: The procedures and technologies for MM-integration into the Mirror Structure (MS) to achieve the required alignment accuracies in a timely manner; stiffness requirements and handling scheme required to constrain deformation under gravity during x-ray testing; temperature control to constrain thermo-elastic deformation during flight; and the role of the Instrument Switching Mechanism (ISM) in constraining HEW and Effective Area errors. Next, we present the key environmental requirements of the MMs, and the need to minimise shock-loading of the MMs is stressed. Methods to achieve this Ø are presented, including: Selection of a large clamp-band launch vehicle interface (LV I/F); lengthening of the shock-path from the LV I/F to the MAM I/F; modal-tuning of the MAM to act as a low-pass filter during launch shock events; use of low-shock HDRMs for the MAM; and the possibility to deploy a passive vibration solution at the LV I/F to reduce loads. 9. Development of distributed fiber-optic sensor system for electric power systems 2 Energy Technology Data Exchange (ETDEWEB) Kim, Young Soo; Kim, Yo Hee; Oh, Sang Ki; Kim, In Soo; Park, Hae Soo; Roh, Jong Dae [Korea Electrotechnology Research Institute, Changwon (Korea, Republic of) 1996-05-01 A specific scheme of distributed fiber-optic sensor system for electric power systems is designed in the second year research. A pulsed laser diode driver is designed and optical spectrum distribution of output light of laser diode is analyzed using optical spectrum analyzer. Optical coupler and optic system for selecting a specific wavelength from Raman backscattering light wave are designed. Optical receiving circuit signal processing electronics are designed and tested with 2.2 km length of optical fiber. Experimental results show that optical output power of the backscattering light wave exponentially decay as the tim axis approaches to the fiber end point. Analysis of Raman backscattering wave knows the temperature change of any specified point of optical fiber. The output power of Raman backscattering wave is about 30 db weaker than that of Rayleigh backscattering wave. Research of high speed averaging circuit should be performed for improvement of signal to noise ratio. And, Parallel signal processing algorithm should be adapted to shorten the measuring time of distributed fiber-optic sensor system. (author). 37 refs. 10. The simulation library of the Belle II software system Science.gov (United States) Kim, D. Y.; Ritter, M.; Bilka, T.; Bobrov, A.; Casarosa, G.; Chilikin, K.; Ferber, T.; Godang, R.; Jaegle, I.; Kandra, J.; Kodys, P.; Kuhr, T.; Kvasnicka, P.; Nakayama, H.; Piilonen, L.; Pulvermacher, C.; Santelj, L.; Schwenker, B.; Sibidanov, A.; Soloviev, Y.; Starič, M.; Uglov, T. 2017-10-01 SuperKEKB, the next generation B factory, has been constructed in Japan as an upgrade of KEKB. This brand new e+ e- collider is expected to deliver a very large data set for the Belle II experiment, which will be 50 times larger than the previous Belle sample. Both the triggered physics event rate and the background event rate will be increased by at least 10 times than the previous ones, and will create a challenging data taking environment for the Belle II detector. The software system of the Belle II experiment is designed to execute this ambitious plan. A full detector simulation library, which is a part of the Belle II software system, is created based on Geant4 and has been tested thoroughly. Recently the library has been upgraded with Geant4 version 10.1. The library is behaving as expected and it is utilized actively in producing Monte Carlo data sets for various studies. In this paper, we will explain the structure of the simulation library and the various interfaces to other packages including geometry and beam background simulation. 11. Neuroprotective Effects and Mechanisms of Curcumin–Cu(II and –Zn(II Complexes Systems and Their Pharmacological Implications Directory of Open Access Journals (Sweden) Fa-Shun Yan 2017-12-01 Full Text Available Alzheimer’s disease (AD is the main form of dementia and has a steadily increasing prevalence. As both oxidative stress and metal homeostasis are involved in the pathogenesis of AD, it would be interesting to develop a dual function agent, targeting the two factors. Curcumin, a natural compound isolated from the rhizome of Curcuma longa, is an antioxidant and can also chelate metal ions. Whether the complexes of curcumin with metal ions possess neuroprotective effects has not been evaluated. Therefore, the present study was designed to investigate the protective effects of the complexes of curcumin with Cu(II or Zn(II on hydrogen peroxide (H2O2-induced injury and the underlying molecular mechanisms. The use of rat pheochromocytoma (PC12 cells, a widely used neuronal cell model system, was adopted. It was revealed that curcumin–Cu(II complexes systems possessed enhanced O2·–-scavenging activities compared to unchelated curcumin. In comparison with unchelated curcumin, the protective effects of curcumin–Cu(II complexes systems were stronger than curcumin–Zn(II system. Curcumin–Cu(II or –Zn(II complexes systems significantly enhanced the superoxide dismutase, catalase, and glutathione peroxidase activities and attenuated the increase of malondialdehyde levels and caspase-3 and caspase-9 activities, in a dose-dependent manner. The curcumin–Cu(II complex system with a 2:1 ratio exhibited the most significant effect. Further mechanistic study demonstrated that curcumin–Cu(II or –Zn(II complexes systems inhibited cell apoptosis via downregulating the nuclear factor κB (NF-κB pathway and upregulating Bcl-2/Bax pathway. In summary, the present study found that curcumin–Cu(II or –Zn(II complexes systems, especially the former, possess significant neuroprotective effects, which indicates the potential advantage of curcumin as a promising agent against AD and deserves further study. 12. DT II (Development Test II) of LADDS (Laundry and Decontamination Dry Cleaning System). Volume 2 Science.gov (United States) 1991-03-12 17111 ED -"[-115"LAD-003 DT II LADM LAWIDY/DRY CLEANER GClLAD INDICATMIC THE SYSTEM lIS NON aHARCING. NO ACTION HAS TAKEN. AT 2030 HOURS, OPSHRS - 94...K / 11. MR 12, i1 gWP0RTABILI1 ANALYSIS CHART PACE: 23 PROJECT NUIRER PROJECT MONE ITE 13 -iS-115-LAD-003 IT Il LADMS LAUNDRY/DR? CLEANER LADD01 INC...S-115-LADO-03 IT II LADMS LAUIDRY/DRY CLEANER LADDOl INC-DOTI: 900918 ---------SCORING INFOR1nATION _ TIR1: L5"-0000" 01 INC CLASS: fNIJOR STEP CLASS 13. Defining the uncertainty of electro-optical identification system performance estimates using a 3D optical environment derived from satellite Science.gov (United States) Ladner, S. D.; Arnone, R.; Casey, B.; Weidemann, A.; Gray, D.; Shulman, I.; Mahoney, K.; Giddings, T.; Shirron, J. 2009-05-01 Current United States Navy Mine-Counter-Measure (MCM) operations primarily use electro-optical identification (EOID) sensors to identify underwater targets after detection via acoustic sensors. These EOID sensors which are based on laser underwater imaging by design work best in "clear" waters and are limited in coastal waters especially with strong optical layers. Optical properties and in particular scattering and absorption play an important role on systems performance. Surface optical properties alone from satellite are not adequate to determine how well a system will perform at depth due to the existence of optical layers. The spatial and temporal characteristics of the 3d optical variability of the coastal waters along with strength and location of subsurface optical layers maximize chances of identifying underwater targets by exploiting optimum sensor deployment. Advanced methods have been developed to fuse the optical measurements from gliders, optical properties from "surface" satellite snapshot and 3-D ocean circulation models to extend the two-dimensional (2-D) surface satellite optical image into a three-dimensional (3-D) optical volume with subsurface optical layers. Modifications were made to an EOID performance model to integrate a 3-D optical volume covering an entire region of interest as input and derive system performance field. These enhancements extend present capability based on glider optics and EOID sensor models to estimate the system's "image quality". This only yields system performance information for a single glider profile location in a very large operational region. Finally, we define the uncertainty of the system performance by coupling the EOID performance model with the 3-D optical volume uncertainties. Knowing the ensemble spread of EOID performance field provides a new and unique capability for tactical decision makers and Navy Operations. 14. Self-characterization of linear and nonlinear adaptive optics systems. Science.gov (United States) Hampton, Peter J; Conan, Rodolphe; Keskin, Onur; Bradley, Colin; Agathoklis, Pan 2008-01-10 We present methods used to determine the linear or nonlinear static response and the linear dynamic response of an adaptive optics (AO) system. This AO system consists of a nonlinear microelectromechanical systems deformable mirror (DM), a linear tip-tilt mirror (TTM), a control computer, and a Shack-Hartmann wavefront sensor. The system is modeled using a single-input-single-output structure to determine the one-dimensional transfer function of the dynamic response of the chain of system hardware. An AO system has been shown to be able to characterize its own response without additional instrumentation. Experimentally determined models are given for a TTM and a DM. 15. Large-objective optical system for TV night observation Science.gov (United States) He, Xin; Pan, Junhua 1993-04-01 A relatively large aperture (375 mm) optical system working at f/1.6 for a low-light-level movable TV camera has been designed. In the process of designing, advantages and defects of a catadioptric system with all spherical surfaces and three kinds of aspherical systems have been compared. The optical system we finally adopted is a reflecting system with two aspherical surfaces. Its primary mirror is hyperboloid and the secondary is a high order planary aspherical surface. Maximal asphericity is 0.048 mm and 0.007 mm, respectively. There are also two small correcting lenses (crown and flint glass) near focus in order to correct curvature of field and astigmatism. Within 2.4 degree(s) field of view, spot diagrams show that the maximal image is 0.035 mm in diameter with a flat focal plane over the wavelength 4861 angstrom to 6563 angstrom. The total weight of the optical elements is about 14 Kg. The optical tube length is about 277 mm and its T/Number is 2. 16. Optical character recognition systems for different languages with soft computing CERN Document Server Chaudhuri, Arindam; Badelia, Pratixa; K Ghosh, Soumya 2017-01-01 The book offers a comprehensive survey of soft-computing models for optical character recognition systems. The various techniques, including fuzzy and rough sets, artificial neural networks and genetic algorithms, are tested using real texts written in different languages, such as English, French, German, Latin, Hindi and Gujrati, which have been extracted by publicly available datasets. The simulation studies, which are reported in details here, show that soft-computing based modeling of OCR systems performs consistently better than traditional models. Mainly intended as state-of-the-art survey for postgraduates and researchers in pattern recognition, optical character recognition and soft computing, this book will be useful for professionals in computer vision and image processing alike, dealing with different issues related to optical character recognition. 17. Optical sensors and their applications for probing biological systems DEFF Research Database (Denmark) Palanco, Marta Espina There is a great interest in exploring and developing new optical sensitive methodologies for probing complex biological systems. In this project we developed non-invasive and sensitive biosensor strategies for studying physiologically relevant chemical and physical properties of plant and mammal......There is a great interest in exploring and developing new optical sensitive methodologies for probing complex biological systems. In this project we developed non-invasive and sensitive biosensor strategies for studying physiologically relevant chemical and physical properties of plant...... be trapped and deformed via variations of the light intensity through the optical fibers. Of key importance was the ability to create stable laminar flows at low velocity fields. We therefore optimized the presented device to be able to trap a large number of cells and to exchange the local environment... 18. Detection system for optical coherence tomography: Czerny-Turner spectrometer Science.gov (United States) Kamińska, Aleksandra 2017-08-01 Research methods based on spectral analysis have powerful impact on development in many field of science. Signal spectrum can be a source of useful and important data. It enables to obtain information about physical and chemical properties of tested materials. This paper has been devoted to describe optical design for high resolution spectrometer, which is significant element of optical coherence tomography (OCT) systems. Designed spectrometer is working in visible range (450-830 nm). Czerny-Turner configuration enables to correcting astigmatism and coma aberration over full bandwidth. Moreover, spectrometer has uncomplicated construction. Merely, two mirrors and diffraction gratings allows to design low - cost spectrometer with satisfying optical properties. Spectrum detection has been realized using CMOS line scan sensors with 6144 pixels. It provides high speed and resolution of the system. 19. SAFARI optical system architecture and design concept NARCIS (Netherlands) Pastor, Carmen; Jellema, Willem; Zuluaga-Ramírez, Pablo; Arrazola, David; Fernández-Rodriguez, M.; Belenguer, Tomás; González Fernández, Luis M.; Audley, Michael D.; Evers, Jaap; Eggens, Martin; Torres Redondo, Josefina; Najarro, Francisco; Roelfsema, P. 2016-01-01 SpicA FAR infrared Instrument, SAFARI, is one of the instruments planned for the SPICA mission. The SPICA mission is the next great leap forward in space-based far-infrared astronomy and will study the evolution of galaxies, stars and planetary systems. SPICA will utilize a deeply cooled 2.5m-class 20. DWDM Fiber-Wireless Access System with Centralized Optical Frequency Comb-based RF Carrier Generation DEFF Research Database (Denmark) Pang, Xiaodan; Beltrán, Marta; Sánchez, José 2013-01-01 We propose and experimentally demonstrate an optical wireless DWDM system at 60 GHz with optical incoherent heterodyne up-conversion using an optical frequency comb. Multiple users with wireline and wireless services are simultaneously supported.......We propose and experimentally demonstrate an optical wireless DWDM system at 60 GHz with optical incoherent heterodyne up-conversion using an optical frequency comb. Multiple users with wireline and wireless services are simultaneously supported.... 1. Fiber Optic Aircraft Systems Electromagnetic Pulse (Emp) Survivability Science.gov (United States) Gage, Byron; Greenwell, Roger; Summerlin, Michael; Zetlen, Bryan 1984-10-01 Mitigation of EMP coupling into sensitive, mission critical equipment is essential for Aircraft required to operate in adverse nuclear environments. As has been demonstrated in several aircraft test-fix-test programs, traditional hardening can eliminate most EMP problems but generally adds weight, volume, and complexity which impacts system reliability, maintainability and hardness surveillance. Fiber optic technology reduces weight, volume, and complexity while reducing overall life cycle costs and can also mitigate or eliminate many EMP related problems. As requirements for data transfer volume increase, aircraft system expansion utilizing present technology within extended design constraints is hampered by mission requirements for extensive EMI, RFI, EMP, lightning and short circuit shielding and protection. The criticality of excessive weight and space needed for shielding protection is well known and so are the problems of bent pins associated with filter pin connec-tors. The use of non-metallic composite structural materials for the aircraft skin further exacerbates the traditional shielding and filtering problems. The complete elimination of shielding and filtering is not possible. However, the use of fiber optics paths, complex penetrations and other intentional or inherent inadvertent conductors and thereby greatly simplifies EMP hardening. The inherent dielectric nature of fiber optics makes it relatively resistant or immune to the upset/damage potential of EMP. Fiber optic technology is also capable of electromagnetic interference and cross talk. The vulnerability of fiber optic technology to other significant factors in the operational environment, i.e., ionizing radiation, should also be examined and assessed. 2. Computationally efficient gradient matrix of optical path length in axisymmetric optical systems. Science.gov (United States) Hsueh, Chun-Che; Lin, Psang-Dain 2009-02-10 We develop a mathematical method for determining the optical path length (OPL) gradient matrix relative to all the system variables such that the effects of variable changes can be evaluated in a single pass. The approach developed avoids the requirement for multiple ray-tracing operations and is, therefore, more computationally efficient. By contrast, the effects of variable changes on the OPL of an optical system are generally evaluated by utilizing a ray-tracing approach to determine the OPL before and after the variable change and then applying a finite-difference (FD) approximation method to estimate the OPL gradient with respect to each individual variable. Utilizing a Petzval lens system for verification purposes, it is shown that the approach developed reduces the computational time by around 90% compared to that of the FD method. 3. New compact and efficient local oscillator optic system for the KSTAR electron cyclotron emission imaging system Energy Technology Data Exchange (ETDEWEB) Nam, Y. B., E-mail: southub@postech.ac.kr; Yun, G. S. [Department of Physics, Pohang University of Science and Technology, Pohang 37673 (Korea, Republic of); Lee, D. J.; Lee, J.; Lee, W. [Ulsan National Institute of Science and Technology, Ulsan 44919 (Korea, Republic of); Kim, C. [Pennsylvania State University, Old Main, State College, Pennsylvania 16801 (United States); Park, H. K. [Ulsan National Institute of Science and Technology, Ulsan 44919 (Korea, Republic of); National Fusion Research Institute, Daejeon 34133 (Korea, Republic of) 2016-11-15 Electron cyclotron emission imaging (ECEI) diagnostic on Korean Superconducting Tokamak Advanced Research utilizes quasi-optical heterodyne-detection method to measure 2D (vertical and radial) T{sub e} fluctuations from two toroidally separated poloidal cross section of the plasma. A cylindrical lens local oscillator (LO) optics with optical path length (OPL) 2–2.5 m has been used in the current ECEI system to couple the LO source to the 24 vertically aligned array of ECE detectors. For efficient and compact LO optics employing the Powell lens is proposed so that the OPL of the LO source is significantly reduced from ∼2.0 m to 0.4 m with new optics. The coupling efficiency of the LO source is expected to be improved especially at the edge channels. Results from the optical simulation together with the laboratory test of the prototype optics will be discussed in this paper. 4. An All-Fiber-Optic Combined System of Noncontact Photoacoustic Tomography and Optical Coherence Tomography Directory of Open Access Journals (Sweden) Jonghyun Eom 2016-05-01 Full Text Available We propose an all-fiber-based dual-modal imaging system that combines noncontact photoacoustic tomography (PAT and optical coherence tomography (OCT. The PAT remotely measures photoacoustic (PA signals with a 1550-nm laser on the surface of a sample by utilizing a fiber interferometer as an ultrasound detector. The fiber-based OCT, employing a swept-source laser centered at 1310 nm, shares the sample arm of the PAT system. The fiber-optic probe for the combined system was homemade with a lensed single-mode fiber (SMF and a large-core multimode fiber (MMF. The compact and robust common probe is capable of obtaining both the PA and the OCT signals at the same position without any physical contact. Additionally, the MMF of the probe delivers the short pulses of a Nd:YAG laser to efficiently excite the PA signals. We experimentally demonstrate the feasibility of the proposed dual-modal system with a phantom made of a fishing line and a black polyethylene terephthalate fiber in a tissue mimicking solution. The all-fiber-optic system, capable of providing complementary information about absorption and scattering, has a promising potential in minimally invasive and endoscopic imaging. 5. The Fiber Optic System for the Advanced Topographic Laser Altimeter System (ATLAS) Instrument Science.gov (United States) Ott, Melanie N.; Thomes, Joe; Onuma, Eleanya; Switzer, Robert; Chuska, Richard; Blair, Diana; Frese, Erich; Matyseck, Marc 2016-01-01 The Advanced Topographic Laser Altimeter System (ATLAS) Instrument has been in integration and testing over the past 18 months in preparation for the Ice, Cloud and Land Elevation Satellite - 2 (ICESat-2) Mission, scheduled to launch in 2017. ICESat-2 is the follow on to ICESat which launched in 2003 and operated until 2009. ATLAS will measure the elevation of ice sheets, glaciers and sea ice or the "cryosphere" (as well as terrain) to provide data for assessing the earth's global climate changes. Where ICESat's instrument, the Geo-Science Laser Altimeter (GLAS) used a single beam measured with a 70 m spot on the ground and a distance between spots of 170 m, ATLAS will measure a spot size of 10 m with a spacing of 70 cm using six beams to measure terrain height changes as small as 4 mm. The ATLAS pulsed transmission system consists of two lasers operating at 532 nm with transmitter optics for beam steering, a diffractive optical element that splits the signal into 6 separate beams, receivers for start pulse detection and a wavelength tracking system. The optical receiver telescope system consists of optics that focus all six beams into optical fibers that feed a filter system that transmits the signal via fiber assemblies to the detectors. Also included on the instrument is a system that calibrates the alignment of the transmitted pulses to the receiver optics for precise signal capture. The larger electro optical subsystems for transmission, calibration, and signal receive, stay aligned and transmitting sufficiently due to the optical fiber system that links them together. The robust design of the fiber optic system, consisting of a variety of multi fiber arrays and simplex assemblies with multiple fiber core sizes and types, will enable the system to maintain consistent critical alignments for the entire life of the mission. Some of the development approaches used to meet the challenging optical system requirements for ATLAS are discussed here. 6. EMC of Electrical Systems - Electromagnetic Coupling ( Part II Directory of Open Access Journals (Sweden) KOVACOVA, I. 2007-04-01 Full Text Available The paper deals with the general analysis of one part of the electromagnetic compatibility (EMC problem - the electromagnetic coupling applied in the field of power electrical systems. The verification simulation analyses and practical measurements of the electromagnetic coupling, which are confirming the correctness of results obtained from theoretical analyses (part I., are presented in part II. So they can be used for predictive stating of EMC quality of individual new electrotechnical products. 7. Airborne Optical Systems Test Bed (AOSTB) Science.gov (United States) 2016-07-01 the operator GUI 5 for onboard data processing, configured many of the computer systems, and helped perform months of testing of the sensor both...coated broadband mirror, and is directed by a periscope downwards and to the RSMA scan mirror. A safety shutter exists on a motorized flip mount...emergency. The shutter consists of a motorized flip mount with a flat mirror 31 that is out of the beam path for propagation, and in the beam path when 8. Temporal Probabilistic Constellation Shaping for WDM Optical Communication Systems DEFF Research Database (Denmark) Yankov, Metodi Plamenov; Forchhammer, Søren 2016-01-01 Finite state machine sources transmitting QPSK are studied as input to WDM optical fiber systems with ideal distributed Raman amplification. The probabilities of successive constellation symbols are shaped for nonlinear transmission and gains of around 500km (5-10%) are demonstrated... 9. the analysis of an optical fibre communication system using laser ... African Journals Online (AJOL) THE ANALYSIS OF AN OPTICAL FIBRE. COMMUNICATION SYSTEM USING LASER RATE. EQUATIONS. HN Kundaeli. Department of Physics, University of Dar-es-Salaam,. P. O. Box 35063, Dar es Salaam, Tanzania. ABSTRACT. Fibre dispersion has been the main limitation to high-speed data transmissions employing ... 10. Simulating channel losses in an underwater optical communication system. Science.gov (United States) Cox, William; Muth, John 2014-05-01 A Monte Carlo numerical simulation for computing the received power for an underwater optical communication system is discussed and validated. Power loss between receiver and transmitter is simulated for a variety of receiver aperture sizes and fields of view. Additionally, pointing-and-tracking losses are simulated. 11. LCLS-II high power RF system overview and progress Energy Technology Data Exchange (ETDEWEB) Yeremian, Anahid Dian 2015-10-07 A second X-ray free electron laser facility, LCLS-II, will be constructed at SLAC. LCLS-II is based on a 1.3 GHz, 4 GeV, continuous-wave (CW) superconducting linear accelerator, to be installed in the first kilometer of the SLAC tunnel. Multiple types of high power RF (HPRF) sources will be used to power different systems on LCLS-II. The main 1.3 GHz linac will be powered by 280 1.3 GHz, 3.8 kW solid state amplifier (SSA) sources. The normal conducting buncher in the injector will use four more SSAs identical to the linac SSAs but run at 2 kW. Two 185.7 MHz, 60 kW sources will power the photocathode dual-feed RF gun. A third harmonic linac section, included for linearizing the bunch energy spread before the first bunch compressor, will require sixteen 3.9 GHz sources at about 1 kW CW. A description and an update on all the HPRF sources of LCLS-II and their implementation is the subject of this paper. 12. PYROLASER - PYROLASER OPTICAL PYROMETER OPERATING SYSTEM Science.gov (United States) Roberts, F. E. 1994-01-01 The PYROLASER package is an operating system for the Pyrometer Instrument Company's Pyrolaser. There are 6 individual programs in the PYROLASER package: two main programs, two lower level subprograms, and two programs which, although independent, function predominantly as macros. The package provides a quick and easy way to setup, control, and program a standard Pyrolaser. Temperature and emissivity measurements may be either collected as if the Pyrolaser were in the manual operations mode, or displayed on real time strip charts and stored in standard spreadsheet format for post-test analysis. A shell is supplied to allow macros, which are test-specific, to be easily added to the system. The Pyrolaser Simple Operation program provides full on-screen remote operation capabilities, thus allowing the user to operate the Pyrolaser from the computer just as it would be operated manually. The Pyrolaser Simple Operation program also allows the use of "quick starts". Quick starts provide an easy way to permit routines to be used as setup macros for specific applications or tests. The specific procedures required for a test may be ordered in a sequence structure and then the sequence structure can be started with a simple button in the cluster structure provided. One quick start macro is provided for continuous Pyrolaser operation. A subprogram, Display Continuous Pyr Data, is used to display and store the resulting data output. Using this macro, the system is set up for continuous operation and the subprogram is called to display the data in real time on strip charts. The data is simultaneously stored in a spreadsheet format. The resulting spreadsheet file can be opened in any one of a number of commercially available spreadsheet programs. The Read Continuous Pyrometer program is provided as a continuously run subprogram for incorporation of the Pyrolaser software into a process control or feedback control scheme in a multi-component system. The program requires the 13. An adaptive optics imaging system designed for clinical use Science.gov (United States) Zhang, Jie; Yang, Qiang; Saito, Kenichi; Nozato, Koji; Williams, David R.; Rossi, Ethan A. 2015-01-01 Here we demonstrate a new imaging system that addresses several major problems limiting the clinical utility of conventional adaptive optics scanning light ophthalmoscopy (AOSLO), including its small field of view (FOV), reliance on patient fixation for targeting imaging, and substantial post-processing time. We previously showed an efficient image based eye tracking method for real-time optical stabilization and image registration in AOSLO. However, in patients with poor fixation, eye motion causes the FOV to drift substantially, causing this approach to fail. We solve that problem here by tracking eye motion at multiple spatial scales simultaneously by optically and electronically integrating a wide FOV SLO (WFSLO) with an AOSLO. This multi-scale approach, implemented with fast tip/tilt mirrors, has a large stabilization range of ± 5.6°. Our method consists of three stages implemented in parallel: 1) coarse optical stabilization driven by a WFSLO image, 2) fine optical stabilization driven by an AOSLO image, and 3) sub-pixel digital registration of the AOSLO image. We evaluated system performance in normal eyes and diseased eyes with poor fixation. Residual image motion with incremental compensation after each stage was: 1) ~2–3 arc minutes, (arcmin) 2) ~0.5–0.8 arcmin and, 3) ~0.05–0.07 arcmin, for normal eyes. Performance in eyes with poor fixation was: 1) ~3–5 arcmin, 2) ~0.7–1.1 arcmin and 3) ~0.07–0.14 arcmin. We demonstrate that this system is capable of reducing image motion by a factor of ~400, on average. This new optical design provides additional benefits for clinical imaging, including a steering subsystem for AOSLO that can be guided by the WFSLO to target specific regions of interest such as retinal pathology and real-time averaging of registered images to eliminate image post-processing. PMID:26114033 14. An adaptive optics imaging system designed for clinical use. Science.gov (United States) Zhang, Jie; Yang, Qiang; Saito, Kenichi; Nozato, Koji; Williams, David R; Rossi, Ethan A 2015-06-01 Here we demonstrate a new imaging system that addresses several major problems limiting the clinical utility of conventional adaptive optics scanning light ophthalmoscopy (AOSLO), including its small field of view (FOV), reliance on patient fixation for targeting imaging, and substantial post-processing time. We previously showed an efficient image based eye tracking method for real-time optical stabilization and image registration in AOSLO. However, in patients with poor fixation, eye motion causes the FOV to drift substantially, causing this approach to fail. We solve that problem here by tracking eye motion at multiple spatial scales simultaneously by optically and electronically integrating a wide FOV SLO (WFSLO) with an AOSLO. This multi-scale approach, implemented with fast tip/tilt mirrors, has a large stabilization range of ± 5.6°. Our method consists of three stages implemented in parallel: 1) coarse optical stabilization driven by a WFSLO image, 2) fine optical stabilization driven by an AOSLO image, and 3) sub-pixel digital registration of the AOSLO image. We evaluated system performance in normal eyes and diseased eyes with poor fixation. Residual image motion with incremental compensation after each stage was: 1) ~2-3 arc minutes, (arcmin) 2) ~0.5-0.8 arcmin and, 3) ~0.05-0.07 arcmin, for normal eyes. Performance in eyes with poor fixation was: 1) ~3-5 arcmin, 2) ~0.7-1.1 arcmin and 3) ~0.07-0.14 arcmin. We demonstrate that this system is capable of reducing image motion by a factor of ~400, on average. This new optical design provides additional benefits for clinical imaging, including a steering subsystem for AOSLO that can be guided by the WFSLO to target specific regions of interest such as retinal pathology and real-time averaging of registered images to eliminate image post-processing. 15. Observation of room temperature optical absorption in InP/GaAs type-II ultrathin quantum wells and quantum dots Energy Technology Data Exchange (ETDEWEB) Singh, S. D., E-mail: devsh@rrcat.gov.in; Porwal, S.; Mondal, Puspen; Srivastava, A. K.; Mukherjee, C.; Dixit, V. K.; Sharma, T. K.; Oak, S. M. [Raja Ramanna Centre for Advanced Technology, Indore-452013, Madhya Pradesh (India) 2014-06-14 Room temperature optical absorption process is observed in ultrathin quantum wells (QWs) and quantum dots (QDs) of InP/GaAs type-II band alignment system using surface photovoltage spectroscopy technique, where no measurable photoluminescence signal is available. Clear signature of absorption edge in the sub band gap region of GaAs barrier layer is observed for the ultrathin QWs and QDs, which red shifts with the amount of deposited InP material. Movement of photogenerated holes towards the sample surface is proposed to be the main mechanism for the generation of surface photovoltage in type-II ultrathin QWs and QDs. QDs of smaller size are found to be free from the dislocations as confirmed by the high resolution transmission electron microscopy images. 16. Fourier transform digital holographic adaptive optics imaging system Science.gov (United States) Liu, Changgeng; Yu, Xiao; Kim, Myung K. 2013-01-01 A Fourier transform digital holographic adaptive optics imaging system and its basic principles are proposed. The CCD is put at the exact Fourier transform plane of the pupil of the eye lens. The spherical curvature introduced by the optics except the eye lens itself is eliminated. The CCD is also at image plane of the target. The point-spread function of the system is directly recorded, making it easier to determine the correct guide-star hologram. Also, the light signal will be stronger at the CCD, especially for phase-aberration sensing. Numerical propagation is avoided. The sensor aperture has nothing to do with the resolution and the possibility of using low coherence or incoherent illumination is opened. The system becomes more efficient and flexible. Although it is intended for ophthalmic use, it also shows potential application in microscopy. The robustness and feasibility of this compact system are demonstrated by simulations and experiments using scattering objects. PMID:23262541 17. An Antenna Measurement System Based on Optical Feeding Directory of Open Access Journals (Sweden) Ryohei Hosono 2013-01-01 the advantage of the system is demonstrated by measuring an ultra-wideband (UWB antenna both by the optical and electrical feeding systems and comparing with a calculated result. Ripples in radiation pattern due to the electrical feeding are successfully suppressed by the optical feeding. For example, in a radiation measurement on the azimuth plane at 3 GHz, ripple amplitude of 1.0 dB that appeared in the electrical feeding is reduced to 0.3 dB. In addition, a circularly polarized (CP antenna is successfully measured by the proposed system to show that the system is available not only for amplitude but also phase measurements. 18. Ring resonator systems to perform optical communication enhancement using soliton CERN Document Server 2014-01-01 The title explain new technique of secured and high capacity optical communication signals generation by using the micro and nano ring resonators. The pulses are known as soliton pulses which are more secured due to having the properties of chaotic and dark soliton signals with ultra short bandwidth. They have high capacity due to the fact that ring resonators are able to generate pulses in the form of solitons in multiples and train form. These pulses generated by ring resonators are suitable in optical communication due to use the compact and integrated rings system, easy to control, flexibi 19. Optical fiber communication systems with Matlab and Simulink models CERN Document Server Binh, Le Nguyen 2014-01-01 ""This book adds an aspect of programming and simulation not so well developed in other books. It is complete in this sense and enables directly linking the physics of optical components and systems to realistic results.""-Martin Rochette, Associate Professor, McGill University, Quebec, Canada""…this will be an excellent textbook since it has all new development and information on optical communication systems…I think this book can easily replace many other textbooks in this field.""-Massoud Moussavi, California State Polytechnic University-Pomona""The book is well written. It describes the fu 20. A Compact Prototype of an Optical Pattern Recognition System Science.gov (United States) Jin, Y.; Liu, H. K.; Marzwell, N. I. 1996-01-01 In the Technology 2006 Case Studies/Success Stories presentation, we will describe and demonstrate a prototype of a compact optical pattern recognition system as an example of a successful technology transfer and continuuing development of state-of-the-art know-how by the close collaboration among government, academia, and small business via the NASA SBIR program. The prototype consists of a complete set of optical pattern recognition hardware with multi-channel storage and retrieval capability that is compactly configured inside a portable 1'X 2'X 3' aluminum case. 1. Design layout for gas monitoring system II (GMS-2) computer system Energy Technology Data Exchange (ETDEWEB) Vo, V.; Philipp, B.L.; Manke, M.P. 1995-08-02 This document provides a general overview of the computer systems software that perform the data acquisition and control for the 241-SY-101 Gas Monitoring System II (GMS-2). It outlines the system layout, and contains descriptions of components and the functions they perform. The GMS-2 system was designed and implemented by Los Alamos National Laboratory and supplied to Westinghouse Hanford Company 2. Integrated optical measurement system for fluorescence spectroscopy in microfluidic channels DEFF Research Database (Denmark) Hübner, Jörg; Mogensen, Klaus Bo; Jørgensen, Anders Michael 2001-01-01 A transportable miniaturized fiber-pigtailed measurement system is presented which allows quantitative fluorescence detection in microliquid handling systems. The microliquid handling chips are made in silica on silicon technology and the optical functionality is monolithically integrated with th...... with two dyes, fluorescein, and Bodipy 650/665 X, showed good linear behavior over a wide range of concentrations. Minimally detected concentrations were 250 pM for fluorescein and 100 nM for Bodipy.... 3. T-6A Texan II Systems Engineering Case Study Science.gov (United States) 2010-01-01 System for the 21st Century: JPATS and the T-6, Richard H Emmons , June 2004, Page 1 11 Ibid, Page 2 T-6A Texan II Engineering Case Study Air Force...12 Also know as Service Life Extension Program 13 A Training System for the 21st Century: JPATS and the T-6, Richard H Emmons , June 2004, Page 3 14...PATS program. This team consisted of Maj Gen Robert Delligatti (ATC Vice Commander); Capt LynnAnne Merten (ATC PATS Acquisition Manager); and Capt 4. Integrated intravascular optical coherence tomography (OCT) - ultrasound (US) imaging system Science.gov (United States) Yin, Jiechen; Yang, Hao-Chung; Li, Xiang; Zhou, Qifa; Hu, Changhong; Zhang, Jun; Shung, K. Kirk; Chen, Zhongping 2010-02-01 Optical coherence tomography (OCT) and intravascular ultrasound (IVUS) are considered two complementary imaging techniques in the detection and diagnosis of atherosclerosis. OCT permits visualization of micron-scale features of atherosclerosis plaque, and IVUS offers full imaging depth of vessel wall. Under the guidance of IVUS, minimal amount of flushing agent will be needed to obtain OCT imaging of the interested area. We report on a dual-modality optical coherence tomography (OCT) - ultrasound (US) system for intravascular imaging. To the best of our knowledge, we have developed the first integrated OCT-US probe that combines OCT optical components with an ultrasound transducer. The OCT optical components mainly consist of a single mode fiber, a gradient index (GRIN) lens for light beam focusing, and a right-angled prism for reflecting light into biological tissue. A 40MHz PZT-5H side-viewing ultrasound transducer was fabricated to obtain the ultrasound image. These components were integrated into a single probe, enabling both OCT and ultrasound imaging at the same time. In vitro OCT and ultrasound images of a rabbit aorta were obtained using this dual-modality imaging system. This study demonstrates the feasibility of an OCT-US system for intravascular imaging which is expected to have a prominent impact on early detection and characterization of atherosclerosis. 5. Temperature control system for optical elements in astronomical instrumentation Science.gov (United States) Verducci, Orlando; de Oliveira, Antonio C.; Ribeiro, Flávio F.; Vital de Arruda, Márcio; Gneiding, Clemens D.; Fraga, Luciano 2014-07-01 Extremely low temperatures may damage the optical components assembled inside of an astronomical instrument due to the crack in the resin or glue used to attach lenses and mirrors. The environment, very cold and dry, in most of the astronomical observatories contributes to this problem. This paper describes the solution implemented at SOAR for remotely monitoring and controlling temperatures inside of a spectrograph, in order to prevent a possible damage of the optical parts. The system automatically switches on and off some heat dissipation elements, located near the optics, as the measured temperature reaches a trigger value. This value is set to a temperature at which the instrument is not operational to prevent malfunction and only to protect the optics. The software was developed with LabVIEWTM and based on an object-oriented design that offers flexibility and ease of maintenance. As result, the system is able to keep the internal temperature of the instrument above a chosen limit, except perhaps during the response time, due to inertia of the temperature. This inertia can be controlled and even avoided by choosing the correct amount of heat dissipation and location of the thermal elements. A log file records the measured temperature values by the system for operation analysis. 6. Optical polarimetric and near-infrared photometric study of the RCW95 Galactic H II region Science.gov (United States) Vargas-González, J.; Roman-Lopes, A.; Santos, F. P.; Franco, G. A. P.; Santos, J. F. C.; Maia, F. F. S.; Sanmartim, D. 2018-02-01 We carried out an optical polarimetric study in the direction of the RCW 95 star-forming region in order to probe the sky-projected magnetic field structure by using the distribution of linear polarization segments which seem to be well aligned with the more extended cloud component. A mean polarization angle of θ = 49.8° ± 7.7°7 was derived. Through the spectral dependence analysis of polarization it was possible to obtain the total-to-selective extinction ratio (RV) by fitting the Serkowski function, resulting in a mean value of RV = 2.93 ± 0.47. The foreground polarization component was estimated and is in agreement with previous studies in this direction of the Galaxy. Further, near-infrared (NIR) images from Vista Variables in the Via Láctea (VVV) survey were collected to improve the study of the stellar population associated with the H II region. The Automated Stellar Cluster Analysis algorithm was employed to derive structural parameters for two clusters in the region, and a set of PAdova and TRieste Stellar Evolution Code (PARSEC) isochrones was superimposed on the decontaminated colour-magnitude diagrams to estimate an age of about 3 Myr for both clusters. Finally, from the NIR photometry study combined with spectra obtained with the Ohio State Infrared Imager and Spectrometer mounted at the Southern Astrophysics Research Telescope we derived the spectral classification of the main ionizing sources in the clusters associated with IRAS 15408-5356 and IRAS 15412-5359, both objects classified as O4V stars. 7. Investigations in silicate glasses. I. Radiation damage. II. Optical nonlinearity. [Gamma rays and electrons Energy Technology Data Exchange (ETDEWEB) Moran, M.J. 1976-11-15 The investigation of two poorly understood but technologically important physical properties of silicate glasses and related materials is described. The use of Electron Paramagnetic Resonance to investigate the nature of radiation-induced damage in glasses exposed to a variety of high-energy radiation sources is discussed first. Second, the measurement of the nonlinear index of refraction coefficient in a variety of optical materials related to the design of high-power laser systems is described. The radiation damage investigations rely heavily on the comparison of experimental results for different experimental situations. The comparison of EPR lineshapes, absolute spin densities and power saturation behavior is used to probe a variety of microscopic and macroscopic aspects of radiation damage in glasses. Comparison of radiation damage associated with exposure to gamma rays and fast neutrons (and combinations thereof) are interpreted in terms of the microscopic damage mechanisms which are expected to be associated with the specific radiations. Comparison of radiation damage behavior in different types of glasses is also interpreted in terms of the behavior expected for the specific materials. The body of data which is generated is found to be internally self-consistent and is also generally consistent with the radiation damage behavior expected for specific situations. A new and versatile technique for measuring the nonlinear index of refraction coefficient, n/sub 2/, in optical materials is described. The technique utilizes a 1 ns pulsed neodymium-glass laser system and time-resolved interferometry to determine the ratio of the coefficient n/sub 2/ of sample materials to the n/sub 2/ of CS/sub 2/. This method avoids some of the complications associated with performing absolute measurements of n/sub 2/ and allows the use of a relatively simple experimental technique. The measurements determine the nonlinear index ratios of the samples with an accuracy of about 8. Joint digital signal processing for superchannel coherent optical communication systems. Science.gov (United States) Liu, Cheng; Pan, Jie; Detwiler, Thomas; Stark, Andrew; Hsueh, Yu-Ting; Chang, Gee-Kung; Ralph, Stephen E 2013-04-08 Ultra-high-speed optical communication systems which can support ≥ 1Tb/s per channel transmission will soon be required to meet the increasing capacity demand. However, 1Tb/s over a single carrier requires either or both a high-level modulation format (i.e. 1024QAM) and a high baud rate. Alternatively, grouping a number of tightly spaced "sub-carriers" to form a terabit superchannel increases channel capacity while minimizing the need for high-level modulation formats and high baud rate, which may allow existing formats, baud rate and components to be exploited. In ideal Nyquist-WDM superchannel systems, optical subcarriers with rectangular spectra are tightly packed at a channel spacing equal to the baud rate, thus achieving the Nyquist bandwidth limit. However, in practical Nyquist-WDM systems, precise electrical or optical control of channel spectra is required to avoid strong inter-channel interference (ICI). Here, we propose and demonstrate a new "super receiver" architecture for practical Nyquist-WDM systems, which jointly detects and demodulates multiple channels simultaneously and mitigates the penalties associated with the limitations of generating ideal Nyquist-WDM spectra. Our receiver-side solution relaxes the filter requirements imposed on the transmitter. Two joint DSP algorithms are developed for linear ICI cancellation and joint carrier-phase recovery. Improved system performance is observed with both experimental and simulation data. Performance analysis under different system configurations is conducted to demonstrate the feasibility and robustness of the proposed joint DSP algorithms. 9. Research on solar-blind UV optical imaging system Science.gov (United States) Wang, Baohua; Wang, Yuanyuan; Zhong, Xiaoming; Ruan, Ningjuan 2015-02-01 Solar blind UV detecting system has many advantages such as strong environmental adaptability, low error rate, small volume and without refrigeration. To in-depth develop UV solar blind detection system research work has important significance for further improving solar blind UV detection technology. The working principle of solar blind UV detection system and the basic components were introduced firstly, and then the key technology of solar blind UV detection system was deeply analyzed. Finally, large coverage solar blind UV optical imaging system was designed according to the actual demand for greater coverage of the solar blind UV detection system. The result shows that the system has good imaging quality, simple and compact structure. This system can be used in various types of solar blind UV detection system, and is of high application value. 10. Predictive value of the APACHE II, SAPS II, SOFA and GCS scoring systems in patients with severe purulent bacterial meningitis. Science.gov (United States) Pietraszek-Grzywaczewska, Iwona; Bernas, Szymon; Łojko, Piotr; Piechota, Anna; Piechota, Mariusz 2016-01-01 Scoring systems in critical care patients are essential for predicting of the patient outcome and evaluating the therapy. In this study, we determined the value of the Acute Physiology and Chronic Health Evaluation II (APACHE II), Simplified Acute Physiology Score II (SAPS II), Sequential Organ Failure Assessment (SOFA) and Glasgow Coma Scale (GCS) scoring systems in the prediction of mortality in adult patients admitted to the intensive care unit (ICU) with severe purulent bacterial meningitis. We retrospectively analysed data from 98 adult patients with severe purulent bacterial meningitis who were admitted to the single ICU between March 2006 and September 2015. Univariate logistic regression identified the following risk factors of death in patients with severe purulent bacterial meningitis: APACHE II, SAPS II, SOFA, and GCS scores, and the lengths of ICU stay and hospital stay. The independent risk factors of patient death in multivariate analysis were the SAPS II score, the length of ICU stay and the length of hospital stay. In the prediction of mortality according to the area under the curve, the SAPS II score had the highest accuracy followed by the APACHE II, GCS and SOFA scores. For the prediction of mortality in a patient with severe purulent bacterial meningitis, SAPS II had the highest accuracy. 11. Urban Integrated Industrial Cogeneration Systems Analysis. Phase II final report Energy Technology Data Exchange (ETDEWEB) 1984-01-01 Through the Urban Integrated Industrial Cogeneration Systems Analysis (UIICSA), the City of Chicago embarked upon an ambitious effort to identify the measure the overall industrial cogeneration market in the city and to evaluate in detail the most promising market opportunities. This report discusses the background of the work completed during Phase II of the UIICSA and presents the results of economic feasibility studies conducted for three potential cogeneration sites in Chicago. Phase II focused on the feasibility of cogeneration at the three most promising sites: the Stockyards and Calumet industrial areas, and the Ford City commercial/industrial complex. Each feasibility case study considered the energy load requirements of the existing facilities at the site and the potential for attracting and serving new growth in the area. Alternative fuels and technologies, and ownership and financing options were also incorporated into the case studies. Finally, site specific considerations such as development incentives, zoning and building code restrictions and environmental requirements were investigated. 12. Progress in Nano-Electro-Optics VII Chemical, Biological, and Nanophotonic Technologies for Nano-Optical Devices and Systems CERN Document Server Ohtsu, Motoichi 2010-01-01 This book focuses on chemical and nanophotonic technology to be used to develop novel nano-optical devices and systems. It begins with temperature- and photo-induced phase transition of ferromagnetic materials. Further topics include: energy transfer in artificial photosynthesis, homoepitaxial multiple quantum wells in ZnO, near-field photochemical etching and nanophotonic devices based on a nonadiabatic process and optical near-field energy transfer, respectively and polarization control in the optical near-field for optical information security. Taken as a whole, this overview will be a valuable resource for engineers and scientists working in the field of nano-electro-optics. 13. New multicolor illumination system for automatic optical inspection Science.gov (United States) Xiong, Guangjie; Ma, Shuyuan; Nie, Xuejun; Tang, Xiaohua 2010-10-01 In automatic optical inspection (AOI), the illumination system affects the quality of input images and the result of image processing in the AOI. This paper developed a new multi-color illumination system specially used in the printed circuit board (PCB) inspection to detect a variety of defects in automated optical inspection system. The new illumination system consists of four kinds of colors of light emitting diode (LED) arrays composed of high-density LED surface light source. In order to detect a variety of defects, the radiation angle of the each LED array is different. The system uses a micro-controller to control the four sets of LED arrays, after acquisition of the image, which can self-adjust the light intensity of the illumination system based on the reference and comparison of histogram of the image in real time and can control different color LED array respectively according to the quality of the tested image. This paper analyzed the structural model of the illumination system and designed the control system. The experimental results show that the new illumination system has important performances such as uniform illumination, adjustable light intensity, fast response, lower heat and etc. The system can provide highly stable illumination for the AOI to obtain high-quality images effectively for detect the defects of PCB, and improve the defect detection rate and reduce the defects of the false alarm rate of AOI. 14. Design of VisSWIR continuous zoom optical system Science.gov (United States) Yang, Mingyang; Yang, Hongtao; Qu, Rui; Mei, Chao; Zhou, Zuofeng 2017-02-01 For 640 pixel×512 pixel cooled staring focal plane array detector, a VisSWIR wideband continuous zoom optical system with 7X zoom range is presented based on the pattern of the negative zoom group and compensating lens group. The zoom system provides continuous changed in the field of view from narrow to the wide. The zoom optical system works in the range of 0.4μm 1.7μm, F number is 4, the pixel of the detector is 15μm. It realizes 20mm 140mm continuous zoom with a smooth zoom path and provided high image quality with the whole zoom range, the zoom ratio is 7:1. The modulation transfer function(MTF) for the system is above 0.5 within the whole focal length range at spatial frequency of 34lp/mm and it almost approaches the diffraction limit. RMS value of spot diameter was investigation, the maximum distortion value is less than 5% and the surface type of all lens applied is spherical. Moreover, the cam curve after optimization is given by the optical design software Code V macro. The design results provide that the zoom system has the small size, high resolution, excellent image quality and the smooth cam curve etc. 15. Fiber-optic perimeter security system based on WDM technology Science.gov (United States) Polyakov, Alexandre V. 2017-10-01 Intelligent underground fiber optic perimeter security system is presented. Their structure, operation, software and hardware with neural networks elements are described. System allows not only to establish the fact of violation of the perimeter, but also to locate violations. This is achieved through the use of WDM-technology division spectral information channels. As used quasi-distributed optoelectronic recirculation system as a discrete sensor. The principle of operation is based on registration of the recirculation period change in the closed optoelectronic circuit at different wavelengths under microstrain exposed optical fiber. As a result microstrain fiber having additional power loss in a fiber optical propagating pulse, which causes a time delay as a result of switching moments of the threshold device. To separate the signals generated by intruder noise and interference, the signal analyzer is used, based on the principle of a neural network. The system detects walking, running or crawling intruder, as well as undermining attempts to register under the perimeter line. These alarm systems can be used to protect the perimeters of facilities such as airports, nuclear reactors, power plants, warehouses, and other extended territory. 16. Fiber optic perimeter system for security in smart city Science.gov (United States) Cubik, Jakub; Kepak, Stanislav; Nedoma, Jan; Fajkus, Marcel; Zboril, Ondrej; Novak, Martin; Jargus, Jan; Vasinek, Vladimir 2017-10-01 Protection of persons and assets is the key challenge of Smart City safeguards technologies. Conventional security technologies are often outdated and easy to breach. Therefore, new technologies that could complement existing systems or replace them are developed. The use of optical fibers and their subsequent application in sensing is a trend of recent years. This article discusses the use of fiber-optic sensors in perimeter protection. The sensor consists of optical fibers and couplers only and being constructed without wires and metal parts bring many advantages. These include an absence of interference with electromagnetic waves, system presence can be difficult to detect as well as affect its operation. Testing installation of perimeter system was carried out under reinforced concrete structure. Subjects walked over the bridge at different speeds and over the different routes. The task for the system was an absolute detection of all subjects. The proposed system should find application mainly in areas with the presence of volatile substances, strong electromagnetic fields, or in explosive areas. 17. Microfocus/Polycapillary-Optic Crystallographic X-Ray System Science.gov (United States) Joy, Marshall; Gubarev, Mikhail; Ciszak, Ewa 2005-01-01 A system that generates an intense, nearly collimated, nearly monochromatic, small-diameter x-ray beam has been developed for use in macromolecular crystallography. A conventional x-ray system for macromolecular crystallography includes a rotating-anode x-ray source, which is massive (.500 kg), large (approximately 2 by 2 by 1 m), and power-hungry (between 2 and 18 kW). In contrast, the present system generates a beam of the required brightness from a microfocus source, which is small and light enough to be mounted on a laboratory bench, and operates at a power level of only tens of watts. The figure schematically depicts the system as configured for observing x-ray diffraction from a macromolecular crystal. In addition to the microfocus x-ray source, the system includes a polycapillary optic . a monolithic block (typically a bundle of fused glass tubes) that contains thousands of straight or gently curved capillary channels, along which x-rays propagate with multiple reflections. This particular polycapillary optic is configured to act as a collimator; the x-ray beam that emerges from its output face consists of quasi-parallel subbeams with a small angular divergence and a diameter comparable to the size of a crystal to be studied. The gap between the microfocus x-ray source and the input face of the polycapillary optic is chosen consistently with the focal length of the polycapillary optic and the need to maximize the solid angle subtended by the optic in order to maximize the collimated x-ray flux. The spectrum from the source contains a significant component of Cu K (photon energy is 8.08 keV) radiation. The beam is monochromatized (for Cu K ) by a nickel filter 10 m thick. In a test, this system was operated at a power of 40 W (current of 897 A at an accelerating potential of 45 kV), with an anode x-ray spot size of 41+/-2 microns. Also tested, in order to provide a standard for comparison, was a commercial rotating-anode x-ray crystallographic system with a 18. Biologically active Schiff bases containing thiophene/furan ring and their copper(II) complexes: Synthesis, spectral, nonlinear optical and density functional studies Science.gov (United States) Gündüzalp, Ayla Balaban; Özsen, İffet; Alyar, Hamit; Alyar, Saliha; Özbek, Neslihan 2016-09-01 Schiff bases; 1,8-bis(thiophene-2-carboxaldimine)-p-menthane (L1) and 1,8-bis(furan-2-carboxaldimine)-p-menthane (L2) have been synthesized and characterized by elemental analysis, 1Hsbnd 13C NMR, UV-vis, FT-IR and LC-MS methods. 1H and 13C shielding tensors for L1 and L2 were calculated with GIAO/DFT/B3LYP/6-311++G(d,p) methods in CDCl3. The vibrational band assignments, nonlinear optical (NLO) activities, frontier molecular orbitals (FMOs) and absorption spectrum have been investigated by the same basis set. Schiff base-copper(II) complexes have been synthesized and structurally characterized with spectroscopic methods, magnetic and conductivity measurements. The spectroscopic data suggest that Schiff base ligands coordinate through azomethine-N and thiophene-S/furan-O donors (as SNNS and ONNO chelating systems) to give a tetragonal geometry around the copper(II) ions. Schiff bases and Cu(II) complexes have been screened for their biological activities on different species of pathogenic bacteria, those are, Gram positive bacteria: Bacillus subtitilus, Yersinia enterotica, Bacillus cereus, Listeria monocytogenes, Micrococcus luteus and Gram negative bacteria: Escherichia coli, Pseudomonas aeroginosa, Shigella dysenteriae, Salmonella typhi, Klebsiella pseudomonas by using microdilution technique (MIC values in mM). Biological activity results show that Cu(II) complexes have higher activities than parent ligands and metal chelation may affect significantly the antibacterial behavior of the organic ligands. 19. Pressure dependent optical gain in Type-II PbSe/PbS core /shell quantum dots Science.gov (United States) Saravanamoorthy, S. N.; John Peter, A.; Lee, Chang Woo 2017-11-01 Pressure induced nonlinear optical properties of Type-II lead based core/shell quantum dot heterostructure are investigated. The pressure dependent exciton binding energies are obtained with the increase of shell thickness for various inner core radii. The exciton interaction energies in a PbSe/PbS core/shell quantum dot inner core radii are found in the presence of hydrostatic pressure using variational method within the single band effective mass approximation. The threshold current density with the hydrostatic pressure and inner dot radii in a PbSe/PbS dot with the constant shell width is obtained. The pressure dependence on peak optical gain with the current density in a PbSe/PbS core/shell quantum dot with the fixed carrier density is found. The pressure related optical gain with the photon energy in a PbSe/PbS core/shell quantum dot is also obtained. The results show that the threshold current density increases by 20% when the pressure increases from 0 GPa to 4 GPa for 20 Å quantum dot radius whereas 16% of increase in threshold current density is observed for 50 Å quantum dot radius for the same increase of pressure values. The peak of the optical gain is observed to increase nonlinearly with the applied hydrostatic pressure and 12% of optical gain is enhanced for every 2 GPa pressure value. The results will be useful for some potential applications in near infrared light sources. 20. High pressure effect on optical gain in type-II InGaAs/GaAsSb nano-heterostructure Energy Technology Data Exchange (ETDEWEB) Yadav, Nisha; Nirmal, H. K.; Yadav, Rashmi; Lal, Pyare; Alvi, P. A., E-mail: drpaalvi@gmail.com [Department of Physics, Banasthali University, Banasthali Vidyapith-304022, Rajasthan (India) 2016-03-09 This paper reports the simulation of optical gain in type-II InGaAs/GaAsSb quantum well based nano-scale heterostructure. In order to simulate the optical gain, the heterostructure has been modeled with the help of six band k.p method. The 6 × 6 diagonalized k.p Hamiltonian has been solved to evaluate the valence sub-bands (i.e. light and heavy hole energies); and then optical matrix elements and optical gain within TE (Transverse Electric) mode has been calculated. The results obtained suggest that peak optical gain of the order of ∼ 9000 /cm in the heterostructure can be achieved at the lasing wavelength ∼ 1.95 µm (SWIR region). The application of high pressure (2 and 5 GPa) on the structure shows that the gain as well as lasing wavelength both approach to higher values. Thus, the structure can be tuned externally by the application of high pressure. 1. Optical diagnostics based on elastic scattering: Recent clinical demonstrations with the Los Alamos Optical Biopsy System Energy Technology Data Exchange (ETDEWEB) Bigio, I.J.; Loree, T.R.; Mourant, J.; Shimada, T. [Los Alamos National Lab., NM (United States); Story-Held, K.; Glickman, R.D. [Texas Univ. Health Science Center, San Antonio, TX (United States). Dept. of Ophthalmology; Conn, R. [Lovelace Medical Center, Albuquerque, NM (United States). Dept. of Urology 1993-08-01 A non-invasive diagnostic tool that could identify malignancy in situ and in real time would have a major impact on the detection and treatment of cancer. We have developed and are testing early prototypes of an optical biopsy system (OBS) for detection of cancer and other tissue pathologies. The OBS invokes a unique approach to optical diagnosis of tissue pathologies based on the elastic scattering properties, over a wide range of wavelengths, of the microscopic structure of the tissue. The use of elastic scattering as the key to optical tissue diagnostics in the OBS is based on the fact that many tissue pathologies, including a majority of cancer forms, manifest significant architectural changes at the cellular and sub-cellular level. Since the cellular components that cause elastic scattering have dimensions typically on the order of visible to near-IR wavelengths, the elastic (Mie) scattering properties will be strongly wavelength dependent. Thus, morphology and size changes can be expected to cause significant changes in an optical signature that is derived from the wavelength dependence of elastic scattering. The data acquisition and storage/display time with the OBS instrument is {approximately}1 second. Thus, in addition to the reduced invasiveness of this technique compared with current state-of-the-art methods (surgical biopsy and pathology analysis), the OBS offers the possibility of impressively faster diagnostic assessment. The OBS employs a small fiber-optic probe that is amenable to use with any endoscope, catheter or hypodermic, or to direct surface examination (e.g. as in skin cancer or cervical cancer). It has been tested in vitro on animal and human tissue samples, and clinical testing in vivo is currently in progress. 2. An Optical Tomography System Using a Digital Signal Processor. Science.gov (United States) Rahim, Ruzairi Abdul; Thiam, Chiam Kok; Rahiman, Mohd Hafiz Fazalul 2008-03-27 The use of a personal computer together with a Data Acquisition System (DAQ) as the processing tool in optical tomography systems has been the norm ever since the beginning of process tomography. However, advancements in silicon fabrication technology allow nowadays the fabrication of powerful Digital Signal Processors (DSP) at a reasonable cost. This allows this technology to be used in an optical tomography system since data acquisition and processing can be performed within the DSP. Thus, the dependency on a personal computer and a DAQ to sample and process the external signals can be reduced or even eliminated. The DSP system was customized to control the data acquisition process of 16x16 optical sensor array, arranged in parallel beam projection. The data collected was used to reconstruct the cross sectional image of the pipeline conveyor. For image display purposes, the reconstructed image was sent to a personal computer via serial communication. This allows the use of a laptop to display the tomogram image besides performing any other offline analysis. 3. Measuring the spatial resolution of an optical system in an undergraduate optics laboratory Science.gov (United States) Leung, Calvin; Donnelly, T. D. 2017-06-01 Two methods of quantifying the spatial resolution of a camera are described, performed, and compared, with the objective of designing an imaging-system experiment for students in an undergraduate optics laboratory. With the goal of characterizing the resolution of a typical digital single-lens reflex (DSLR) camera, we motivate, introduce, and show agreement between traditional test-target contrast measurements and the technique of using Fourier analysis to obtain the modulation transfer function (MTF). The advantages and drawbacks of each method are compared. Finally, we explore the rich optical physics at work in the camera system by calculating the MTF as a function of wavelength and f-number. For example, we find that the Canon 40D demonstrates better spatial resolution at short wavelengths, in accordance with scalar diffraction theory, but is not diffraction-limited, being significantly affected by spherical aberration. The experiment and data analysis routines described here can be built and written in an undergraduate optics lab setting. 4. Optical communication with two-photon coherent states. II - Photoemissive detection and structured receiver performance Science.gov (United States) Shapiro, J. H.; Yuen, H. P.; Machado Mata, J. A. 1979-01-01 In a previous paper (1978), the authors developed a method of analyzing the performance of two-photon coherent state (TCS) systems for free-space optical communications. General theorems permitting application of classical point process results to detection and estimation of signals in arbitrary quantum states were derived. The present paper examines the general problem of photoemissive detection statistics. On the basis of the photocounting theory of Kelley and Kleiner (1964) it is shown that for arbitrary pure state illumination, the resulting photocurrent is in general a self-exciting point process. The photocount statistics for first-order coherent fields reduce to those of a special class of Markov birth processes, which the authors term single-mode birth processes. These general results are applied to the structure of TCS radiation, and it is shown that the use of TCS radiation with direct or heterodyne detection results in minimal performance increments over comparable coherent-state systems. However, significant performance advantages are offered by use of TCS radiation with homodyne detection. The abstract quantum descriptions of homodyne and heterodyne detection are derived and a synthesis procedure for obtaining quantum measurements described by arbitrary TCS is given. 5. New optical sensor systems for photogrammetry and remote sensing Science.gov (United States) Eckardt, Andreas; Börner, Anko; Jahn, Herbert; Hilbert, Stefan; Walter, Ingo 2006-09-01 Recent developments in the fields of detectors on one hand and a significant change of national and international political and commercial constraints on the other hand led to a large number of proposals for spaceborne sensor systems focusing on Earth observation in the last months. Due to the commercial availability of TDI lines and fast readable CCD-Chips new sensor concepts are feasible for high resolution sensor systems regarding geometry and radiometry und their data products. Systemic approaches are essential for the design of complex sensor systems for dedicated tasks. Starting with system theory optically, mechanical and electrical components are designed and deployed. Single modules and the entire system have to be calibrated using suitable procedures. The paper gives an overview about current activities at German Aerospace Center on the field of innovative sensor systems for photogrammetry and remote sensing. 6. Optics designs and system MTF for laser scanning displays Science.gov (United States) Urey, Hakan; Nestorovic, Ned; Ng, Baldwin S.; Gross, Abraham A. 1999-07-01 The Virtual Retinal DisplayTM (VRDTM) technology is a new display technology being developed at Microvision Inc. The displayed image is scanned onto the viewer's retina using low- power red, green, and blue light sources. Microvision's proprietary miniaturized scanner designs make VRD system very well suited for head-mounted displays. In this paper we discuss some of the advantages of the VRD technology, various ocular designs for HMD and other applications, and details of constructing a system MTF budget for laser scanning systems that includes electronics, modulators, scanners, and optics. 7. Competitive adsorption of Cd(II), Zn(II) and Ni(II) from their binary and ternary acidic systems using tourmaline. Science.gov (United States) Liu, Haibin; Wang, Cuiping; Liu, Jingting; Wang, Baolin; Sun, Hongwen 2013-10-15 8. 46 CFR 128.220 - Class II non-vital systems-materials and pressure design. Science.gov (United States) 2010-10-01 ... schedule in wall thickness. (c) Each Class II non-vital piping-system must be certified by the builder as... 46 Shipping 4 2010-10-01 2010-10-01 false Class II non-vital systems-materials and pressure design... SUPPLY VESSELS MARINE ENGINEERING: EQUIPMENT AND SYSTEMS Materials and Pressure Design § 128.220 Class II... 9. The CMS Level-1 Trigger system for LHC Run II CERN Document Server Tapper, Alexander 2017-01-01 During LHC Run II the centre-of-mass energy of pp collisions has increased up to 13 TeV and the instantaneous luminosity has progressed towards $2 \\times 10^{34} \\rm{cm}^{-2}\\rm{s}^{-1}$. In order to guarantee a successful and ambitious physics programme under these conditions, the CMS trigger system has been upgraded. The upgraded CMS Level-1 trigger is designed to improve performance at high luminosity and large number of simultaneous inelastic collisions per crossing. The trigger design, implementation and commissioning are summarised and performance results are described. 10. Holographic spectrum-splitting optical systems for solar photovoltaics Science.gov (United States) Zhang, Deming Solar energy is the most abundant source of renewable energy available. The relatively high cost prevents solar photovoltaic (PV) from replacing fossil fuel on a larger scale. In solar PV power generation the cost is reduced with more efficient PV technologies. In this dissertation, methods to improve PV conversion efficiency with holographic optical components are discussed. The tandem multiple-junction approach has achieved very high conversion efficiency. However it is impossible to manufacture tandem PV cells at a low cost due to stringent fabrication standards and limited material types that satisfy lattice compatibility. Current produced by the tandem multi-junction PV cell is limited by the lowest junction due to series connection. Spectrum-splitting is a lateral multi-junction concept that is free of lattice and current matching constraints. Each PV cell can be optimized towards full absorption of a spectral band with tailored light-trapping schemes. Holographic optical components are designed to achieve spectrum-splitting PV energy conversion. The incident solar spectrum is separated onto multiple PV cells that are matched to the corresponding spectral band. Holographic spectrum-splitting can take advantage of existing and future low-cost technologies that produces high efficiency thin-film solar cells. Spectrum-splitting optical systems are designed and analyzed with both transmission and reflection holographic optical components. Prototype holograms are fabricated and high optical efficiency is achieved. Light-trapping in PV cells increases the effective optical path-length in the semiconductor material leading to improved absorption and conversion efficiency. It has been shown that the effective optical path length can be increased by a factor of 4n2 using diffusive surfaces. Ultra-light-trapping can be achieved with optical filters that limit the escape angle of the diffused light. Holographic reflection gratings have been shown to act as angle 11. Optics CERN Document Server Mathieu, Jean Paul 1975-01-01 Optics, Parts 1 and 2 covers electromagnetic optics and quantum optics. The first part of the book examines the various of the important properties common to all electromagnetic radiation. This part also studies electromagnetic waves; electromagnetic optics of transparent isotropic and anisotropic media; diffraction; and two-wave and multi-wave interference. The polarization states of light, the velocity of light, and the special theory of relativity are also examined in this part. The second part is devoted to quantum optics, specifically discussing the classical molecular theory of optical p 12. Optical engineering of an excimer laser ophthalmic surgery system Science.gov (United States) Yoder, Paul R., Jr. 1991-04-01 We outline the interrelationships between the various major subsystems of an apparatus being developed in the USA for in vivo surgical recontouring of the human cornea to correct refractive errors and/or to remove damaged tissue. We then discuss five significant optical engineering problems that had to be solved during the design of this apparatus in order for the overall system to function properly. 13. Fiber optic high-definition large-screen display system Science.gov (United States) Griffith, Lawrence L.; Reidinger, Michael J.; Feigles, Edward M. 1990-08-01 TRU-LYTE Systems, Inc. is developing an HDTV display that will exceed displays in the large screen display (LSD) market. Due to the present design and manufacturing techniques of LCDs, ELs, and CRTs there are limitations with LSD applications. One of the possible solutions is a hybrid of fiber optic technology and transmissive active matrix LCDs. In this design, multiple LCD modules are coupled with an equal number of fiber optic modules. These modules are designed so that strands of fiber optics are placed in a coherent manner from a rear panel to a predetermined spaced front panel. An image projected onto the rear panel will result in an enlarged image being displayed on the front panel. Imageboard modules would then be manufactured using this design of the building block method. The determining factors would include the desired output intensity, size restrictions, and cost factors. Research has also developed a technology that allows for consistent wide-angle viewing of the image displa'ed by the optical fibers. Applications for this product range from HDTV to stadium scoreboards. 14. Adaptive optics ophthalmologic systems using dual deformable mirrors Energy Technology Data Exchange (ETDEWEB) Jones, S; Olivier, S; Chen, D; Sadda, S; Joeres, S; Zawadzki, R; Werner, J S; Miller, D 2007-02-01 Adaptive Optics (AO) have been increasingly combined with a variety of ophthalmic instruments over the last decade to provide cellular-level, in-vivo images of the eye. The use of MEMS deformable mirrors in these instruments has recently been demonstrated to reduce system size and cost while improving performance. However, currently available MEMS mirrors lack the required range of motion for correcting large ocular aberrations, such as defocus and astigmatism. In order to address this problem, we have developed an AO system architecture that uses two deformable mirrors, in a woofer/tweeter arrangement, with a bimorph mirror as the woofer and a MEMS mirror as the tweeter. This setup provides several advantages, including extended aberration correction range, due to the large stroke of the bimorph mirror, high order aberration correction using the MEMS mirror, and additionally, the ability to ''focus'' through the retina. This AO system architecture is currently being used in four instruments, including an Optical Coherence Tomography (OCT) system and a retinal flood-illuminated imaging system at the UC Davis Medical Center, a Scanning Laser Ophthalmoscope (SLO) at the Doheny Eye Institute, and an OCT system at Indiana University. The design, operation and evaluation of this type of AO system architecture will be presented. 15. Optical wireless networked-systems: applications to aircrafts Science.gov (United States) 2011-01-01 This paper focuses on leveraging the progress in semiconductor technologies to facilitate production of efficient light-based in-flight entertainment (IFE), distributed sensing, navigation and control systems. We demonstrate the ease of configuring "engineered pipes" using cheap lenses, etc. to achieve simple linear transmission capacity growth. Investigation of energy-efficient, miniaturized transceivers will create a wireless medium, for both inter and intra aircrafts, providing enhanced security, and improved quality-of-service for communications links in greater harmony with onboard systems. The applications will seamlessly inter-connect multiple intelligent devices in a network that is deployable for aircrafts navigation systems, onboard sensors and entertainment data delivery systems, and high-definition audio-visual broadcasting systems. Recent experimental results on a high-capacity infrared (808 nm) system are presented. The light source can be applied in a hybrid package along with a visible lighting LED for both lighting and communications. Also, we present a pragmatic combination of light communications through "Spotlighting" and existing onboard power-lines. It is demonstrated in details that a high-capacity IFE visible light system communicating over existing power-lines (VLC/PLC) may lead to savings in many areas through reduction of size, weight and energy consumption. This paper addresses the challenges of integrating optimized optical devices in the variety of environments described above, and presents mitigation and tailoring approaches for a multi-purpose optical network. 16. Self-Homodyne Detection in Optical Communication Systems Directory of Open Access Journals (Sweden) Benjamin J. Puttnam 2014-05-01 Full Text Available We review work on self-homodyne detection (SHD for optical communication systems. SHD uses a transmitted pilot-tone (PT, originating from the transmitter laser, to exploit phase noise cancellation at a coherent receiver and to enable transmitter linewidth tolerance and potential energy savings. We give an overview of SHD performance, outlining the key contributors to the optical signal-to-noise ratio penalty compared to equivalent intradyne systems, and summarize the advantages, differences and similarities between schemes using polarization-division multiplexed PTs (PDM-SHD and those using space-division multiplexed PTs (SDM-SHD. For PDM-SHD, we review the extensive work on the transmission of advanced modulation formats and techniques to minimize the trade-off with spectral efficiency, as well as recent work on digital SHD, where the SHD receiver is combined with an polarization-diversity ID front-end receiver to provide both polarization and modulation format alignment. We then focus on SDM-SHD systems, describing experimental results using multi-core fibers (MCFs with up to 19 cores, including high capacity transmission with broad-linewidth lasers and experiments incorporating SDM-SHD in networking. Additionally, we discuss the requirement for polarization tracking of the PTs at the receiver and path length alignment and review some variants of SHD before outlining the future challenges of self-homodyne optical transmission and gaps in current knowledge. 17. Impact of ADC parameters on linear optical sampling systems Science.gov (United States) Nguyen, Trung-Hien; Gay, Mathilde; Gomez-Agis, Fausto; Lobo, Sébastien; Sentieys, Olivier; Simon, Jean-Claude; Peucheret, Christophe; Bramerie, Laurent 2017-11-01 Linear optical sampling (LOS), based on the coherent photodetection of an optical signal under test with a low repetition-rate signal originating from a pulsed local oscillator (LO), enables the characterization of the temporal electric field of optical sources. Thanks to this technique, low-speed photodetectors and analog-to-digital converters (ADCs) can be integrated in the LOS system providing a cost-effective tool for characterizing high-speed signals. However, the impact of photodetector and ADC parameters on such LOS systems has not been explored in detail so far. These parameters, including the integration time of the track-and-hold function, the effective number of bits (ENOB) of the ADC, as well as the combined limited bandwidth of the photodetector and ADC are experimentally and numerically investigated in a LOS system for the first time. More specifically, by reconstructing 10-Gbit/s non-return-to-zero on-off keying (NRZ-OOK) and 10-Gbaud NRZ-quadrature phase-shift-keying (QPSK) signals, it is shown that a short integration time provides a better recovered signal fidelity. Furthermore, an ENOB of 6 bits and an ADC bandwidth normalized to the sampling rate of 2.8 are found to be sufficient in order to reliably monitor the considered signals. 18. Optical design of automotive headlight system incorporating digital micromirror device. Science.gov (United States) Hung, Chuan-Cheng; Fang, Yi-Chin; Huang, Ming-Shyan; Hsueh, Bo-Ren; Wang, Shuan-Fu; Wu, Bo-Wen; Lai, Wei-Chi; Chen, Yi-Liang 2010-08-01 In recent years, the popular adaptive front-lighting automobile headlight system has become a main emphasis of research that manufacturers will continue to focus great efforts on in the future. In this research we propose a new integral optical design for an automotive headlight system with an advanced light-emitting diode and digital micromirror device (DMD). Traditionally, automobile headlights have all been designed as a low beam light module, whereas the high beam light module still requires using accessory lamps. In anticipation of this new concept of integral optical design, we have researched and designed a single optical system with high and low beam capabilities. To switch on and off the beams, a DMD is typically used. Because DMDs have the capability of redirecting incident light into a specific angle, they also determine the shape of the high or low light beam in order to match the standard of headlight illumination. With collocation of the multicurvature reflection lens design, a DMD can control the light energy distribution and thereby reinforce the resolution of the light beam. 19. A pupil tracking system for adaptive optics retinal imaging Science.gov (United States) Sahin, Betul; Harms, Fabrice; Lamory, Barbara; Vabre, Laurent 2008-04-01 High-resolution imaging of the retina is a challenge due to the optical aberrations introduced by the eye, a living system in constant change and motion. Adaptive Optics (AO) is particularly suited to the continuous, dynamic correction of aberrations as they change over time. In particular, eye pupil displacements induce fast-changing wave front errors which lead to a need for faster wave front sensors. We propose a new approach for ocular adaptive optics by adding a Pupil Tracking System (PTS) into the AO loop. This system is different from the existing eye tracking devices by its speed, high precision in a short range and therefore its suitability for integration in an AO loop. Performance tests done using an artificial eye with a pupil diameter of 7 mm have shown promising results. These tests have demonstrated that the device achieves an accuracy of <15 μm in a +/-2 mm range of eye movements with a standard deviation <10 μm, and requires less than 12 ms for each detection. 20. ASPHERICAL SURFACES APPROXIMATION IN AUTOMATED DESIGN OF OPTICAL SYSTEMS Directory of Open Access Journals (Sweden) T. V. Ivanova 2015-07-01 Full Text Available Subject of Research. The paper deals with the problems of higher order aspherical surfaces approximation using different equation types. The objects of research are two types of equations for higher order aspherical surfaces description used in different software for optical systems design (SАRО, OPAL, ZEMAX, CODE-V, etc. and dependent on z-coordinate or on a radial coordinate on the surface. Conversion from one type of equations to another is considered in view of application in different software for optical systems design. Methods. The subject matter of the method lies in usage of mean square method approximation for recalculation of high-order aspherical surface. Iterative algorithm for recalculation is presented giving the possibility to recalculate coefficients for different types of equations with required accuracy. Recommendations are given for choosing recalculation parameters such as the number of result equation coefficients, the number of points for recalculation and point allocation on a surface. Main Results. Example of recalculation for aspherical surface and accuracy estimation, including result aberration comparison between initial surface and recalculated surface are presented. The example has shown that required accuracy of surface representation was obtained. Practical Relevance. This technique is usable for recalculation of higher order aspherical surfaces in various types of software for optical systems design and also for research of optimal higher order aspherical surfaces description. 1. High performance X-ray and neutron microfocusing optics. Phase II final report. Energy Technology Data Exchange (ETDEWEB) Gregory Hirsch 2000-01-14 The use of extremely small diameter x-ray beams at synchrotron radiation facilities has become an important experimental technique for investigators in many other scientific disciplines. While there have been several different optical elements developed for producing such microbeams, this SBIR project was concerned with one particular device: the tapered-monocapillary optic. 2. PERFORMANCE ANALYSIS OF OPTICAL CDMA SYSTEM USING VC CODE FAMILY UNDER VARIOUS OPTICAL PARAMETERS Directory of Open Access Journals (Sweden) HASSAN YOUSIF AHMED 2012-06-01 Full Text Available The intent of this paper is to study the performance of spectral-amplitude coding optical code-division multiple-access (OCDMA systems using Vector Combinatorial (VC code under various optical parameters. This code can be constructed by an algebraic way based on Euclidian vectors for any positive integer number. One of the important properties of this code is that the maximum cross-correlation is always one which means that multi-user interference (MUI and phase induced intensity noise are reduced. Transmitter and receiver structures based on unchirped fiber Bragg grating (FBGs using VC code and taking into account effects of the intensity, shot and thermal noise sources is demonstrated. The impact of the fiber distance effects on bit error rate (BER is reported using a commercial optical systems simulator, virtual photonic instrument, VPITM. The VC code is compared mathematically with reported codes which use similar techniques. We analyzed and characterized the fiber link, received power, BER and channel spacing. The performance and optimization of VC code in SAC-OCDMA system is reported. By comparing the theoretical and simulation results taken from VPITM, we have demonstrated that, for a high number of users, even if data rate is higher, the effective power source is adequate when the VC is used. Also it is found that as the channel spacing width goes from very narrow to wider, the BER decreases, best performance occurs at a spacing bandwidth between 0.8 and 1 nm. We have shown that the SAC system utilizing VC code significantly improves the performance compared with the reported codes. 3. Optical System and Desing Of The New 1.6 Meter Wide-Field Telescope With Active Optics Science.gov (United States) Papushev, Pavel; Denisenko, Sergey; Kamus, Sergey; Pimenov, Yury; Tergoev, Vladim 2006-08-01 In this report we present and discuss the design, construction and capabilities of the two meters class wide field survey telescope. The designs based on modified R-C system with two or three lens correctors in visible and near infrared (2,2 mkm) spectral range. The optical systems of the 1.6 meters telescope with up to 3 degrees field of view and less than 15% obscuration area are considered in detail. Optical performance of system, its mount and separate element of the active optics system are examined. 4. Geometrical co-calibration of a tomographic optical system with CT for intrinsically co-registered imaging. Science.gov (United States) Cao, Liji; Breithaupt, Mathies; Peter, Jörg 2010-03-21 A mathematical approach for geometric co-calibration of a dual-modal small-animal imaging system is presented. The system comprises an optical imaging setup for in vivo bioluminescence and fluorescence detection, as well as an x-ray CT, both mounted on a common rotatable gantry enabling fully simultaneous imaging at axially overlapping fields-of-view. Geometric co-calibration is performed once by imaging a single cylindrical light-emitting source with both modalities over 360 degrees at two axial positions, respectively. Given the three-dimensional coordinates of the source positions in the reconstructed CT volume data along with their two-dimensional locations projected at the optical detector plane, the following intrinsic system parameters are calculated: (i) the intrinsic geometric parameters of the optical detection system-five parameters for each view and (ii) the relative positional relationship between the optical and CT systems-two parameters for each view. After co-calibration is performed, experimental studies using phantoms demonstrate the high degree of intrinsic positional accuracy between the optical and CT measurements. The most important advantage of this approach is that dual-modal data fusion is accomplished without any post-registration strategies. 5. Geometrical co-calibration of a tomographic optical system with CT for intrinsically co-registered imaging Energy Technology Data Exchange (ETDEWEB) Cao Liji; Breithaupt, Mathies; Peter, Joerg [Division of Medical Physics in Radiology, German Cancer Research Center, Im Neuenheimer Feld 280, 69120 Heidelberg (Germany)], E-mail: l.cao@dkfz.de 2010-03-21 A mathematical approach for geometric co-calibration of a dual-modal small-animal imaging system is presented. The system comprises an optical imaging setup for in vivo bioluminescence and fluorescence detection, as well as an x-ray CT, both mounted on a common rotatable gantry enabling fully simultaneous imaging at axially overlapping fields-of-view. Geometric co-calibration is performed once by imaging a single cylindrical light-emitting source with both modalities over 360 deg. at two axial positions, respectively. Given the three-dimensional coordinates of the source positions in the reconstructed CT volume data along with their two-dimensional locations projected at the optical detector plane, the following intrinsic system parameters are calculated: (i) the intrinsic geometric parameters of the optical detection system-five parameters for each view and (ii) the relative positional relationship between the optical and CT systems-two parameters for each view. After co-calibration is performed, experimental studies using phantoms demonstrate the high degree of intrinsic positional accuracy between the optical and CT measurements. The most important advantage of this approach is that dual-modal data fusion is accomplished without any post-registration strategies. 6. Final Report for the MANNRRSS II Program Management of Nevada's Natural Resources with Remote Sensing Systems, Beatty, NV Energy Technology Data Exchange (ETDEWEB) Lester Miller; Brian Horowitz; Chris Kratt; Tim Minor; Stephen F. Zitzer; James. V. Taranik; Zan L. Aslett; Todd O. Morken 2009-06-04 This document provides the Final Report on the Management of Nevada’s Natural Resources with Remote Sensing Systems (MANNRRSS) II program. This is a U.S. Department of Energy (DOE)-funded project tasked with utilizing hyperspectral and ancillary electro-optical instrumentation data to create an environmental characterization of an area directly adjacent to the Nevada Test Site (NTS). 7. Detecting tissue optical and mechanical properties with an ultrasound modulated optical imaging system in reflection detection geometry. Science.gov (United States) Cheng, Yi; Li, Sinan; Eckersley, Robert J; Elson, Daniel S; Tang, Meng-Xing 2015-01-01 Tissue optical and mechanical properties are correlated to tissue pathologic changes. This manuscript describes a dual-mode ultrasound modulated optical imaging system capable of sensing local optical and mechanical properties in reflection geometry. The optical characterisation was achieved by the acoustic radiation force assisted ultrasound modulated optical tomography (ARF-UOT) with laser speckle contrast detection. Shear waves generated by the ARF were also tracked optically by the same system and the shear wave speed was used for the elasticity measurement. Tissue mimicking phantoms with multiple inclusions buried at 11 mm depth were experimentally scanned with the dual-mode system. The inclusions, with higher optical absorption and/or higher stiffness than background, were identified based on the dual results and their stiffnesses were quantified. The system characterises both optical and mechanical properties of the inclusions compared with the ARF-UOT or the elasticity measurement alone. Moreover, by detecting the backward scattered light in reflection detection geometry, the system is more suitable for clinical applications compared with transmission geometry. 8. Multi-platform laser communication networking optical antenna system design Science.gov (United States) Zhang, Tao 2016-10-01 In this paper, a new conclusion based on rotating parabolic model and a different scheme of laser communication networking antenna system has been put forward in the paper. Based on rotating parabolic antenna, a new theory of the optical properties have been deduced, which can realize larger dynamic, duplex, networking communications among multiple platforms in 360° azimuth and pitch range. Meanwhile, depending on the operation mode of the system, multiple mathematical optimization models have been established. Tracking communication range, emission energy efficiency and receiving energy efficiency have been analyzed and optimized. Relationship among opening up and low apertures, the lens unit aperture, focal length of lens unit as well as rotating parabolic focal length have been analyzed. Tracking pitching range and emission energy utilization has carried on the theoretical derivation and optimization and networking platform link between energy receiver and transmitter has been analyzed. Taking some parameters of this new system into calculation, optimized results can be utilized with MATLAB software for its application and system of communication engineering. The rotating parabolic internal can form a hollow structure, which is utilized for miniaturization, light-weighted design and realize duplex communication in a wide range and distance. Circular orbit guidance is the modern way used in dynamic tracking system. The new theory and optical antenna system has widespread applications value as well. 9. 78 FR 17187 - Notice of Intent To Grant Exclusive Patent License; Fiber Optic Sensor Systems Technology... Science.gov (United States) 2013-03-20 ... Department of the Navy Notice of Intent To Grant Exclusive Patent License; Fiber Optic Sensor Systems... to grant to Fiber Optic Sensor Systems Technology Corporation, a revocable, nonassignable, exclusive... its intent to grant to Fiber Optic Sensor Systems Technology Corporation a revocable, nonassignable... 10. LISA Pathfinder: Optical Metrology System monitoring during operations Science.gov (United States) Audley, Heather E.; LISA Pathfinder Collaboration 2017-05-01 The LISA Pathfinder (LPF) mission has demonstrated excellent performance. In addition to having surpassed the main mission goals, data has been collected from the various subsystems throughout the duration of the mission. This data is a valuable resource, both for a more complete understanding of the LPF satellite and the differential acceleration measurements, as well as for the design of the future Laser Interferometer Space Antenna (LISA) mission. Initial analysis of the Optical Metrology System (OMS) data was performed as part of daily system monitoring, and more in-depth analyses are ongoing. This contribution presents an overview of these activities along with an introduction to the OMS. 11. LED Based Optical Wireless Communication System for WBAN Directory of Open Access Journals (Sweden) ATTIYA BAQAI 2016-10-01 Full Text Available In this paper the O-WBANs (Optical Wireless Body Area Networks is presented as an alternative to the radio and microwave BANs which are plagued with issues like interference, power hungry, hazardous and costly spectrum. Various experiments performed in this work demonstrate the feasibility of LED (Light-Emitting Diode based optical systems to be used for BANs. The system cost is kept as low as possible. Mainly the effects on the optical link are observed under ambient light and with different modulation schemes by varying link distance and line of sight in this paper. Experimental results reveal the satisfactory link availability up to the distance of 7 feet (around 2m, the optimum range of WBAN and within the field of view of 30-60º. The low cost solution presented in this paper meets the WBANs data rate requirement for physiological data (i.e. 10-100 kbps. As IR (Infra-Red signals do not provide any interference with the signals generated by the medical equipment of hospitals which is the problem in case of RF or microwave signals, additionally the IR signals are confined within a room hence IR signaling can prove to be potential candidate for WBAN fulfilling its security and limited access requirement. 12. Optical Sensor for the Determination of Lead(II Based On Immobilization of Dithizone onto Chitosan-Silica Membrane Directory of Open Access Journals (Sweden) Yuspian Nur 2017-04-01 Full Text Available Optical sensor based on incorporation of dithizone and chitosan-silica for the determination of lead (II ion has been conducted. The optical sensor was made by composite of chitosan and silica in the ratio 2:1. Sol gel technique was chosen to make the membrane because it has some advantages such as its low temperature and chemical reactivity. The sensor has a good selectivity at pH 5 with response times at ± 180 s. Linearity responses was obtained with concentration range from 0.2 to 1.1 ppm with r2 = 0.9921. Average of %RSD on precision and accuracy (%PK test were 1.46 and 100.96%, respectively. Limit of detection (LOD and limit of quantification (LOQ were 0.11 ppm and 0.37 ppm, respectively Science.gov (United States) Wang, Yukun; Xu, Huanyu; Li, Dayu; Wang, Rui; Jin, Chengbin; Yin, Xianghui; Gao, Shijie; Mu, Quanquan; Xuan, Li; Cao, Zhaoliang 2018-01-18 The performance of free-space optics communication (FSOC) is greatly degraded by atmospheric turbulence. Adaptive optics (AO) is an effective method for attenuating the influence. In this paper, the influence of the spatial and temporal characteristics of turbulence on the performance of AO in a FSOC system is investigated. Based on the Greenwood frequency (GF) and the ratio of receiver aperture diameter to atmospheric coherent length (D/r 0 ), the relationship between FSOC performance (CE) and AO parameters (corrected Zernike modes number and bandwidth) is derived for the first time. Then, simulations and experiments are conducted to analyze the influence of AO parameters on FSOC performance under different GF and D/r 0 . The simulation and experimental results show that, for common turbulence conditions, the number of corrected Zernike modes can be fixed at 35 and the bandwidth of the AO system should be larger than the GF. Measurements of the bit error rate (BER) for moderate turbulence conditions (D/r 0  = 10, f G  = 60 Hz) show that when the bandwidth is two times that of GF, the average BER is decreased by two orders of magnitude compared with f G /f 3dB  = 1. These results and conclusions can provide important guidance in the design of an AO system for FSOC. 14. Very deep spectroscopy of the bright Saturn nebula NGC 7009 - II. Analysis of the rich optical recombination spectrum Science.gov (United States) Fang, X.; Liu, X.-W. 2013-03-01 In Paper I, we presented a deep, long-slit spectrum of the bright Saturn nebula NGC 7009. Numerous permitted lines emitted by the C+, N+, O+ and Ne+ ions were detected. Gaussian profile fitting to the spectrum yielded more than 1000 lines, the majority of which are optical recombination lines (ORLs) of heavy-element ions. In the current paper, we present a critical analysis of the rich optical recombination spectrum of NGC 7009, in the context of the bi-abundance nebular model proposed by Liu et al. Transitions from individual multiplets are checked carefully for potential blended lines. The observed relative intensities are compared with the theoretical predictions based on high-quality effective recombination coefficients, now available for the recombination line spectrum of a number of heavy-element ions. The possibility of plasma diagnostics using the ORLs of various heavy-element ions is discussed in detail. The line ratios that can be used to determine electron temperature are presented for each ion, although there is still a lack of adequate atomic data and some of the lines are still not detected in the spectrum of NGC 7009 due to weakness and/or line blending. Plasma diagnostics based on the N II and O II recombination spectra both yield electron temperatures close to 1000 K, which is lower than those derived from the collisionally excited line (CEL) ratios (e.g. the [O III] and [N II] nebular-to-auroral line ratios; see Paper I for details) by nearly one order of magnitude. The very low temperatures yielded by the O II and N II ORLs indicate that they originate from very cold regions. The C2+/H+, N2+/H+, O2+/H+ and Ne2+/H+ ionic abundance ratios derived from ORLs are consistently higher, by about a factor of 5, than the corresponding values derived from CELs. In calculating the ORL ionic abundance ratios, we have used the newly available high-quality effective recombination coefficients, and adopted an electron temperature of ˜1000 K, as given by the ORL 15. Okayama optical polarimetry and spectroscopy system (OOPS) I. General description. Science.gov (United States) Sasaki, T.; Yutani, M.; Shimizu, Y.; Kurakami, T. An imaging polarimeter and spectrometer (OOPS) was built for the 91 cm telescope at the Okayama Astrophysical Observatory. The optics consists of a collimator lens (f = 135 mm) and camera lens (f = 50 mm) with several optical elements in the parallel beam section to enable several observing modes: polarimetry, grism low-dispersion spetroscopy, spectro-polarimetry, and direct imaging. A Rochon prism coupled with a rotating half wave plate makes two-channel polarized images on a CCD. Control system is designed to realize a quick change of observing modes and flexible control for scheduled observation in cooperation with the 91 cm telescope control and the autoguider with the help of message-based communication. 16. Fiber Optic Displacement Sensor System for Structural Health Monitoring Energy Technology Data Exchange (ETDEWEB) Lee, Kun Ho; Ahn, Byung Jun; Kim, Dae Hyun [Seoul National University of Science and Technology, Seoul (Korea, Republic of) 2011-08-15 It has been doing to research on novel techniques for structural health monitoring by applying various sensor techniques to measure the deflection in mechanical and civil structures. Several electric-based displacement sensors have many difficulties for using them because of EMI (Electro-Magnetic Interference) noise of many lead-wires when they are installed to many points in the structures. In this paper, it is proposed an affordable intensity-based fiber optic sensor to measure small displacement solving the problems of conventional sensors. In detail, the sensor head was designed on the basis of the principle of bending loss and a basic experiment was performed to obtain the sensitivity, the linearity and the stroke of the sensor. Moreover, a prototype was designed and manufactured to be easily installed to a structure and a real-time control software was also successfully developed to drive the fiber optic sensor system. 17. Optical cuff for optogenetic control of the peripheral nervous system Science.gov (United States) Michoud, Frédéric; Sottas, Loïc; Browne, Liam E.; Asboth, Léonie; Latremoliere, Alban; Sakuma, Miyuki; Courtine, Grégoire; Woolf, Clifford J.; Lacour, Stéphanie P. 2018-02-01 Objective. Nerves in the peripheral nervous system (PNS) contain axons with specific motor, somatosensory and autonomic functions. Optogenetics offers an efficient approach to selectively activate axons within the nerve. However, the heterogeneous nature of nerves and their tortuous route through the body create a challenging environment to reliably implant a light delivery interface. Approach. Here, we propose an optical peripheral nerve interface—an optocuff—, so that optogenetic modulation of peripheral nerves become possible in freely behaving mice. Main results. Using this optocuff, we demonstrate orderly recruitment of motor units with epineural optical stimulation of genetically targeted sciatic nerve axons, both in anaesthetized and in awake, freely behaving animals. Behavioural experiments and histology show the optocuff does not damage the nerve thus is suitable for long-term experiments. Significance. These results suggest that the soft optocuff might be a straightforward and efficient tool to support more extensive study of the PNS using optogenetics. 18. Advances in lasers and optical micro-nano-systems Science.gov (United States) Laurell, F.; Fazio, E. 2010-09-01 Lasers represent a well consolidated technology: nevertheless, research in this field remains very active and productive, in both basic and applied directions. At the moment significant attention is given to those sources that bring together high power and compactness. Such high power lasers find important applications for material treatments and such applications are presented by Ehsani et al and Saiedeh Saghafi et al, in the treatment of dielectric thin films (Alteration of optical and morphological properties of polycarbonate illuminated by visible/IR laser beams) or of biological tissues like pistachio seeds (Investigating the effects of laser beams (532 and 660 nm) in annihilation of pistachio mould fungus using spectrophotometry analysis). In particular the latter paper show how laser sources can find very important applications in new domains, preserving goods and food without the need for preservatives or pesticides by simply sterilizing them using light. Optical Micro and Nano Systems presents a new domain for exploration. In this framework this special issue is very attractive, because it assembles papers reporting new results in three directions: new techniques for monitoring integrated micro- and nano-systems, new integrated systems and novel high performance metamaterial configurations. Integrated micro-components can be monitored and controlled using reflectance measurements as presented by Piombini et al (Toward the reflectance measurement of micro components). Speckle formation during laser beam reflection can also be a very sophisticated tool for detecting ultra-precise displacements, as presented by Filter et al (High resolution displacement detection with speckles : accuracy limits in linear displacement speckle metrology). Three dimensional integrated optical structures is indeed a big challenge and a peculiarity of photonics, they can be formed through traditional holography or using more sophisticated and novel ! technologies. Thus, special 19. Protocols for Relay-Assisted Free-Space Optical Systems CERN Document Server Chatzidiamantis, Nestor D; Kriezis, Emmanouil E; Karagiannidis, George K; Schober, Robert 2011-01-01 We investigate transmission protocols for relay-assisted free-space optical (FSO) systems, when multiple parallel relays are employed and there is no direct link between the source and the destination. As alternatives to all-active FSO relaying, where all the available relays transmit concurrently, we propose schemes that select only a single relay to participate in the communication between the source and the destination in each transmission slot. This selection is based on the channel state information (CSI) obtained either from all or from some of the FSO links. Thus, the need for synchronizing the relays' transmissions is avoided and the slowly varying nature of the atmospheric channel is exploited. For both relay selection and all-active relaying, novel closed-form expressions for their outage performance are derived, assuming the versatile Gamma-Gamma channel model. Furthermore, based on the derived analytical results, the problem of allocating the optical power resources to the FSO links is addressed, ... 20. Fabrication and optical properties of type-II InP/InAs nanowire/quantum-dot heterostructures Energy Technology Data Exchange (ETDEWEB) Yan, Xin; Zhang, Xia; Li, Junshuai; Wu, Yao; Li, Bang; Ren, Xiaomin [State Key Laboratory of Information Photonics and Optical Communications, Beijing University of Posts and Telecommunications, Beijing, 100876 (China) 2016-02-15 The growth and optical properties of InAs quantum dots on a pure zinc blende InP nanowire are investigated. The quantum dots are formed in Stranski-Krastanov mode and exhibit pure zinc blende crystal structure. A substantial blueshift of the dots peak with a cube-root dependence on the excitation power is observed, suggesting a type-II band alignment. The peak position of dots initially red-shifts and then blue-shifts with increasing temperature, which is attributed to the carrier redistribution among the quantum dots. (copyright 2016 WILEY-VCH Verlag GmbH and Co. KGaA, Weinheim) 1. High-Frequency Flush Mounted Miniature LOX Fiber-Optic Pressure Sensor II Project Data.gov (United States) National Aeronautics and Space Administration — Luna Innovations has teamed with the University of Alabama, Huntsville, to develop a miniature flush-mounted fiber-optic pressure sensor that will allow accurate,... 2. Sol-gel based optical sensor for determination of Fe (II): a novel probe for iron speciation. Science.gov (United States) 2015-02-05 A highly selective optical sensor for Fe (II) ions was developed based on entrapment of a sensitive reagent, 2,4,6-tri(2-pyridyl)-s-triazine (TPTZ), in a silica sol-gel thin film coated on a glass substrate. The thin films fabricated based on tetraethoxysilane (TEOS) as precursor, sol-gel pH∼3, water:alkoxyde ratio of 4:1 and TPTZ concentration of 0.112 mol L(-1). The influence of sol-gel parameters on sensing behavior of the fabricated sensor was also investigated. The fabricated sensor can be used for determination of Fe (II) ion with an outstanding high selectivity over a dynamic range of 5-115 ng mL(-1) and a detection limit of 1.68 ng mL(-1). It also showed reproducible results with relative standard deviation of 3.5% and 1.27% for 10 and 90 ng mL(-1) of Fe (II), respectively, along with a fast response time of ∼120 s. Total iron also was determined after reduction of Fe (III) to Fe (II) using ascorbic acid as reducing agent. Then, the concentration of Fe (III) was calculated by subtracting the concentration of Fe (II) from the total iron concentration. Interference studies showed a good selectivity for Fe (II) with trapping TPTZ into sol-gel matrix and appropriately adjusting the structure of doped sol-gel. The sensor was compared with other sensors and was applied to determine iron in different water samples with good results. Copyright © 2014 Elsevier B.V. All rights reserved. 3. Noise and signal interference in optical fiber transmission systems an optimum design approach CERN Document Server Bottacchi, Stefano 2008-01-01 A comprehensive reference to noise and signal interference in optical fiber communications Noise and Signal Interference in Optical Fiber Transmission Systems is a compendium on specific topics within optical fiber transmission and the optimization process of the system design. It offers comprehensive treatment of noise and intersymbol interference (ISI) components affecting optical fiber communications systems, containing coverage on noise from the light source, the fiber and the receiver. The ISI is modeled with a statistical approach, leading to new useful computational m 4. THE RADIO AND OPTICAL LUMINOSITY EVOLUTION OF QUASARS. II. THE SDSS SAMPLE Energy Technology Data Exchange (ETDEWEB) Singal, J.; Petrosian, V. [Kavli Institute for Particle Astrophysics and Cosmology, SLAC National Accelerator Laboratory and Stanford University, 382 Via Pueblo Mall, Stanford, CA 94305-4060 (United States); Stawarz, L. [Institute of Space and Astronautical Science (ISAS), Japan Aerospace Exploration Agency (JAXA), 3-1-1 Yoshinodai, Chuo-ku, Sagamihara, Kanagawa 252-5510 (Japan); Lawrence, A., E-mail: jsingal@stanford.edu [University of Edinburgh Institute for Astronomy, Scottish Universities Physics Alliance (SUPA), Royal Observatory, Blackford Hill, Edinburgh (United Kingdom) 2013-02-10 We determine the radio and optical luminosity evolutions and the true distribution of the radio-loudness parameter R, defined as the ratio of the radio to optical luminosity, for a set of more than 5000 quasars combining Sloan Digital Sky Survey optical and Faint Images of the Radio Sky at Twenty cm (FIRST) radio data. We apply the method of Efron and Petrosian to access the intrinsic distribution parameters, taking into account the truncations and correlations inherent in the data. We find that the population exhibits strong positive evolution with redshift in both wavebands, with somewhat greater radio evolution than optical. With the luminosity evolutions accounted for, we determine the density evolutions and local radio and optical luminosity functions. The intrinsic distribution of the radio-loudness parameter R is found to be quite different from the observed one and is smooth with no evidence of a bimodality in radio loudness for log R {>=} -1. The results we find are in general agreement with the previous analysis of Singal et al., which used POSS-I optical and FIRST radio data. 5. The Radio and Optical Luminosity Evolution of Quasars II - The SDSS Sample Energy Technology Data Exchange (ETDEWEB) Singal, J.; Petrosian, V.; Stawarz, L.; Lawrence, A. 2012-12-28 We determine the radio and optical luminosity evolutions and the true distribution of the radio loudness parameter R, defined as the ratio of the radio to optical luminosity, for a set of more than 5000 quasars combining SDSS optical and FIRST radio data. We apply the method of Efron and Petrosian to access the intrinsic distribution parameters, taking into account the truncations and correlations inherent in the data. We find that the population exhibits strong positive evolution with redshift in both wavebands, with somewhat greater radio evolution than optical. With the luminosity evolutions accounted for, we determine the density evolutions and local radio and optical luminosity functions. The intrinsic distribution of the radio loudness parameter R is found to be quite different than the observed one, and is smooth with no evidence of a bi-modality in radio loudness. The results we find are in general agreement with the previous analysis of Singal et al., 2011 which used POSS-I optical and FIRST radio data. 6. Optical secure image verification system based on ghost imaging Science.gov (United States) Wu, Jingjing; Haobogedewude, Buyinggaridi; Liu, Zhengjun; Liu, Shutian 2017-09-01 The ghost imaging can perform Fourier-space filtering by tailoring the configuration. We proposed a novel optical secure image verification system based on this theory with the help of phase matched filtering. In the verification process, the system key and the ID card which contain the information of the correct image and the information to be verified are put in the reference and the test paths, respectively. We demonstrate that the ghost imaging configuration can perform an incoherent correlation between the system key and the ID card. The correct verification manifests itself with a correlation peak in the ghost image. The primary image and the image to be verified are encrypted and encoded into pure phase masks beforehand for security. Multi-image secure verifications can also be implemented in the proposed system. 7. 4F-based optical phase imaging system DEFF Research Database (Denmark) 2013-01-01 The invention relates to 4F-based optical phase imaging system and in particular to reconstructing quantitative phase information of an object when using such systems. The invention applies a two-dimensional, complex spatial light modulator (SLM) to impress a complex spatial synthesized modulation...... in addition to the complex spatial modulation impressed by the object. This SLM is arranged so that the synthesized modulation is superimposed with the object modulation and is thus placed at an input plane to the phase imaging system. By evaluating output images from the phase imaging system, the synthesized...... modulation is selected to optimize parameters in the output image which improves the reconstruction of qualitative and quantitative object phase information from the resulting output images.... 8. Technicon H*1 Hematology System: Optical Design Considerations Science.gov (United States) Colella, G. M.; Tycko, D. H.; Groner, W. 1988-06-01 The Technicon H*1 systemTM is a clinical laboratory flow cytometer which performs a complete hematology profile, providing quantitative information on the various types of cells in a blood sample. A light-scattering method, using a HeNe laser, determines in a single flow channel the red cell count, platelet count, and the distributions of red cell volume, red cell hemoglobin concentration, and platelet volume. To accomplish this the scattered light from each red cell in the sample is measured in real time at two angular intervals. The cell volume and the hemoglobin concentration within the cell are derived from these two measurements. Severe accuracy and precision specifications are placed on the medically important red cell count (RBC) and the mean red cell volume (MCV). From the point of view of optical system design, the dominant factor is the requirement that RBC and MCV have precision and accuracy of the order of 2%. Signal-to-noise and scattering-angle definition requirements dictated the choice of a HeNe laser light source. The optics includes an illumination system for producing a sharply defined, uniformly illuminated scattering region and a detection system which must accurately define the accepted scattering angles. In previous cytometric methods for determining MCV only a single quantity was measured for each cell. Such methods cannot disentangle the independent effects of cell size and hemoglobin concentration on the measurement, thus compromising MCV accuracy. The present double-angle scattering method overcomes this accuracy problem. The H*1 red cell method, the supporting optical design and data demonstrating that the use of this technique eliminates interference between the observed red cell indices are presented. 9. Blade System Design Study. Part II, final project report (GEC). Energy Technology Data Exchange (ETDEWEB) Griffin, Dayton A. (DNV Global Energy Concepts Inc., Seattle, WA) 2009-05-01 As part of the U.S. Department of Energy's Low Wind Speed Turbine program, Global Energy Concepts LLC (GEC)1 has studied alternative composite materials for wind turbine blades in the multi-megawatt size range. This work in one of the Blade System Design Studies (BSDS) funded through Sandia National Laboratories. The BSDS program was conducted in two phases. In the Part I BSDS, GEC assessed candidate innovations in composite materials, manufacturing processes, and structural configurations. GEC also made recommendations for testing composite coupons, details, assemblies, and blade substructures to be carried out in the Part II study (BSDS-II). The BSDS-II contract period began in May 2003, and testing was initiated in June 2004. The current report summarizes the results from the BSDS-II test program. Composite materials evaluated include carbon fiber in both pre-impregnated and vacuum-assisted resin transfer molding (VARTM) forms. Initial thin-coupon static testing included a wide range of parameters, including variation in manufacturer, fiber tow size, fabric architecture, and resin type. A smaller set of these materials and process types was also evaluated in thin-coupon fatigue testing, and in ply-drop and ply-transition panels. The majority of materials used epoxy resin, with vinyl ester (VE) resin also used for selected cases. Late in the project, testing of unidirectional fiberglass was added to provide an updated baseline against which to evaluate the carbon material performance. Numerous unidirectional carbon fabrics were considered for evaluation with VARTM infusion. All but one fabric style considered suffered either from poor infusibility or waviness of fibers combined with poor compaction. The exception was a triaxial carbon-fiberglass fabric produced by SAERTEX. This fabric became the primary choice for infused articles throughout the test program. The generally positive results obtained in this program for the SAERTEX material have led to its 10. New optical architecture for holographic data storage system compatible with Blu-ray Disc™ system Science.gov (United States) Shimada, Ken-ichi; Ide, Tatsuro; Shimano, Takeshi; Anderson, Ken; Curtis, Kevin 2014-02-01 A new optical architecture for holographic data storage system which is compatible with a Blu-ray Disc™ (BD) system is proposed. In the architecture, both signal and reference beams pass through a single objective lens with numerical aperture (NA) 0.85 for realizing angularly multiplexed recording. The geometry of the architecture brings a high affinity with an optical architecture in the BD system because the objective lens can be placed parallel to a holographic medium. Through the comparison of experimental results with theory, the validity of the optical architecture was verified and demonstrated that the conventional objective lens motion technique in the BD system is available for angularly multiplexed recording. The test-bed composed of a blue laser system and an objective lens of the NA 0.85 was designed. The feasibility of its compatibility with BD is examined through the designed test-bed. 11. A Sort-Last Rendering System over an Optical Backplane Directory of Open Access Journals (Sweden) Yasuhiro Kirihata 2005-06-01 Full Text Available Sort-Last is a computer graphics technique for rendering extremely large data sets on clusters of computers. Sort-Last works by dividing the data set into even-sized chunks for parallel rendering and then composing the images to form the final result. Since sort-last rendering requires the movement of large amounts of image data among cluster nodes, the network interconnecting the nodes becomes a major bottleneck. In this paper, we describe a sort-last rendering system implemented on a cluster of computers whose nodes are connected by an all-optical switch. The rendering system introduces the notion of the Photonic Computing Engine, a computing system built dynamically by using the optical switch to create dedicated network connections among cluster nodes. The sort-last volume rendering algorithm was implemented on the Photonic Computing Engine, and its performance is evaluated. Prelimi- nary experiments show that performance is affected by the image composition time and average payload size. In an attempt to stabilize the performance of the system, we have designed a flow control mechanism that uses feedback messages to dynamically adjust the data flow rate within the computing engine. 12. Progress on PEP-II magnet power conversion system Energy Technology Data Exchange (ETDEWEB) Bellomo, P.; Genova, L. [Stanford Linear Accelerator Center, Menlo Park, CA (United States); Jackson, T. [Lawrence Berkeley National Lab., CA (United States); Shimer, D. [Lawrence Livermore National Lab., CA (United States) 1996-06-04 The various power systems for supplying the PEP-II DC magnets rely exclusively on switchmode conversion, utilizing a variety of means depending on the requirements. All of the larger power supplies, ranging from 10 to 200 kW, are powered from DC sources utilizing rectified 480 V AC. Choppers can be used for the series connected strings, but for smaller groups and individual magnets, inverters driving high-frequency transformers with rectifiers comprise the best approach. All of the various systems use a building block approach of multiple standard-size units connected in series or parallel to most cost-effectively deal with a great range of voltage and current requirements. Utilization of existing infrastructure from PEP-I has been a cost-effective determinant. Equipment is being purchased either off-the-shelf, through performance specification, or by hardware purchase based on design-through-prototype. The corrector magnet power system, utilizing inexpensive, off-the-shelf, four-quadrant switching motor-controllers, has already proven very reliable: 120 of the total of 900 units have been running on the injection system for four months with no failures. 13. Sirepo: a web-based interface for physical optics simulations - its deployment and use at NSLS-II Science.gov (United States) Rakitin, Maksim S.; Chubar, Oleg; Moeller, Paul; Nagler, Robert; Bruhwiler, David L. 2017-08-01 "Sirepo" is an open source cloud-based software framework which provides a convenient and user-friendly web-interface for scientific codes such as Synchrotron Radiation Workshop (SRW) running on a local machine or a remote server side. SRW is a physical optics code allowing to simulate the synchrotron radiation from various insertion devices (undulators and wigglers) and bending magnets. Another feature of SRW is a support of high-accuracy simulation of fully- and partially-coherent radiation propagation through X-ray optical beamlines, facilitated by so-called "Virtual Beamline" module. In the present work, we will discuss the most important features of Sirepo/SRW interface with emphasis on their use for commissioning of beamlines and simulation of experiments at National Synchrotron Light Source II. In particular, "Flux through Finite Aperture" and "Intensity" reports, visualizing results of the corresponding SRW calculations, are being routinely used for commissioning of undulators and X-ray optical elements. Material properties of crystals, compound refractive lenses, and some other optical elements can be dynamically obtained for the desired photon energy from the databases publicly available at Argonne National Lab and at Lawrence Berkeley Lab. In collaboration with the Center for Functional Nanomaterials (CFN) of BNL, a library of samples for coherent scattering experiments has been implemented in SRW and the corresponding Sample optical element was added to Sirepo. Electron microscope images of artificially created nanoscale samples can be uploaded to Sirepo to simulate scattering patterns created by synchrotron radiation in different experimental schemes that can be realized at beamlines. 14. System modeling of spent fuel transfers at EBR-II Energy Technology Data Exchange (ETDEWEB) Imel, G.R. [Argonne National Lab., Idaho Falls, ID (United States); Houshyar, A. [Western Michigan Univ., Kalamazoo, MI (United States). Industrial Engineering Dept. 1994-12-31 15. 75 FR 34988 - Notice of Intent To Grant Exclusive Patent License; Fiber Optic Sensor Systems Technology... Science.gov (United States) 2010-06-21 ... Department of the Navy Notice of Intent To Grant Exclusive Patent License; Fiber Optic Sensor Systems... Navy hereby gives notice of its intent to grant to Fiber Optic Sensor Systems Technology Corporation a.... Patent No. 7,149,374: Fiber Optic Pressure Sensor, Navy Case No. 84,557.//U.S. Patent No. 7,379,630... 16. Synthesis, structural, optical band gap and biological studies on iron (III), nickel (II), zinc (II) and mercury (II) complexes of benzyl α-monoxime pyridyl thiosemicarbazone Science.gov (United States) Bedier, R. A.; Yousef, T. A.; Abu El-Reash, G. M.; El-Gammal, O. A. 2017-07-01 New ligand, (E)-2-((E)-2-(hydroxyimino)-1,2-diphenylethylidene)-N-(pyridin-2 yl) hydrazinecarbothioamide (H2DPPT) and its complexes [Fe(DPPT)Cl(H2O)], [Ni(H2DPPT)2Cl2], [Zn(HDPPT)(OAc)] and [Hg(HDPPT)Cl](H2O)4 were isolated and characterized by various of physico-chemical techniques. IR spectra show that H2DPPT coordinates to the metal ions as neutral NN bidentate, mononegative NNS tridentate and binegative NNSN tetradentate, respectively. From the modeling studies, the bond length, bond angle, HOMO, LUMO and dipole moment had been calculated to confirm the geometry of the ligands and their investigated complexes. The thermal studies showed the type of water molecules involved in metal complexes Furthermore, the kinetic and thermodynamic parameters for the different decomposition steps were calculated using the Coats-Redfern and Horowitz-Metzger methods. Also, the optical band gap (Eg) has been calculated to elucidate the conductivity of the isolated complexes. The optical transition energy (Eg) is direct and equals 3.34 and 3.44 ev for Ni and Fe complexes, respectively. The ligand and their metal complexes were screened for antibacterial activity against the following bacterial species, Bacillus thuringiensis, Staphylococcus aureus, Pseudomonas aeuroginosa and Escherichia coli. The results revealed that the metal complexes have more potent antibacterial compared with the ligand. Also, the degradation effect of the investigated compounds was tested showing that, Ni complex exhibited powerful and complete degradation effect on DNA. 17. The CaSSIS imaging system: optical performance overview Science.gov (United States) Gambicorti, L.; Piazza, D.; Pommerol, A.; Roloff, V.; Gerber, M.; Ziethe, R.; El-Maarry, M. R.; Weigel, T.; Johnson, M.; Vernani, D.; Pelo, E.; Da Deppo, V.; Cremonese, G.; Ficai Veltroni, I.; Thomas, N. 2016-07-01 The Colour and Stereo Surface Imaging System (CaSSIS) is the high-resolution scientific imager on board the European Space Agency's (ESA) ExoMars Trace Gas Orbiter (TGO) which was launched on 14th March 2016 to Mars. CaSSIS will observe the Martian surface from an altitude of 400 km with an optical system based on a modified TMA telescope (Three Mirrors Anastigmatic configuration) with a 4th powered folding mirror. The camera EPD (Entrance Pupil Diameter) is 135 mm, and the expected focal length is 880 mm, giving an F# 6.5 in the wavelength range of 400- 1100 nm with a distortion designed to be less than 2%. CaSSIS will operate in a "push-frame" mode with a monolithic Filter Strip Assembly (FSA) produced by Optics Balzers Jena GmbH selecting 4 colour bands and integrated on the focal plane by Leonardo-Finmeccanica SpA (under TAS-I responsibility). The detector is a spare of the Simbio-Sys detector of the Italian Space Agency (ASI), developed by Raytheon Vision Systems. It is a 2kx2k hybrid Si-PIN array with a 10 μm pixel pitch. A scale of 4.6 m/px from the nominal orbit is foreseen to produce frames of 9.4 km × 47 km on the Martian surface. The University of Bern was in charge of the full instrument integration as well as the characterization of the focal plane and calibration of the entire instrument. The paper will present an overview of the CaSSIS telescope and FPA optical performance. The preliminary results of on-ground calibration and the first commissioning campaign (April 2016) will be described. 18. Bedside arterial blood gas monitoring system using fluorescent optical sensors Science.gov (United States) Bartnik, Daniel J.; Rymut, Russell A. 1995-05-01 We describe a bedside arterial blood gas (ABG) monitoring system which uses fluorescent optical sensors in the measurement of blood pH, PCO2 and PO2. The Point-of-Care Arterial Blood Gas Monitoring System consists of the SensiCathTM optical sensor unit manufactured by Optical Sensors Incorporated and the TramTM Critical Care Monitoring System with ABG Module manufactured by Marquette Electronics Incorporated. Current blood gas measurement techniques require a blood sample to be removed from the patient and transported to an electrochemical analyzer for analysis. The ABG system does not require removal of blood from the patient or transport of the sample. The sensor is added to the patient's existing arterial line. ABG measurements are made by drawing a small blood sample from the arterial line in sufficient quantity to ensure an undiluted sample at the sensor. Measurements of pH, PCO2 and PO2 are made within 60 seconds. The blood is then returned to the patient, the line flushed and results appear on the bedside monitor. The ABG system offers several advantages over traditional electrochemical analyzers. Since the arterial line remains closed during the blood sampling procedure the patient's risk of infection is reduced and the caregiver's exposure to blood is eliminated. The single-use, disposable sensor can be measure 100 blood samples over 72 hours after a single two-point calibration. Quality Assurance checks are also available and provide the caregiver the ability to assess system performance even after the sensor is patient attached. The ABG module integrates with an existing bedside monitoring system. This allows ABG results to appear on the same display as ECG, respiration, blood pressure, cardiac output, SpO2, and other clinical information. The small module takes up little space in the crowded intensive care unit. Performance studies compare the ABG system with an electrochemical blood gas analyzer. Study results demonstrated accurate and precise blood 19. Optical response of Ce(III and Eu(II doped hybrid materials synthesised by Sol-Gel processing Directory of Open Access Journals (Sweden) Cordoncillo, E. 2000-02-01 Full Text Available This study deals with the preparation of two hybrid organic-inorganic matrices via sol–gel, starting from alkylalkoxisilanes Si(CH3(OCH2CH33 (MTEOS–SiH(CH3(OCH2CH32 (MDES, A system, and SiH(CH3(OCH2CH32 (MDES SiH(OCH2CH33 (TREOS, B system, together with zirconium n-propoxide. Another type-A is carried out by adding acetylacetone, A-acac system. The matrices are characterised by infrared spectroscopy, nuclear magnetic resonance (NMR-MAS, and chemical analysis. Optical characteristics of the matrices have been studied. A-acac and B matrices are doped with an Eu(III salt, and A and B matrices are doped with a Ce(IV salt. Absorption and emission studies show the presence of Eu(II and Ce(III. The transition metal alkoxide that catalysed cleavage of the Si–H bonds was used to reduce in situ at room temperature, the rare earth cations. Depending on chemical strategy, the resulting hybrid materials can be processed as transparent bulks or coatings which exhibit a good transparency in the UV–visible domain. Both the undoped and the rare earth doped matrices exhibit a strong blue emission. En este trabajo se aborda la preparación de dos matrices híbridas orgánico-inorgánicas por vía sol-gel, a partir de mezclas de alquilalcoxisilanos Si(CH3(OCH2CH33 (MTEOS–SiH(CH3(OCH2CH32 (MDES, sistema A, y SiH(CH3(OCH2CH32 (MDES–SiH(OCH2CH33 (TREOS, sistema B, en presencia de n-propóxido de circonio. Se efectúa una variación del sistema A por adición de acetilacetona, sistema A-acac. Las matrices se caracterizan por espectroscopia infrarroja, resonancia magnética nuclear (RMN-MAS y análisis químico. Se estudian las características ópticas de los materiales obtenidos. Las matrices A-acac y B se dopan con una sal de Eu(III y las matrices A y B con una sal de Ce(IV. Los estudios de absorción y emisión indican la presencia de Eu(II y Ce(III, es decir estos estados de oxidación se han generado in situ a temperatura ambiente en los 20. Recent progress in multidimensional optical sensing and imaging systems (MOSIS) Science.gov (United States) Shen, Xin; Javidi, Bahram 2017-05-01 We present recent progress of the previously reported Multidimensional Optical Sensing and Imaging Systems (MOSIS) 2.0 for target recognition, material inspection and integrated visualization. The degrees of freedom of MOSIS 2.0 include three-dimensional (3D) imaging, polarimetric imaging and multispectral imaging. Each of these features provides unique information about a scene. 3D computationally reconstructed images mitigate the occlusion in front of the object, which can be used for 3D object recognition. The degree of polarization (DoP) of the light reflected from object surface is measured by 3D polarimetric imaging. Multispectral imaging is able to segment targets with specific spectral properties. 1. Wavefront reconstruction in adaptive optics systems using nonlinear multivariate splines. Science.gov (United States) de Visser, Cornelis C; Verhaegen, Michel 2013-01-01 This paper presents a new method for zonal wavefront reconstruction (WFR) with application to adaptive optics systems. This new method, indicated as Spline based ABerration REconstruction (SABRE), uses bivariate simplex B-spline basis functions to reconstruct the wavefront using local wavefront slope measurements. The SABRE enables WFR on nonrectangular and partly obscured sensor grids and is not subject to the waffle mode. The performance of SABRE is compared to that of the finite difference (FD) method in numerical experiments using data from a simulated Shack-Hartmann lenslet array. The results show that SABRE offers superior reconstruction accuracy and noise rejection capabilities compared to the FD method. 2. A beamforming system based on the acousto-optic effect DEFF Research Database (Denmark) Torras Rosell, Antoni; Barrera Figueroa, Salvador; Jacobsen, Finn 2012-01-01 Beamforming techniques are usually based on microphone arrays. The present work uses a beam of light as a sensor element, and describes a beamforming system that locates sound sources based on the acousto-optic effect, this is, the interaction between sound and light. The use of light as a sensing...... element makes this method immune to spatial aliasing. This feature is illustrated by means of simulation and experimental results. For ease of comparison, the study is supplemented with results obtained with a line array of microphones... 3. Phase recovery algorithm in coherent homodyne optical systems Science.gov (United States) Ibragimov, R. Z. 2017-04-01 In paper we present low-complexity feed-forward algorithm for digital signal processing in coherent optical 16QAM systems, employing adaptive filtering for decreasing linewidth phase noise of coherent laser. Simulations show performance that using dual polarization for 16 quadrature-amplitude modulation (QAM) transmission is possible for laser beat linewidths of ΔfTs = 0.3 • 10-3 and ΔfTs = 7.9 • 10-5 when soft-decision phase estimator is employed. 4. Compact microscope-based optical tweezers system for molecular manipulation Science.gov (United States) Sischka, Andy; Eckel, Rainer; Toensing, Katja; Ros, Robert; Anselmetti, Dario 2003-11-01 A compact single beam optical tweezers system for force measurements and manipulation of individual double-stranded deoxyribonucleic acid (DNA) molecules was integrated into a commercial inverted optical microscope. A maximal force of 150 pN combined with a force sensitivity of less than 0.5 pN allows measurements of elastic properties of single molecules which complements and overlaps the force regime accessible with atomic force microscopy (AFM). The manipulation and measurement performance of this system was tested with individual λ-DNA molecules and renders new aspects of dynamic forces phenomena with higher precision in contrast to AFM studies. An integrated liquid handling system with a fluid cell allows investigation of the force response of individual DNA molecules in the presence of DNA binding agents. Comparison of YOYO-1-, ethidium bromide intercalated DNA, and distamycin-A complexed DNA revealed accurate and reproducible differences in the force response to an external load. This opens the possibility to use it as a single molecule biosensor to investigate DNA binding agents and even to identify molecular binding mechanisms. 5. Ultrasensitive detection system for fiber optic-based ultraviolet spectroscopy Science.gov (United States) Klein, Karl-Friedrich; Belz, Mathias; Dress, Peter; Schelle, B.; Boyle, William J. O.; Grattan, Kenneth T. V.; Franke, Hilmar 1998-03-01 A fiber optic based deep UV-absorption sensor system is characterized, using fibers for light delivery and a liquid core waveguide (LCW) for analyzing liquids. UN-improved fibers with 500 micrometers core diameter are capable of transmitting light intensities below 230 nm with spectral radiant powers above 500 nW/nm at 214 nm. Their short-term behavior and lifetime in respect to UV-stability have been investigated, using a broadband deuterium lamp. To raise the sensitivity of the total system, the absorption path length has been increased significantly using the lightguiding properties of the LCW consisting of a cylindrical glass tube with a Teflon AF 2400 inner coating of about 50 micrometers thickness. Due to lower refractive index of Teflon in comparison to water, the LCW concept offers significant advantages, especially for long optical pathlengths. However, the basic attenuation of the liquid in dependence on the wavelength as to be taken into account. Results on the use of such a system monitoring concentrations of acetylsalicylic acid, acetone and toluene in water are reported and discussed. 6. A new optical encryption system for image transformation Science.gov (United States) Yao, Shuyu; Chen, Linfei; Chang, Guojun; He, Bingyu 2017-12-01 This paper introduces a new optical image encryption system based on Fresnel diffraction and phase iterative algorithm, which can realize the conversion between different images. The method is based on the optical system of free space transmission, and uses the iterative phase retrieval algorithm to encode an image into two phase masks and a ciphertext. Unlike the existed methods, the ciphertext is a visible image, which can be used to achieve the conversion of one image to another image. In order to enhance the security, two phase masks are combined into a wide-scale phase mask by the double image cross pixel scrambling approach. In the decryption process, the wide-scale phase mask is re-decrypted into two random phase masks using a random shift matrix. The ciphertext and the first phase mask are placed on the input plane and the second random phase mask is placed on the transformation plane. The Fresnel diffraction principle can be used to obtain the plaintext information on the output plane. Theoretical analysis and simulation results show that the encryption system is feasible and quite safe. 7. New education system for construction of optical holography setup - Tangible learning with Augmented Reality Science.gov (United States) Yamaguchi, Takeshi; Yoshikawa, Hiroshi 2013-02-01 In case of teaching optical system construction, it is difficult to prepare the optical components for the attendance student. However the tangible learning is very important to master the optical system construction. It helps learners understand easily to use an inexpensive learning system that provides optical experiments experiences. Therefore, we propose the new education system for construction of optical setup with the augmented reality. To use the augmented reality, the proposed system can simulate the optical system construction by the direct hand control. Also, this system only requires an inexpensive web camera, printed makers and a personal computer. Since this system does not require the darkroom and the expensive optical equipments, the learners can study anytime, anywhere when they want to do. In this paper, we developed the system that can teach the optical system construction of the Denisyuk hologram and 2-step transmission type hologram. For the tangible learning and the easy understanding, the proposed system displays the CG objects of the optical components on the markers which are controlled by the learner's hands. The proposed system does not only display the CG object, but also display the light beam which is controlled by the optical components. To display the light beam that is hard to be seen directly, the learners can confirm about what is happening by the own manipulation. For the construction of optical holography setup, we arrange a laser, mirrors, a PBS (polarizing beam splitter), lenses, a polarizer, half-wave plates, spatial filters, an optical power meter and a recording plate. After the construction, proposed system can check optical setup correctly. In comparison with the learners who only read a book, the learners who use the system can construct the optical holography setup more quickly and correctly. 8. Development of Diesel Exhaust Aftertreatment System for Tier II Emissions Energy Technology Data Exchange (ETDEWEB) Yu, R. C.; Cole, A. S., Stroia, B. J.; Huang, S. C. (Cummins, Inc.); Howden, Kenneth C.; Chalk, Steven (U.S. Dept. of Energy) 2002-06-01 9. The CMS Level-1 trigger system for LHC Run II Science.gov (United States) 2017-03-01 The Compact Muon Solenoid (CMS) experiment implements a sophisticated two-level online selection system that achieves a rejection factor of nearly 105. During Run II, the LHC has increased the centre-of-mass energy of proton-proton collisions up to 13 TeV and may progressively reach an instantaneous luminosity of 2×1034 cm-2 s-1 or higher. In order to guarantee a successful and ambitious physics programme under this intense environment, the CMS Trigger and Data acquisition system has been upgraded. The upgraded CMS Level-1 (L1) trigger benefits from the recent μTCA technology and is designed to maintain the performance under high instantaneous luminosity conditions. More sophisticated, innovative algorithms are now the core of the first decision layer of CMS: this drastically reduces the trigger rate and improves the trigger efficiency for a wide variety of physics processes. In this document, we present the overall architecture of the upgraded Level-1 trigger system. The performance of single object triggers, measured on collision data recorded during the 2016 running period, are also summarised. 10. Angiotensin II and insulin crosstalk in the cardiovascular system. Science.gov (United States) Liu, Zhen-qi 2006-12-01 Under normal physiology, insulin exerts vasodilatory and pro-survival actions via the phosphatidylinositol 3-kinase (PI3-kinase) pathway and vasoconstrictive and mitogenic actions via the mitogen-activated protein kinase (MAPK) pathway in the vasculature. In the insulin resistant states, insulin signals through the PI3-kinase pathway are blunted but its signals through the MAPK cascade remain intact. This imbalance predisposes insulin resistant patients to hypertension and atherosclerosis. The renin-angiotensin system (RAS) is expressed both systemically and locally in the cardiovascular system. Insulin resistance up-regulates the local RAS which contributes to the pathogenesis of hypertension, heart failure, and atherosclerosis. Angiotensin II impairs insulin signaling, induces inflammation via the NF-kappaB pathway, reduces nitric oxide availability and facilitates vasoconstriction, leading to insulin resistance and endothelial dysfunction. Thus the RAS, insulin resistance and inflammation perpetuate each other and coordinately contribute to endothelial dysfunction, vascular injury and atherosclerosis. RAS inhibition decreases cardiovascular and renal morbidity and mortality and the incidence of new onset Type 2 diabetes. 11. Performance of the Keck Observatory adaptive optics system Energy Technology Data Exchange (ETDEWEB) van Dam, M A; Mignant, D L; Macintosh, B A 2004-01-19 In this paper, the adaptive optics (AO) system at the W.M. Keck Observatory is characterized. The authors calculate the error budget of the Keck AO system operating in natural guide star mode with a near infrared imaging camera. By modeling the control loops and recording residual centroids, the measurement noise and band-width errors are obtained. The error budget is consistent with the images obtained. Results of sky performance tests are presented: the AO system is shown to deliver images with average Strehl ratios of up to 0.37 at 1.58 {micro}m using a bright guide star and 0.19 for a magnitude 12 star. 12. Design of optical system for binocular fundus camera. Science.gov (United States) Wu, Jun; Lou, Shiliang; Xiao, Zhitao; Geng, Lei; Zhang, Fang; Wang, Wen; Liu, Mengjia 2017-12-01 A non-mydriasis optical system for binocular fundus camera has been designed in this paper. It can capture two images of the same fundus retinal region from different angles at the same time, and can be used to achieve three-dimensional reconstruction of fundus. It is composed of imaging system and illumination system. In imaging system, Gullstrand Le Grand eye model is used to simulate normal human eye, and Schematic eye model is used to test the influence of ametropia in human eye on imaging quality. Annular aperture and black dot board are added into illumination system, so that the illumination system can eliminate stray light produced by corneal-reflected light and omentoscopic lens. Simulation results show that MTF of each visual field at the cut-off frequency of 90lp/mm is greater than 0.2, system distortion value is -2.7%, field curvature is less than 0.1 mm, radius of Airy disc is 3.25um. This system has a strong ability of chromatic aberration correction and focusing, and can image clearly for human fundus in which the range of diopters is from -10 D to +6 D(1 D = 1 m -1 ). 13. Machine learning techniques for optical communication system optimization DEFF Research Database (Denmark) Zibar, Darko; Wass, Jesper; Thrane, Jakob In this paper, machine learning techniques relevant to optical communication are presented and discussed. The focus is on applying machine learning tools to optical performance monitoring and performance prediction.......In this paper, machine learning techniques relevant to optical communication are presented and discussed. The focus is on applying machine learning tools to optical performance monitoring and performance prediction.... 14. Comparative experiments between a conventional piezo hydrophone system and a fibre optic hydrophone system NARCIS (Netherlands) Cheng, L.K.; Bruijn, D. de 2000-01-01 We report here the results of comparative measurements between Fibre Optic (FO) hydrophone system with our Hydrostatic Pressure Conmpensation (HPC) mechanism and a conventional piezo hydrophone system (consisting of an ITC 1042 hydrophone and a B&K2035 Analyser). The sensitivity and the dynamic 15. Investigation of the limits of a fibre optic sensor system for measurement of temperature distribution DEFF Research Database (Denmark) Brehm, Robert; Johnson, Frank 2006-01-01 The aim of this project is to develop an innovative temperature sensor system which is able to measure the temperature distribution along a fibre optical cable. This technique for temperature measurement is based on Optical Time Domain Reflectometry (OTDR).......The aim of this project is to develop an innovative temperature sensor system which is able to measure the temperature distribution along a fibre optical cable. This technique for temperature measurement is based on Optical Time Domain Reflectometry (OTDR).... 16. Preparation of a novel pH optical sensor using orange (II) based on agarose membrane as support. Science.gov (United States) 2016-04-01 A novel and cost effective optical pH sensor was prepared using covalent immobilization of orange (II) indicator on the agarose membrane as solid support. The fabricated optical sensor was fixed into a sample holder of a spectrophotometer instrument for pH monitoring. Variables affecting sensor performance including pH of dye bonding to agarose membrane and dye concentration were optimized. The sensor responds to the pH changes in the range of 3.0-10.0 with a response time of 2.0 min and appropriate reproducibility (RSD ≤ 0.9%). No significant variation was observed on sensor response after increasing the ionic strength in the range of 0.0-0.5M of sodium chloride. Determination of pH using the proposed optical sensor is quick, simple, inexpensive, selective and sensitive in the pH range of 3.0-10.0. Copyright © 2015 Elsevier B.V. All rights reserved. 17. Region-of-interest diffuse optical tomography system Energy Technology Data Exchange (ETDEWEB) Saikia, Manob Jyoti; Kanhirodan, Rajan, E-mail: rajan@physics.iisc.ernet.in [Department of Physics, Indian Institute of Science, Bangalore 560012 (India) 2016-01-15 Diffuse optical tomography (DOT) using near-infrared light is a promising tool for non-invasive imaging of deep tissue. This technique is capable of quantitative reconstruction of absorption (μ{sub a}) and scattering coefficient (μ{sub s}) inhomogeneities in the tissue. The rationale for reconstructing the optical property map is that the absorption coefficient variation provides diagnostic information about metabolic and disease states of the tissue. The aim of DOT is to reconstruct the internal tissue cross section with good spatial resolution and contrast from noisy measurements non-invasively. We develop a region-of-interest scanning system based on DOT principles. Modulated light is injected into the phantom/tissue through one of the four light emitting diode sources. The light traversing through the tissue gets partially absorbed and scattered multiple times. The intensity and phase of the exiting light are measured using a set of photodetectors. The light transport through a tissue is diffusive in nature and is modeled using radiative transfer equation. However, a simplified model based on diffusion equation (DE) can be used if the system satisfies following conditions: (a) the optical parameter of the inhomogeneity is close to the optical property of the background, and (b) μ{sub s} of the medium is much greater than μ{sub a} (μ{sub s} > > μ{sub a}). The light transport through a highly scattering tissue satisfies both of these conditions. A discrete version of DE based on finite element method is used for solving the inverse problem. The depth of probing light inside the tissue depends on the wavelength of light, absorption, and scattering coefficients of the medium and the separation between the source and detector locations. Extensive simulation studies have been carried out and the results are validated using two sets of experimental measurements. The utility of the system can be further improved by using multiple wavelength light sources. In such 18. Dielectric and electro-optical properties of polymer-stabilized liquid crystal. II. Polymer PiBMA dispersed in MBBA Science.gov (United States) Tripathi, Pankaj Kumar; Pande, Mukti; Singh, Shri 2016-09-01 In continuation of our earlier work (Pande et al. in Appl Phys A 122:217-226, 2016), we report the results of dielectric and electro-optical properties of pure MBBA and PSLC (polymer PiBMA dispersed in MBBA) systems. The polymer networks domains formed are found to be anisotropic and are oriented in the direction of electric field for both the planar and homeotropic alignment cells. The dielectric anisotropy, optical anisotropy, response time, threshold voltage, splay elastic constant and rotational viscosity were observed for both the LC and PSLC systems with electric field. The liquid crystal properties are affected significantly with increasing concentration of polymer in pure LC material. It has been observed that the polymer networks interaction plays a major role in changing the properties of PSLC system. The effect of dielectric loss and dielectric permittivity on both pure LC and PSLC systems is also discussed. 19. Optic disc segmentation for glaucoma screening system using fundus images Directory of Open Access Journals (Sweden) Almazroa A 2017-11-01 Full Text Available Ahmed Almazroa,1,2 Weiwei Sun,3 Sami Alodhayb,4 Kaamran Raahemifar,5 Vasudevan Lakshminarayanan6 1King Abdullah International Medical Research Center (KAIMRC, Riyadh, Saudi Arabia; 2Ophthalmology and Visual Science Department, University of Michigan, Ann Arbor, MI, USA; 3School of Resource and Environmental Sciences, Wuhan University, Wuchang, Wuhan, Hubei, China; 4Bin Rushed Ophthalmic Center, Riyadh, Saudi Arabia; 5Department of Electrical and Computer Engineering, University of Ryerson, Toronto, ON, 6School of Optometry, University of Waterloo, ON, Canada Abstract: Segmenting the optic disc (OD is an important and essential step in creating a frame of reference for diagnosing optic nerve head pathologies such as glaucoma. Therefore, a reliable OD segmentation technique is necessary for automatic screening of optic nerve head abnormalities. The main contribution of this paper is in presenting a novel OD segmentation algorithm based on applying a level set method on a localized OD image. To prevent the blood vessels from interfering with the level set process, an inpainting technique was applied. As well an important contribution was to involve the variations in opinions among the ophthalmologists in detecting the disc boundaries and diagnosing the glaucoma. Most of the previous studies were trained and tested based on only one opinion, which can be assumed to be biased for the ophthalmologist. In addition, the accuracy was calculated based on the number of images that coincided with the ophthalmologists’ agreed-upon images, and not only on the overlapping images as in previous studies. The ultimate goal of this project is to develop an automated image processing system for glaucoma screening. The disc algorithm is evaluated using a new retinal fundus image dataset called RIGA (retinal images for glaucoma analysis. In the case of low-quality images, a double level set was applied, in which the first level set was considered to be localization 20. Jacobian and Hessian matrices of optical path length for computing the wavefront shape, irradiance, and caustics in optical systems. Science.gov (United States) Lin, Psang Dain; Liu, Chien-Sheng 2012-11-01 The first- and second-order derivative matrices of the ray (i.e., ∂R¯(i)/∂X¯(0) and ∂(2)R¯(i)/∂X¯02) and optical path length (i.e., ∂OPL(i)/∂X¯(0) and (2)OPL(i)/∂X¯02) were derived with respect to the variable vector X¯(0) of the source ray in an optical system by our previous papers. Using the first and second fundamental forms of the wavefront, these four matrices are used to investigate the local principal curvatures of the wavefront at each boundary surface encountered by a ray traveling through the optical system. The proposed method not only yields the data needed to compute the irradiance of the wavefront but also provides the information required to determine the caustics. Importantly, the proposed methodology is applicable to both axisymmetric and nonaxisymmetric optical systems. 1. Page 1 i. Optic Interference Figures of Amethystine Quartz—Part II ... ... amethyst in any random section plate is that optical twinning of an interpenetrant variety occurs. * Brewster's brushes had been referred to in our earlier paper” as the idiophanous figure. —a term reserved in the present paper for the figure presented with a polariser or analyser alone in conformity with the more prevalent ... 2. Nonterminals, homomorphisms and codings in different variations of OL-systems. II. Nondeterministic systems DEFF Research Database (Denmark) Nielsen, Mogens; Rozenberg, Grzegorz; Salomaa, Arto 1974-01-01 Continuing the work begun in Part I of this paper, we consider now variations of nondeterministic OL-systems. The present Part II of the paper contains a systematic classification of the effect of nonterminals, codings, weak codings, nonerasing homomorphisms and homomorphisms for all basic variat... 3. Theory of aberration fields for general optical systems with freeform surfaces. Science.gov (United States) Fuerschbach, Kyle; Rolland, Jannick P; Thompson, Kevin P 2014-11-03 This paper utilizes the framework of nodal aberration theory to describe the aberration field behavior that emerges in optical systems with freeform optical surfaces, particularly φ-polynomial surfaces, including Zernike polynomial surfaces, that lie anywhere in the optical system. If the freeform surface is located at the stop or pupil, the net aberration contribution of the freeform surface is field constant. As the freeform optical surface is displaced longitudinally away from the stop or pupil of the optical system, the net aberration contribution becomes field dependent. It is demonstrated that there are no new aberration types when describing the aberration fields that arise with the introduction of freeform optical surfaces. Significantly it is shown that the aberration fields that emerge with the inclusion of freeform surfaces in an optical system are exactly those that have been described by nodal aberration theory for tilted and decentered optical systems. The key contribution here lies in establishing the field dependence and nodal behavior of each freeform term that is essential knowledge for effective application to optical system design. With this development, the nodes that are distributed throughout the field of view for each aberration type can be anticipated and targeted during optimization for the correction or control of the aberrations in an optical system with freeform surfaces. This work does not place any symmetry constraints on the optical system, which could be packaged in a fully three dimensional geometry, without fold mirrors. 4. Expert system applications in support of system diagnostics and prognostics at EBR-II Energy Technology Data Exchange (ETDEWEB) Lehto, W.K.; Gross, K.C. (Argonne National Lab., Idaho Falls, ID (USA); Argonne National Lab., IL (USA)) 1989-01-01 Expert systems have been developed to aid in the monitoring and diagnostics of the Experimental Breeder Reactor-II (EBR-II) at the Idaho National Engineering Laboratory (INEL) in Idaho Falls, Idaho. Systems have been developed for failed fuel surveillance and diagnostics and reactor coolant pump monitoring and diagnostics. A third project is being done jointly by ANL-W and EG G Idaho to develop a transient analysis system to enhance overall plant diagnostic and prognostic capability. The failed fuel surveillance and diagnosis system monitors, processes, and interprets information from nine key plant sensors. It displays to the reactor operator diagnostic information needed to make proper decisions regarding technical specification conformance during reactor operation with failed fuel. 8 refs., 9 figs., 2 tabs. 5. Ultrawideband photonic crystal fiber coupler for multiband optical imaging system. Science.gov (United States) Ryu, Seon Young; Choi, Hae Young; Choi, Eun Seo; Tomov, Ivan; Chen, Zhongping; Lee, Byeong Ha 2010-04-01 We report a photonic crystal fiber (PCF) coupler having an ultrawide spectral bandwidth keeping single mode operation. The use of the PCF coupler in a fiber-based optical coherence tomography (OCT) system enables us to handle the wide spectral bands of various light sources, including superluminescent diodes (SLDs) at 1300 nm and 820 nm, Ti:sapphire lasers, and white-light sources. The multiband imaging performances of the PCF-based OCT system are demonstrated by obtaining dental images at 1300 nm and 820 nm with the same setup. In addition, we show that the PCF coupler could cover the spectrum over a one octave span and guide both the fundamental wave (1030 nm) and the second harmonic wave (515 nm) simultaneously. 6. Biological elements carry out optical tasks in coherent imaging systems Science.gov (United States) Ferraro, P.; Bianco, V.; Paturzo, M.; Miccio, L.; Memmolo, P.; Merola, F.; Marchesano, V. 2016-03-01 We show how biological elements, like live bacteria species and Red Blood Cells (RBCs) can accomplish optical functionalities in DH systems. Turbid media allow coherent microscopy despite the strong light scattering these provoke, acting on light just as moving diffusers. Furthermore, a turbid medium can have positive effects on a coherent imaging system, providing resolution enhancement and mimicking the action of noise decorrelation devices, thus yielding an image quality significantly higher than the quality achievable through a transparent medium in similar recording conditions. Besides, suspended RBCs are demonstrated to behave as controllable liquid micro-lenses, opening new possibilities in biophotonics for endoscopy imaging purposes, as well as telemedicine for point-of-care diagnostics in developing countries and low-resource settings. 7. Evaluation of systems and components for hybrid optical firing sets Energy Technology Data Exchange (ETDEWEB) Landry, M.J.; Rupert, J.W.; Mittas, A. 1989-06-01 High-energy density light appears to be a unique energy form that may be used to enhance the nuclear safety of weapon systems. Hybrid optical firing sets (HOFS) utilize the weak-link/strong-link exclusion region concept for nuclear safety; this method is similar to present systems, but uses light to transmit power across the exclusion region barrier. This report describes the assembling, operating, and testing of fourteen HOFS. These firing sets were required to charge a capacitor-discharge unit to 2.0 and 2.5 kV (100 mJ) in less than 1 s. First, we describe the components, the measurement techniques used to evaluate the components, and the different characteristics of the measured components. Second, we describe the HOFS studied, the setups used for evaluating them, and the resulting characteristics. Third, we make recommendations for improving the overall performance and suggest the best HOFS for packaging. 36 refs., 145 figs., 14 tabs. 8. Photosynthetic antennae systems: energy transport and optical absorption Energy Technology Data Exchange (ETDEWEB) Reineker, P. E-mail: peter.reineker@physik.uni-ulm.de; Supritz, Ch.; Warns, Ch.; Barvik, I 2004-06-01 The energy transport and the optical line shape of molecular aggregates, modeling bacteria photosynthetic light-harvesting systems (chlorosomes in the case of Chlorobium tepidum or Chloroflexus aurantiacus and LH2 in the case of Rhodopseudomonas acidophila) is investigated theoretically. The molecular units are described by two-level systems with an average excitation energy {epsilon} and interacting with each other through nearest-neighbor interactions. For LH2 an elliptical deformation of the ring is also allowed. Furthermore, dynamic and in the case of LH2 also quasi-static fluctuations of the local excitation energies are taken into account, simulating fast molecular vibrations and slow motions of the protein backbone, respectively. The fluctuations are described by Gaussian Markov processes in the case of the chlorosomes and by colored dichotomic Markov processes, with exponentially decaying correlation functions, with small ({lambda}{sub s}) and large ({lambda}) decay constants, in the case of LH2. 9. Semiconductor Mode-Locked Lasers for Optical Communication Systems DEFF Research Database (Denmark) Yvind, Kresten 2003-01-01 The thesis deals with the design and fabrication of semiconductor mode-locked lasers for use in optical communication systems. The properties of pulse sources and characterization methods are described as well as requirements for application in communication systems. Especially, the importance of...... as the first 10 GHz all-active monolithic laser with both short pulses and low jitter.Results from external cavity mode-locked lasers are also reported along with an investigation of the influence of the operating conditions on the performance of the device. Antireflection coatings are a critical limiting...... factor for external cavity devices and a chapter is devoted to calculations on coating design.The fabrication process for the lasers is outlined and solutions to the challenges encountered in realizing the proposed structures are reported.... 10. Stabilized high-accuracy optical tracking system (SHOTS) Science.gov (United States) Ruffatto, Donald; Brown, H. Donald; Pohle, Richard H.; Reiley, Michael F.; Haddock, Delmar D. 2001-08-01 This paper describes an 0.75 meter aperture, Stabilized High-accuracy Optical Tracking System (SHOTS), two of which are being developed by Textron Systems Corporation, under contract to the Navy's Space and Naval Warfare Systems Center, San Diego (SPAWAR-SD). The SHOTS design is optimized to meet the requirements of the Navy's Theater Ballistic Missile Defense (TBMD) testing program being conducted at the Kauai Pacific Missile Range Facility (PMRF). The SHOTS utilizes a high-precision, GPS aided inertial navigation unit (INU) coupled with a 3-axis, rate gyro stabilized mount which allows precision pointing to be achieved on either land or sea-based platforms. The SHOTS mount control system architecture, acquisition, tracking and pointing (ATP) functionality and methodology which allows the system to meet the TBMD mission data collection requirements are discussed. High frame rate visible and MWIR sensors are incorporated into the system design to provide the capability of capturing short duration events, e.g., missile-target intercepts. These sensors along with the supporting high speed data acquisition, recording and control subsystems are described. Simulations of the SHOTS imaging performance in TBMD measurement scenarios are presented along with an example of the image improvement being achieved with post-processing image reconstruction algorithms. 11. Using a low-noise interferometric fiber optic gyro in a pointing, acquisition, and tracking system Science.gov (United States) Kaufmann, John; Hakimi, Farhad; Boroson, Don 2013-03-01 Heritage pointing, acquisition, and tracking (PAT) systems have relied on optical tracking with a cooperative remote terminal to stabilize the line-of-sight of optical communications links. A hybrid approach, using new interferometric fiberoptic gyro (IFOG) technology to sense and correct local angular disturbances, blended with optical tracking, is shown to yield two significant advantages over traditional all-optical tracking: (1) line-of-sight stabilization over a very wide disturbance frequency range, down to extremely low frequencies (architect an optimal hybrid IFOG/optical PAT system. In addition, flow-down benefits that can simplify PAT system hardware will be discussed. 12. MEMS tracking mirror system for a bidirectional free-space optical link. Science.gov (United States) Jeon, Sungho; Toshiyoshi, Hiroshi 2017-08-20 We report on a bidirectional free-space optical system that is capable of automatic connection and tracking of an optical link between two nodes. A piezoelectric micro-electro-mechanical systems (MEMS) optical scanner is used to steer a laser beam of two wavelengths superposed to visually present a communication zone, to search for the position of the remote node by means of the retro-reflector optics, and to transmit the data between the nodes. A feedback system is developed to control the MEMS scanner to dynamically establish the optical link within a 10-ms transition time and to keep track of the moving node. 13. MSM optical detector on the basis of II-type ZnSe/ZnTe superlattice Energy Technology Data Exchange (ETDEWEB) Kuznetzov, P. I., E-mail: pik218@ire216.msk.su; Averin, S. V., E-mail: sva278@ire216.msk.su; Zhitov, V. A.; Zakharov, L. Yu.; Kotov, V. M. [Russian Academy of Sciences, Kotel’nikov Institute of Radioengineering and Electronics (Fryazino Branch) (Russian Federation) 2017-02-15 On the basis of a type-II ZnSe/ZnTe superlattice, a MSM (metal—semiconductor–metal) photodetector is fabricated and investigated. The detector features low dark currents and a high sensitivity. The spectral characteristic of the detector provides the possibility of the selective detection of three separate spectral portions of visible and near-infrared radiation. 14. Optical designs of reflection and refraction collection optics for a JT-60SA core Thomson scattering system Science.gov (United States) Tojo, H.; Hatae, T.; Hamano, T.; Sakuma, T.; Itami, K. 2013-09-01 Collection optics for core measurements in a JT-60SA Thomson scattering system were designed. The collection optics will be installed in a limited space and have a wide field of view and wide wavelength range. Two types of the optics are therefore suggested: refraction and reflection types. The reflection system, with a large primary mirror, avoids large chromatic aberrations. Because the size limit of the primary mirror and vignetting due to the secondary mirror affect the total collection throughput, conditions that provide the high throughput are found through an optimization. A refraction system with four lenses forming an Ernostar system is also employed. The use of high-refractive-index glass materials enhances the freedom of the lens curvatures, resulting in suppression of the spherical and coma aberration. Moreover, sufficient throughput can be achieved, even with smaller lenses than that of a previous design given in [H. Tojo, T. Hatae, T. Sakuma, T. Hamano, K. Itami, Y. Aida, S. Suitoh, and D. Fujie, Rev. Sci. Instrum. 81, 10D539 (2010)]. The optical resolutions of the reflection and refraction systems are both sufficient for understanding the spatial structures in plasma. In particular, the spot sizes at the image of the optics are evaluated as {˜}0.3 mm and {˜}0.4 mm, respectively. The throughput for the two systems, including the pupil size and transmissivity, are also compared. The results show that good measurement accuracy (temperatures ({refraction system. 15. Interaction of Zn(II) with hematite nanoparticles and microparticles: Part 2. ATR-FTIR and EXAFS study of the aqueous Zn(II)/oxalate/hematite ternary system. Science.gov (United States) Ha, Juyoung; Trainor, Thomas P; Farges, François; Brown, Gordon E 2009-05-19 Sorption of Zn(II) to hematite nanoparticles (HN) (av diam=10.5 nm) and microparticles (HM) (av diam=550 nm) was studied in the presence of oxalate anions (Ox2-(aq)) in aqueous solutions as a function of total Zn(II)(aq) to total Ox2-(aq) concentration ratio (R=[Zn(II)(aq)]tot/[Ox2-(aq)]tot) at pH 5.5. Zn(II) uptake is similar in extent for both the Zn(II)/Ox/HN and Zn(II)/Ox/HM ternary systems and the Zn(II)/HN binary system at [Zn(II)(aq)](tot)system than for the Zn(II)/Ox/HM ternary and the Zn(II)/HN and Zn(II)/HM binary systems at [Zn(II)(aq)]tot>4 mM. In contrast, Zn(II) uptake for the Zn(II)/HM binary system is a factor of 2 greater than that for the Zn(II)/Ox/HM and Zn(II)/Ox/HN ternary systems and the Zn(II)/HN binary system at [Zn(II)(aq)]totternary system at both R values examined (0.16 and 0.68), attenuated total reflectance Fourier transform infrared (ATR-FTIR) results are consistent with the presence of inner-sphere oxalate complexes and outer-sphere ZnOx(aq) complexes, and/or type A ternary complexes. In addition, extended X-ray absorption fine structure (EXAFS) spectroscopic results suggest that type A ternary surface complexes (i.e., >O2-Zn-Ox) are present. In the Zn(II)/Ox/HN ternary system at R=0.15, ATR-FTIR results indicate the presence of inner-sphere oxalate and outer-sphere ZnOx(aq) complexes; the EXAFS results provide no evidence for inner-sphere Zn(II) complexes or type A ternary complexes. In contrast, ATR-FTIR results for the Zn/Ox/HN sample with R = 0.68 are consistent with a ZnOx(s)-like surface precipitate and possibly type B ternary surface complexes (i.e., >O2-Ox-Zn). EXAFS results are also consistent with the presence of ZnOx(s)-like precipitates. We ascribe the observed increase of Zn(II)(aq) uptake in the Zn(II)/Ox/HN ternary system at [Zn(II)(aq)]tot>or=4 mM relative to the Zn(II)/Ox/HM ternary system to formation of a ZnOx(s)-like precipitate at the hematite nanoparticle/water interface. 16. Development of the visual system of the chick. II. Mechanisms of axonal guidance. Science.gov (United States) Thanos, S; Mey, J 2001-07-01 The quest to understand axonal guidance mechanisms requires exact and multidisciplinary analyses of axon navigation. This review is the second part of an attempt to synthesise experimental data with theoretical models of the development of the topographic connection of the chick retina with the tectum. The first part included classic ideas from developmental biology and recent achievements on the molecular level in understanding cytodifferentiation and histogenesis [J. Mey, S. Thanos, Development of the visual system of the chick. (I) Cell differentiation and histogenesis, Brain Res. Rev. 32 (2000) 343-379]. The present part deals with the question of how millions of fibres exit from the eye, traverse over several millimetres and spread over the optic tectum to assemble a topographic map, whose precision accounts for the sensory performance of the visual system. The following topics gained special attention in this review. (i) A remarkable conceptual continuity between classic embryology and recent molecular biology has revealed that positional cellular specification precedes and determines the formation of the retinotectal map. (ii) Graded expression of asymmetric genes, transcriptional factors and receptors for signal transduction during early development seem to play a crucial role in determining the spatial identity of neurons within surface areas of retina and optic tectum. (iii) The chemoaffinity hypothesis constitutes the conceptual framework for development of the retinotopic organisation of the primary visual pathway. Studies of repulsive factors in vitro developed the original hypothesis from a theoretical postulate of chemoattraction to an empirically supported concept based on chemorepulsion. (iv) The independent but synchronous development of retina and optic tectum in topo-chronologically corresponding patterns ensures that ingrowing retinal axons encounter receptive target tissue at appropriate locations, and at the time when connections are due to 17. 78 FR 24061 - Minimum Technical Standards for Class II Gaming Systems and Equipment Science.gov (United States) 2013-04-24 ... Gaming Systems and Equipment AGENCY: National Indian Gaming Commission. ACTION: Final rule. SUMMARY: The... for Class II gaming systems and equipment to harmonize the charitable gaming exemption amount in the... guidance to equipment manufacturers and distributors of Class II gaming systems. The standards did not... 18. 78 FR 11795 - Minimum Technical Standards for Class II Gaming Systems and Equipment Science.gov (United States) 2013-02-20 ... Gaming Systems and Equipment AGENCY: National Indian Gaming Commission. ACTION: Notice of proposed... final rule amending its technical standards for Class II gaming systems and equipment, and the rule... in the Federal Register called Minimum Technical Standards for Class II Gaming Systems and Equipment... 19. POLARIS-II - An acousto-optic imaging spectropolarimeter for ground-based astronomy Science.gov (United States) Glenar, D. A.; Hillman, J. J.; Saif, B.; Bergstralh, J. 1992-01-01 A compact, acousto-optic tunable filter (AOTF) imaging spectropolarimeter for ground based astronomy from 400-1100 nm has been constructed at NASA/GSFC. The key components of this instrument are a TeO2 non-collinear AOTF, CCD camera, and an all-reflective optical relay assembly which uses a single elliptical mirror to produce side-by-side orthogonally polarized spectral images. The instrument was used at the Lowell Observatory 42-inch telescope for 'first light' planetary imaging and measurements of photometric standard stars. Narrow-band images of Saturn near 700 nm appear to show polarization effects which result from multiple scattering by aerosols. The instrument has recently been upgraded in order to integrate the RF drive electronics and eliminate contamination by scattered light. Design of the instrument and some initial results are presented. 20. Transition operators in electromagnetic-wave diffraction theory. II - Applications to optics Science.gov (United States) Hahne, G. E. 1993-01-01 The theory developed by Hahne (1992) for the diffraction of time-harmonic electromagnetic waves from fixed obstacles is briefly summarized and extended. Applications of the theory are considered which comprise, first, a spherical harmonic expansion of the so-called radiation impedance operator in the theory, for a spherical surface, and second, a reconsideration of familiar short-wavelength approximation from the new standpoint, including a derivation of the so-called physical optics method on the basis of quasi-planar approximation to the radiation impedance operator, augmented by the method of stationary phase. The latter includes a rederivation of the geometrical optics approximation for the complete Green's function for the electromagnetic field in the presence of a smooth- and a convex-surfaced perfectly electrically conductive obstacle. 1. Infrared Spectra and Optical Constants of Astronomical Ices: II. Ethane and Ethylene Science.gov (United States) Hudson, Reggie L.; Gerakines, Perry A.; Moore, M. H. 2014-01-01 Infrared spectroscopic observations have established the presence of hydrocarbon ices on Pluto and other TNOs, but the abundances of such molecules cannot be deduced without accurate optical constants (n, k) and reference spectra. In this paper we present our recent measurements of near- and mid-infrared optical constants for ethane (C2H6) and ethylene (C2H4) in multiple ice phases and at multiple temperatures. As in our recent work on acetylene (C2H2), we also report new measurements of the index of refraction of each ice at 670 nm. Comparisons are made to earlier work where possible, and electronic versions of our new results are made available. 2. Thermo-optically induced reorganizations in the main light harvesting antenna of plants. II DEFF Research Database (Denmark) Holm, Jens Kai; Varkonyi, Zsuzsanna; Kovacs, Laszlo 2005-01-01 We have investigated the circular dichroism spectral transients associated with the light-induced reversible reorganizations in chirally organized macrodomains of pea thylakoid membranes and loosely stacked lamellar aggregates of the main chlorophyll a/b light harvesting complexes (LHCII) isolate......, LHCII-only domains might play an important role in light adaptation and photoprotection of plants.......We have investigated the circular dichroism spectral transients associated with the light-induced reversible reorganizations in chirally organized macrodomains of pea thylakoid membranes and loosely stacked lamellar aggregates of the main chlorophyll a/b light harvesting complexes (LHCII) isolated...... from the same membranes. These reorganizations have earlier been assigned to originate from a thermo-optic effect. According to the thermo-optic mechanism, fast local thermal transients due to dissipation of the excess excitation energy induce elementary structural changes in the close vicinity... 3. Optical fibre dosemeter systems for clinical applications based on radioluminescence and optically stimulated luminescence from Al2O3:C DEFF Research Database (Denmark) Marckmann, C.J.; Andersen, C.E.; Aznar, M.C. 2006-01-01 Optical fibre dosemeter systems based on radioluminescence and optically stimulated luminescence (OSL) from carbon-doped aluminium oxide (Al2O3:C) crystals were developed for in vivo real-time dose rate and absorbed dose measurements in radiotherapy and mammography. A technique was also developed...... for making ultra-small dosemeter probes that can easily be placed inside patients in radiation treatment. These probes have shown excellent properties in both head and neck intensity-modulated radiation therapy treatment and in mammography. The dose-response of the OSL signal for the new optical fibre... 4. Optical key distribution system using atmospheric turbulence as the randomness generating function: classical optical protocol for information assurance Science.gov (United States) Drake, Marvin D.; Bas, Christophe F.; Gervais, David; Renda, Priscilla F.; Townsend, Daniel; Rushanan, Joseph J.; Francoeur, Joe; Donnangelo, Nick; Stenner, Michael D. 2013-05-01 We describe an experimental laboratory system that generates and distributes random binary sequence bit streams between two optical terminals (labeled Alice and Bob). The random binary sequence is generated through probing the optical channel of a turbulent atmosphere between the two terminals with coincident laser beams. The two laser beams experience differential phase delays while propagating through the atmospheric optical channel. The differential phase delays are detected and sampled at each terminal to yield raw random bit streams. The random bit streams are processed to remove bit errors and, through privacy amplification, to yield a bit stream known only to Alice and Bob. The same chaotic physical mechanism that provides randomness also provides confidentiality. The laboratory system yielded secret key bit rates of a few bits/second. For external optical channels over longer channel lengths with atmospheric turbulence levels, secret bit rates of 10 s of bits/second are predicted. 5. Optical interconnection networks for high-performance computing systems. Science.gov (United States) Biberman, Aleksandr; Bergman, Keren 2012-04-01 Enabled by silicon photonic technology, optical interconnection networks have the potential to be a key disruptive technology in computing and communication industries. The enduring pursuit of performance gains in computing, combined with stringent power constraints, has fostered the ever-growing computational parallelism associated with chip multiprocessors, memory systems, high-performance computing systems and data centers. Sustaining these parallelism growths introduces unique challenges for on- and off-chip communications, shifting the focus toward novel and fundamentally different communication approaches. Chip-scale photonic interconnection networks, enabled by high-performance silicon photonic devices, offer unprecedented bandwidth scalability with reduced power consumption. We demonstrate that the silicon photonic platforms have already produced all the high-performance photonic devices required to realize these types of networks. Through extensive empirical characterization in much of our work, we demonstrate such feasibility of waveguides, modulators, switches and photodetectors. We also demonstrate systems that simultaneously combine many functionalities to achieve more complex building blocks. We propose novel silicon photonic devices, subsystems, network topologies and architectures to enable unprecedented performance of these photonic interconnection networks. Furthermore, the advantages of photonic interconnection networks extend far beyond the chip, offering advanced communication environments for memory systems, high-performance computing systems, and data centers. 6. EBR-II argon cooling system restricted fuel handling I and C upgrade Energy Technology Data Exchange (ETDEWEB) Start, S.E.; Carlson, R.B.; Gehrman, R.L. [Argonne National Lab., Idaho Falls, ID (United States). Engineering Div. 1995-06-01 The instrumentation and control of the Argon Cooling System (ACS) restricted fuel handling control system at Experimental Breeder Reactor II (EBR-II) is being upgraded from a system comprised of many discrete components and controllers to a computerized system with a graphical user interface (GUI). This paper describes the aspects of the upgrade including reasons for the upgrade, the old control system, upgrade goals, design decisions, philosophies and rationale, and the new control system hardware and software. 7. Caltrans WeatherShare Phase II System: An Application of Systems and Software Engineering Process to Project Development Science.gov (United States) 2009-08-25 In cooperation with the California Department of Transportation, Montana State University's Western Transportation Institute has developed the WeatherShare Phase II system by applying Systems Engineering and Software Engineering processes. The system... 8. Generalized optomechanics and its applications quantum optical properties of generalized optomechanical system CERN Document Server Li, Jin Jin 2013-01-01 A mechanical oscillator coupled to the optical field in a cavity is a typical cavity optomechanical system. In our textbook, we prepare to introduce the quantum optical properties of optomechanical system, i.e. linear and nonlinear effects. Some quantum optical devices based on optomechanical system are also presented in the monograph, such as the Kerr modulator, quantum optical transistor, optomechanical mass sensor, and so on. But most importantly, we extend the idea of typical optomechanical system to coupled mechanical resonator system and demonstrate that the combined two-level structure 9. Optical mass memory system (AMM-13). AMM-13 system segment specification Science.gov (United States) Bailey, G. A. 1980-01-01 The performance, design, development, and test requirements for an optical mass data storage and retrieval system prototype (AMM-13) are established. This system interfaces to other system segments of the NASA End-to-End Data System via the Data Base Management System segment and is designed to have a storage capacity of 10 to the 13th power bits (10 to the 12th power bits on line). The major functions of the system include control, input and output, recording of ingested data, fiche processing/replication and storage and retrieval. 10. Optical Synchronization Systems for Femtosecond X-Ray Sources CERN Document Server Wilcox, Russell; Staples, John W 2005-01-01 In femtosecond pump/probe experiments using short x-ray and optical pulses, precise synchronization must be maintained between widely separated lasers in a synchrotron or FEL facility. We are developing synchronization systems using optical signals for applications requiring different ranges of timing error. For the sub-100fs range we use an amplitude modulated CW laser at 1GHz to transmit RF phase information, and control the delay through a 100m fiber by observing the retroreflected signal. Initial results show 40fs peak-to-peak error above 10Hz, and 200fs long term drift, mainly due to amplitude sensitivity in the analog mixers. For the sub-10fs range we will lock two single-frequency lasers separated by several teraHertz to a master modelocked fiber laser, transmit the two frequencies over fiber, and lock two comb lines of a slave laser to these frequencies, thus synchronizing the two modelocked laser envelopes. For attosecond synchronization we propose a stabilized, free space link using bulk lens wavegu... 11. Printed polymer photonic devices for optical interconnect systems Science.gov (United States) Subbaraman, Harish; Pan, Zeyu; Zhang, Cheng; Li, Qiaochu; Guo, L. J.; Chen, Ray T. 2016-03-01 Polymer photonic device fabrication usually relies on the utilization of clean-room processes, including photolithography, e-beam lithography, reactive ion etching (RIE) and lift-off methods etc, which are expensive and are limited to areas as large as a wafer. Utilizing a novel and a scalable printing process involving ink-jet printing and imprinting, we have fabricated polymer based photonic interconnect components, such as electro-optic polymer based modulators and ring resonator switches, and thermo-optic polymer switch based delay networks and demonstrated their operation. Specifically, a modulator operating at 15MHz and a 2-bit delay network providing up to 35.4ps are presented. In this paper, we also discuss the manufacturing challenges that need to be overcome in order to make roll-to-roll manufacturing practically viable. We discuss a few manufacturing challenges, such as inspection and quality control, registration, and web control, that need to be overcome in order to realize true implementation of roll-to-roll manufacturing of flexible polymer photonic systems. We have overcome these challenges, and currently utilizing our inhouse developed hardware and software tools, security, imaging, energy, lighting etc. 12. Automatic bias control system of high speed electro-optic modulator in DPSK Systems Science.gov (United States) Xu, Miao; Tong, Shoufeng; Wang, Dashuai 2015-04-01 In DPSK communication system, the traditional way for bias voltage is loading a fixed bias voltage on the electro-optic modulator. For the influence of the temperature changes, the half-wave voltage of the electro-optic modulator may change and the DC bias supply voltage will have a certain degree of random fluctuations at the meantime which will cause the DC bias point of the modulator drift and consequently the communication systems are affected. To enhance the stability of the DPSK optical communication system and control the bias of Mach-Zehnder electro-optic modulators automatically, a PID control method has been proposed in this paper. After the actual operation, a DPSK signal transmission with transfer rate of 5Gb/s is built. Using the complex spectrum analyzer, stable signal and the constellation can be received. The automatic control system basically meets the needs of the DPSK transmission system of high stability, high accuracy and capacity of resisting disturbance. 13. Machine learning techniques for optical communication system optimization OpenAIRE Zibar, Darko; Wass, Jesper; Thrane, Jakob; Piels, Molly 2017-01-01 In this paper, machine learning techniques relevant to optical communication are presented and discussed. The focus is on applying machine learning tools to optical performance monitoring and performance prediction. 14. Solar Electric Generating System II finite element analysis Energy Technology Data Exchange (ETDEWEB) Dohner, J.L.; Anderson, J.R. 1994-04-01 On June 2, 1992, Landers earthquake struck the Solar Electric Generating System II, located in Daggett, California. The 30 megawatt power station, operated by the Daggett Leasing Corporation (DLC), suffered substantial damage due to structural failures in the solar farm. These failures consisted of the separation of sliding joints supporting a distribution of parabolic glass mirrors. At separation, the mirrors fell to the ground and broke. It was the desire of the DLC and the Solar Thermal Design Assistance Center (STDAC) of Sandia National Laboratories (SNL) and to redesign these joints so that, in the event of future quakes, costly breakage will be avoided. To accomplish this task, drawings of collector components were developed by the STDAC, from which a detailed finite element computer model of a solar collector was produced. This nonlinear dynamic model, which consisted of over 8,560 degrees of freedom, underwent model reduction to form a low order nonlinear dynamic model containing only 40 degrees of freedom. This model was then used as a design tool to estimate joint dynamics. Using this design tool, joint configurations were modified, and an acceptable joint redesign determined. The results of this analysis showed that the implementation of metal stops welded to support shafts for the purpose of preventing joint separation is a suitable joint redesign. Moreover, it was found that, for quakes of Landers` magnitude, mirror breakage due to enhanced vibration in the trough assembly is unlikely. 15. Rooftop PV system. Final technical progress report, Phase II Energy Technology Data Exchange (ETDEWEB) NONE 1995-08-01 Under this four-year PV:BONUS Program, ECD and United Solar are developing and demonstrating two new lightweight flexible building integrated Photovoltaic (BIPV) modules specifically designed as exact replacements for conventional asphalt shingles and standing seam metal roofing. These modules can be economically and aesthetically integrated into new residential and commercial buildings, and address the even larger roofing replacement market. The modules are designed to be installed by roofing contractors without special training which minimizes the installation and balance of system costs. The modules will be fabricated from high-efficiency, multiple-junction a-Si alloy solar cells developed by ECD and United Solar. Under the Phase I Program, which ended in March 1994, we developed two different concept designs for rooftop PV modules: (1) the United Solar overlapping (asphalt shingle replacement) shingle-type modules and (2) the ECD metal roof-type modules. We also developed a plan for fabricating, testing and demonstrating these modules. Candidate demonstration sites for our rooftop PV modules were identified and preliminary engineering designs for these demonstrations were developed; a marketing study plan was also developed. The major objectives of the Phase II Program, which started in June 1994 was (1) to develop, test, and qualify these new rooftop modules; (2) to develop mechanical and electrical engineering specifications for the demonstration projects; and (3) to develop a marketing/commercialization plan. 16. Fiber-Optic Strain Monitoring System for DUSEL Science.gov (United States) Wang, H.; Maclaughlin, M.; Noni, N.; Turner, A.; Murdoch, L.; Fratta, D. 2008-12-01 The opportunity to understand the response of rock masses to stresses deep within the earth's crust as a function of spatial and temporal scale is at the center of the geomechanics research program proposed for DUSEL. Within the 10-km3 volume of the former Homestake mine, deformations are expected from effective stress changes caused by mine dewatering, seasonal water table changes, and new excavations as well as from long-term creep of drifts and shafts. Data from a whole-mine deformation monitoring and measurement system are integral to calibrating a mine-scale, mechanical and hydrological finite-element model of laboratory and detector space. A synergistic objective of a long-term, state-of-the-art monitoring system is to ensure shaft, tunnel, and cavern stability as well as occupant safety. Fiber-optic sensors are highly stable over long periods of time and they can be daisy-chained to simply significantly the logistics of data acquisition of dozens of sensors on a string. Temperature measurements over large spatial scales can delineate fluid-flow paths and serve simultaneously as a detection system for anomalous temperatures. Two types of fiber-optic sensors are available: distributed strain and temperature (DST) and Fiber Bragg Grating (FBG). DST systems can be installed over kilometers of distance with measurement resolutions of 1-to-10 microstrains and 0.1°C over intervals of one-to-two meters. FBG strain gages and displacement transducers function the same as their electrical counterparts, save for the underlying physics in that displacements are measured as a shift in the spacing of a Bragg grating embedded into the optical fiber. These systems are highly scalable as more than 50,000 points of temperature and strain measurements can be collected from a single daisy-chained fiber-optic cable. Other fiber-based sensors, e.g., acceleration, air pressure, and gases, are also available and can become part of a fiber-based monitoring infrastructure. We 17. The analysis of an optical fibre communication system using laser ... African Journals Online (AJOL) Fibre dispersion has been the main limitation to high-speed data transmissions employing optical fibres. Various techniques and methods for combating this limitation have therefore been proposed. Such techniques include electrical and optical dispersion compensation, optical and electrical pulse shaping, and ... 18. Evaluation on licensability of KNGR system design (II) Energy Technology Data Exchange (ETDEWEB) Park, G. C.; Seo, K. R. [Seoul National Univ., Seoul (Korea, Republic of); Kim, J. K. [Hanyang Univ., Seoul (Korea, Republic of)] (and others) 2001-01-15 The CE methodology of DBA analysis are reviewed. Though UPTF test is different from the KNGR in the geometrical configuration of the down-corner, this was used as a reference to investigate the validity of CEFLASH and COMPERC-II codes in DBA analysis of KNGR. it revealed that CEFLASH is conservative but COMPERC-II for the Refill and Reflood phase is needed for the detailed investigation on ECC bypass, Entrainment, condensation phenomena, CCFL. The direct bypass rate based on the benchmark problem was quantitatively measured. The test model was scaled by the linear scaling methodology, and the accident conditions were the reflood phase at a CB-DEGB LBLOCA. The initial total air flow rate was determined considering the volume scaling factor (1/1000) on criteria of the KNGR design value. The volume rates of ECC water injected: through the DVI nozzles were changed for several case. The direct bypass ratio was about 22 % on the condition of the KNGR scaled-down air volume flow and ECC water velocity at about 1 m/s. The fast neutron fluence at the Reactor Pressure Vessel(RPV) of KNGR designed for 60 years of lifetime was calculated by Monte Carlo simulations and Discrete Ordinates Method for reactor pressure vessel integrity assessment. KNGR core geometry was modeled on a three-dimensional. In the full-scope Monte Carlo method, the maximum fast neutron flux at inner vessel belt line was estimated as 2.738 x 10{sup 10} neutrons/cm{sup 2}{center_dot}see. In the ROCS+MCNP4B calculation, the maximum flux of 2.769 x 10{sup 10} neutrons/ cm{sup 2}{center_dot}see at the RPV was obtained by tallying neutrons crossing the inner surface of the RPV. In ROCS+TORT Calculation, the maximum flux of 3.190 x 10{sup 10} neutrons/cm{sup 2} {center_dot}see was obtained at inner RPV belt line. The lifetime of KNGR was estimated on the basis of conservative end of life fluence limit value of the ABB-CE System 80+. Approximately, 72 Effective Full Power Years (EFPYs), equivalent to 90 calendar 19. A versatile fibre optic sensor interrogation system for the Ariane Launcher based on an electro-optically tuneable laser diode Science.gov (United States) Plattner, M. P.; Hirth, F.; Müller, M. S.; Hoffmann, L.; Buck, T. C.; Koch, A. W. 2017-11-01 Availability of reliable flight sensor data and knowledge of the structural behaviour are essential for safe operation of the Ariane launcher. The Ariane launcher is currently monitored by hundreds of electric sensors during test and qualification. Fibre optic sensors are regarded as a potential technique to overcome limitations of recent monitoring systems for the Ariane launcher [1]. These limitations include cumbersome application of sensors and harness as well as a very limited degree of distributed sensing capability. But, in order to exploit the various advantages of fibre optic sensors (high degree of multiplexing, distributed sensing capability, lower mass impact, etc.) dedicated measurement systems have to be developed and investigated. State-of-the-art fibre optic measurement systems often use free beam setups making them bulky and sensitive to vibration impact. Therefore a new measurement system is developed as part of the ESAstudy [2]. 20. Performance of Fractionally Spaced MLSE in OOK and PAM4 Bandwidth Limited Optical Systems Directory of Open Access Journals (Sweden) C. Prodaniuc 2015-11-01 Full Text Available We analyze the performance of fractionally spaced maximum likelihood sequence estimation (MLSE equalizers in OOK and PAM4 optical systems using optical and electrical components with cut-off frequencies less than the baud rate. It has been demonstrated that signals suffering from optical and electrical impairments can be efficiently equalized in cheap direct-detection optical receivers using MLSE equalizers with one or two samples, depending on the extent of bandwidth limitations.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4723166823387146, "perplexity": 4461.06460286302}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-17/segments/1524125948426.82/warc/CC-MAIN-20180426164149-20180426184149-00004.warc.gz"}
https://hal-ens.archives-ouvertes.fr/hal-01793650
# Asymmetry in elastic properties and the evolution of large continental strike-slip faults Abstract : We use geodetic studies to quantify several cases of significant asymmetry in interseismic and coseismic effects along large continental strike-slip faults using simple two-dimensional edge dislocation models. We first show that asymmetric elastic loading characterizes the present Main Marmara Fault, a portion of the North Anatolian Fault along the northern margin of the Sea of Marmara. The ratio of asymmetry there is about 10. This ratio is even larger, about 30, along the northern Sumatra fault near lake Toba caldera. We then examine two profiles near Point Reyes and Point Arena across the northern San Andreas Fault that have been previously proposed as affected by asymmetry both in interseismic and coseismic effects. We show that an asymmetry ratio of 1.6 in interseismic loading exists near Point Arena, with the southwest side of the fault being more rigid than the northeast one. On the other hand, we do not find significant asymmetry for the Point Reyes profile that was previously described as highly asymmetric. We examine coseismic motion during the 1906 earthquake along the same two profiles. Ratios of 1.2 and 1.7 are found for the Point Arena and Point Reyes profiles, respectively. We discuss the possible causes of asymmetry. Contrasts in seismic velocity in the brittle portion suggest ratios generally not exceeding 2.5 for the dynamic rigidity in the upper brittle section. Larger ratios may involve other complex causes such as differences between static and dynamic rigidities, contrasts in rheology in the deeper creeping sections, and postseismic transients. We conclude that asymmetry should be systematically included within the parameters to be inverted when dealing with the mechanics of large-scale strike-slip faults. Citation: Le Pichon, X., C. Kreemer, and N. Chamot-Rooke (2005), Asymmetry in elastic properties and the evolution of large continental strike-slip faults, Type de document : Article dans une revue Domaine : Littérature citée [48 références] https://hal-ens.archives-ouvertes.fr/hal-01793650 Contributeur : Marine Laffont <> Soumis le : mercredi 16 mai 2018 - 17:30:23 Dernière modification le : jeudi 24 octobre 2019 - 14:50:03 Archivage à long terme le : lundi 24 septembre 2018 - 15:24:56 ### Fichier 2005_LePichon&al.pdf Fichiers éditeurs autorisés sur une archive ouverte ### Citation Xavier Le Pichon, Corné Kreemer, Nicolas Chamot-Rooke. Asymmetry in elastic properties and the evolution of large continental strike-slip faults. Journal of Geophysical Research, American Geophysical Union, 2005, 110 (B3), pp.B03405. ⟨10.1029/2004JB003343⟩. ⟨hal-01793650⟩ ### Métriques Consultations de la notice ## 80 Téléchargements de fichiers
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8094603419303894, "perplexity": 8440.799155054925}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875142323.84/warc/CC-MAIN-20200217115308-20200217145308-00323.warc.gz"}
https://infoscience.epfl.ch/record/128722?ln=en
## Reduced basis method for multi-parameter dependent steady Navier-Stokes equations: applications to natural convection in a cavity This work focuses on the approximation of parametric steady Navier-- Stokes equations by the reduced basis method. For a particular instance of the parameters under consideration, we are able to solve the underlying partial differential equations, compute an output, and give sharp error bounds. The computations are split into an offline part, where the value of the parameters is not yet identified, but only within a range of interest, and an online part, where the problem is solved for an instance of the parameters. The offline part is expensive and is used to build a reduced basis and prepare all the ingredients -- mainly matrix-vector and scalar products, but also eigenvalue computations -- necessary for the online part, which is fast. We provide a model problem -- describing natural convection phenomena in a laterally heated cavity -- characterized by three parameters: Grashof and Prandtl numbers and the aspect ratio of the cavity. We show the feasibility and efficiency of the a posteriori error estimation by the natural norm approach considering several test cases by varying two different parameters. The gain in terms of CPU time with respect to a parallel finite element approximation is of three magnitude orders with an acceptable -- indeed less than 0.1% -- error on the selected outputs. Published in: Journal of Computational Physics, 228, 12, 4359-4378 Year: 2009 Keywords: Note: EPFL-IACS report 12.2008 Laboratories:
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8887831568717957, "perplexity": 479.20960048064364}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912204300.90/warc/CC-MAIN-20190325194225-20190325220225-00526.warc.gz"}
http://nbviewer.jupyter.org/github/BenLangmead/comp-genomics-class/blob/master/notebooks/Minimizers.ipynb
### Minimizers¶ One way to partition the space of all possible $k$-mers is by minimal $l$-mer, where $l < k$. For example, the minimal 2-mer in the string ABC is AB and the minimal 4-mer in the string abracadabra is abra. In this context, the minimal $l$-mer is called a minimizer, and we'll call such a partitioning scheme a $k, l$-minimizing scheme. In [1]: def string_to_kmers(s, length): return [s[i:i+length] for i in xrange(len(s)-length+1)] def minimizer(k, l): """ Given k-mer, return its minimal l-mer """ assert l <= k return min(string_to_kmers(k, l)) In [2]: minimizer('ABC', 2) Out[2]: 'AB' In [3]: minimizer('abracadabra', 4) Out[3]: 'abra' But if our goal is to partition the space of $k$-mers, couldn't we use a hash function instead? Say $k$ is 10 and $l$ is 4. A 10,4-minimizing scheme is a way for dividing the space of $4^{10}$ 10-mers (a million or so) into $4^4 = 256$ partitions. We can accomplish this with a hash function that maps $k$-mers to integers in $[0, 255]$. Why would we prefer minimizers over hash functions? The answer is that two strings that share long substrings tend to have the same minimizer, but not the same hash value. For example, the strings abracadabr and bracadabra have the substring bracadabr in common, and they have the same minimal 4-mer: In [4]: minimizer('abracadabr', 4), minimizer('bracadabra', 4) Out[4]: ('abra', 'abra') But their hash values (modulo 256) are not the same: In [5]: # you might need to 'pip install mmh3' first import mmh3 Out[5]: (26, 224) #### Partition size distribution¶ A feature of hash functions is that they divide the 10-mers quite uniformly (evenly) among the 256 buckets. 10,4-minimzers divide them much less uniformly. This becomes clear when you consider that, given a random 10-mer, the 4-mer TTTT is very unlikely to be its minimizer, whereas the 4-mer AAAA is much more likely. We can also show this empirically by partitioning a collection of random 10-mers: In [6]: import random random.seed(629) def random_kmer(k): return ''.join([random.choice('ACGT') for _ in xrange(k)]) In [7]: %matplotlib inline import matplotlib.pyplot as plt def plot_counts(counter, title=None): idx = range(256) cnts = map(lambda x: counter.get(x, 0), idx) plt.bar(idx, cnts, ec='none') plt.xlim(0, 256) plt.ylim(0, 35) if title is not None: plt.title(title) plt.show() In [8]: from collections import Counter # hash 1000 random 10-mers cnt = Counter([mmh3.hash(s) % 256 for s in [random_kmer(10) for _ in xrange(1000)]]) plot_counts(cnt, 'Frequency of partitions using hash mod 256') In [9]: def lmer_to_int(mer): """ Maps AAAA to 0, AAAC to 1, etc. Works for any length argument. """ cum = 0 charmap = {'A':0, 'C':1, 'G':2, 'T':3} for c in mer: cum *= 4 cum += charmap[c] return cum In [10]: # get minimal 4-mers from 1000 random 10-mers cnt = Counter([lmer_to_int(minimizer(s, 4)) for s in [random_kmer(10) for _ in xrange(1000)]]) plot_counts(cnt, 'Frequency of partitions using minimal 4-mer; AAAA at left, TTTT at right')
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4887550175189972, "perplexity": 5375.748106148461}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376823183.3/warc/CC-MAIN-20181209210843-20181209232843-00488.warc.gz"}
http://nrich.maths.org/7223/note
### Whole Number Dynamics I The first of five articles concentrating on whole number dynamics, ideas of general dynamical systems are introduced and seen in concrete cases. ### Whole Number Dynamics II This article extends the discussions in "Whole number dynamics I". Continuing the proof that, for all starting points, the Happy Number sequence goes into a loop or homes in on a fixed point. ### Whole Number Dynamics III In this third of five articles we prove that whatever whole number we start with for the Happy Number sequence we will always end up with some set of numbers being repeated over and over again. # Difference Dynamics ### Why do this problem? This is an engaging investigation that quickly yields results and leads to conjectures and a lot of mathematical thinking and discussion. Young learners, even at the primary stage, can understand and carry out the iterative process and see the cyclical patterns that emerge. It is not difficult to make and test conjectures. Also it is easy to understand that, after the first sequence, all the sequences have only positive terms and so, by taking differences, the terms in the sequences can't increase. This means that, from any chosen starting sequence, there are a strictly limited number of possible sequences that can follow, and sooner or later a sequence must repeat itself starting a cycle. Thus it is easy to prove that all the sequences of sequences end in cycles or a sequence of zeros. Not only can very young learners create the sequences of sequences, notice the cycles and make conjectures, but also this proof is very accessible and there are more questions to explore making this a low threshold high ceiling investigation. This is a simple example of a dynamical system and it can lead to discussion of how dynamical systems are used to model population dynamics and other natural phenomena. ### Possible approach If different groups in the class choose their own sequences from which to start, it won't be long before they notice that they are all getting the same sort of patterns. It is easier to see how the process works by starting with sequences of length three rather than length two. When everybody finds that before very long they have produced a cycle, and nobody can find a sequence that goes on indefinitely, then perhaps suggest that they try sequences of length four and see if the same thing happens. They will soon find that with length four the iteration always seems to stop with the zero sequence. From that point the teacher can either encourage the learners to discuss why the iteration always seems to stop with a zero sequence or cycle and, depending on the class and time available, reach a well argued proof. Alternatively the class can try starting with sequences of different lengths 2, 3, 4, 5, and 6 say, and try to discover if the lengths of the sequences determine whether the sequences go to zero or end in a cycle. ### Key questions Look at your chain of sequences, have you seen that sequence in the chain before? What will happen next? Is it worth continuing this chain or do you already know how the chain continues? Can you describe what is happening to your chain of sequences? (Encourage language like "it loops back on itself", don't introduce the term 'cycle' too early rather, if possible, let the term emerge in discussion). Can the terms that occur in the sequences get bigger? You said the terms can't get bigger so what is the biggest value any term can take in your chain? Then how many different values can the terms take? So how many different sequences is it possible to have in your chain? Can the chain go on for ever without any sequence being repeated? Would the same thing be true for any chain? Does the same thing happen when you start with sequences of different length? If you get to the zero sequence what can you say about the sequence in the chain just before it? ### Possible extension Prove that when sequence $\mathbf{a}$ maps to the next sequence in the chain $\mathbf{b}$ then $\mathbf{a}$ is a constant sequence if and only if $\mathbf{b}$ is the zero sequence. Read the article Difference Dynamics Discussion. ### Possible support Suggest the learners start with sequences of small terms which will converge very quickly. For example: $(1, 4, 3), (3, 1, 2), (2, 1, 1,), (1,0,1), (1,1,0), (0,1,1), (1,0,1)...$ and $(1,5,3,7), (4,2,4,5), (2,2,1,1), (0,1,0,1), (1,1,1,1), (0,0,0,0)....$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.635906994342804, "perplexity": 581.9748323348119}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1410657118605.64/warc/CC-MAIN-20140914011158-00093-ip-10-196-40-205.us-west-1.compute.internal.warc.gz"}
https://www.jottr.org/2017/03/18/dofuture/
doFuture 0.4.0 is available on CRAN. The doFuture package provides a universal foreach adaptor enabling any future backend to be used with the foreach() %dopar% { ... } construct. As shown below, this will allow foreach() to parallelize on not only multiple cores, multiple background R sessions, and ad-hoc clusters, but also cloud-based clusters and high performance compute (HPC) environments. 1,300+ R packages on CRAN and Bioconductor depend, directly or indirectly, on foreach for their parallel processing. By using doFuture, a user has the option to parallelize those computations on more compute environments than previously supported, especially HPC clusters. Notably, all plyr code with .parallel = TRUE will be able to take advantage of this without need for modifications - this is possible because internally plyr makes use of foreach for its parallelization. With doFuture, foreach can process your code in more places than ever before. Alright, it may not be able to process this programmer’s 62,500 punched cards. ## What is new in doFuture 0.4.0? • Load balancing: The doFuture %dopar% backend will now partition all iterations (elements) and distribute them uniformly such that the each backend worker will receive exactly one partition equally sized to those sent to the other workers. This approach speeds up the processing significantly when iterating over a large set of elements that each has a relatively small processing time. • Globals: Global variables and packages needed in order for external R workers to evaluate the foreach expression are now identified by the same algorithm as used for regular future constructs and future::future_lapply(). For full details on updates, please see the NEWS file. The doFuture package installs out-of-the-box on all operating systems. ## A quick example Here is a bootstrap example using foreach adapted from help("clusterApply", package = "parallel"). I use this example to illustrate how to perform foreach() iterations in parallel on a variety of backends. library("boot") run <- function(...) { cd4.rg <- function(data, mle) MASS::mvrnorm(nrow(data), mle$m, mle$v) cd4.mle <- list(m = colMeans(cd4), v = var(cd4)) boot(cd4, corr, R = 10000, sim = "parametric", ran.gen = cd4.rg, mle = cd4.mle) } ## Attach doFuture (and foreach), and tell foreach to use futures library("doFuture") registerDoFuture() ## Sequentially on the local machine plan(sequential) system.time(boot <- foreach(i = 1:100, .packages = "boot") %dopar% { run() }) ## user system elapsed ## 298.728 0.601 304.242 # In parallel on local machine (with 8 cores) plan(multiprocess) system.time(boot <- foreach(i = 1:100, .packages = "boot") %dopar% { run() }) ## user system elapsed ## 452.241 1.635 68.740 # In parallel on the ad-hoc cluster machine (5 machines with 4 workers each) nodes <- rep(c("n1", "n2", "n3", "n4", "n5"), each = 4L) plan(cluster, workers = nodes) system.time(boot <- foreach(i = 1:100, .packages = "boot") %dopar% { run() }) ## user system elapsed ## 2.046 0.188 22.227 # In parallel on Google Compute Engine (10 r-base Docker containers) vms <- lapply(paste0("node", 1:10), FUN = googleComputeEngineR::gce_vm, template = "r-base") vms <- lapply(vms, FUN = gce_ssh_setup) vms <- as.cluster(vms, docker_image = "henrikbengtsson/r-base-future") plan(cluster, workers = vms) system.time(boot <- foreach(i = 1:100, .packages = "boot") %dopar% { run() }) ## user system elapsed ## 0.952 0.040 26.269 # In parallel on a HPC cluster with a TORQUE / PBS scheduler # (Note, the below timing includes waiting time on job queue) plan(future.BatchJobs::batchjobs_torque, workers = 10) system.time(boot <- foreach(i = 1:100, .packages = "boot") %dopar% { run() }) ## user system elapsed ## 15.568 6.778 52.024 ## About .export and .packages When using doFuture::registerDoFuture(), there is no need to manually specify which global variables (argument .export) to export. By default, the doFuture backend automatically identifies and exports all globals needed. This is done using recursive static-code inspection. The same is true for packages that need to be attached; those will also be handled automatically and there is no need to specify them manually via argument .packages. This is in line with how it works for regular future constructs, e.g. y %<-% { a * sum(x) }. Having said this, you may still want to specify arguments .export and .packages because of the risk that your foreach() statement may not work with other foreach adaptors, e.g. doParallel and doSNOW. Exactly when and where a failure may occur depends on the nestedness of your code and the location of your global variables. Specifying .export and .packages manually skips such automatic identification. Finally, I recommend that you as a developer always try to write your code in such way the users can choose their own futures: The developer decides what should be parallelized - the user chooses how. Happy futuring!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.22586119174957275, "perplexity": 9820.605345212807}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057337.81/warc/CC-MAIN-20210922072047-20210922102047-00541.warc.gz"}
https://www.groundai.com/project/kinetic-energy-and-microcanonical-nonanalyticities-in-finite-and-infinite-systems/
Kinetic energy and microcanonical nonanalyticities in finite and infinite systems # Kinetic energy and microcanonical nonanalyticities in finite and infinite systems ## Abstract In contrast to the canonical case, microcanonical thermodynamic functions can show nonanalyticities also for finite systems. In this paper we contribute to the understanding of these nonanalyticities by working out the relation between nonanalyticities of the microcanonical entropy and its configurational counterpart. If the configurational microcanonical entropy has a nonanalyticity at , then the microcanonical entropy has a nonanalyticity at the same value of its argument for any finite value of the number of degrees of freedom . The presence of the kinetic energy weakens the nonanalyticities such that, if the configurational entropy is times differentiable, the entropy is -times differentiable. In the thermodynamic limit, however, the behaviour is very different: The nonanalyticities do not longer occur at the same values of the arguments, but the nonanalyticity of the microcanonical entropy is shifted to a larger energy. These results give a general explanation of the peculiar behaviour previously observed for the mean-field spherical model. With the hypercubic model we provide a further example illustrating our results. ###### pacs: 05.20.Gg, 05.20.-y, 05.70.Fh ## 1 Introduction In equilibrium statistical mechanics, nonanalyticities of thermodynamic functions are associated with thermodynamic phase transitions: more precisely, one commonly defines a phase transition point as the value of an external parameter (like temperature or pressure) where some thermodynamic function is nonanalytic.1 Such an identification is satisfactory in the canonical ensemble: as originally suggested by Kramers [1], nonanalyticities of thermodynamic functions calculated in the canonical ensemble may show up only in the thermodynamic limit , where is the number of degrees of freedom [2]. Moreover, such nonanalyticities separate different phases, i.e., regions of the parameters where the collective properties of the system are different. When studying physical models, one usually finds that thermodynamic functions have only a small number of nonanalyticities. In the microcanonical ensemble, however, the situation is different. First, nonanalyticities of the microcanonical entropy may even be present at finite . In principle, this fact should have been known for a long time, because even one-degree-of-freedom systems like a simple pendulum or a particle in a double-well potential do show nonanalyticities of the entropy. Still, it came as a surprise to many researchers in the field to see how frequently such nonanalyticities are encountered in many-particle systems [3, 4, 5, 6]. The behaviour of such nonanalyticities as a function of the number of degrees of freedom is remarkable: their number may grow with even exponentially, and their “strength” generically decreases linearly with . The latter means that the first derivatives of the entropy are continuous, where is of order [7, 8]. Since the usual thermodynamic quantities of interest are given by low-order derivatives of the entropy, these nonanalyticities of order can be observed only for very small from noisy data. In the thermodynamic limit most of these nonanalyticities disappear. Those (if any) that survive are typically associated to thermodynamic phase transitions and coincide with the canonical nonanalyticities if equivalence of statistical ensembles holds. Therefore the finite- nonanalyticities are not easily associated to any phenomenon that one would call a phase transition in the usual sense, and it seems inappropriate to define phase transitions in the microcanonical ensemble only based on the presence of nonanalyticities of the microcanonical entropy. The general relation between nonanalyticities of the microcanonical entropy and thermodynamic phase transitions, i.e., the question which of the finite-system nonanalyticities survive in the thermodynamic limit, is a subject of active research (see e.g. [9, 10]). It remains an open problem, despite some recent advances where—under suitable conditions—the “flatness” of stationary points was shown to be relevant to whether its thermodynamic limit contribution is vanishing or not [11, 8]. In the present paper we will investigate the effect of a kinetic energy term on these nonanalyticities. The general properties of microcanonical nonanalyticities have been studied so far only when no kinetic term is present and the Hamiltonian is identified with the interaction potential energy. Standard kinetic energy terms in the Hamiltonian, i.e., quadratic forms in the momenta, are known to yield only trivial contributions to thermodynamics in the canonical ensemble. In the microcanonical ensemble a standard kinetic energy term may have a more noticeable effect: for instance, it may restore equivalence between canonical and microcanonical ensembles when only partial equivalence holds in the absence of a kinetic energy [12]. But adding a kinetic energy term has also a remarkable effect on nonanalyticities of the microcanonical entropy: For the exactly solvable mean-field spherical model it has been shown recently that, in the presence of a standard kinetic energy term, a nonanalyticity of the entropy which for any finite is located at a fixed value of the energy per degree of freedom, jumps discontinuously to a different value of the energy in the thermodynamic limit [6]. In this paper we argue that such a behaviour of nonanalyticities in the thermodynamic limit is not a peculiarity of that model, but rather a general property of all the nonanalyticities that survive as . Moreover, we discuss the behaviour of nonanalyticities at finite , showing that the kinetic energy weakens them: If a nonanalyticity of order is present in the configurational entropy, adding a kinetic energy term to the Hamiltonian will increase the order roughly to , i.e., the first derivatives of the entropy will be continuous. The paper is organized as follows. After giving some definitions and fixing some notation, in section 2 the behaviour of microcanonical nonanalyticities is discussed in general. More specifically, in section 2.1 we recall the results on nonanalyticities of the configurational entropy, while section 2.2 is devoted to the effect of the kinetic energy. In section 3 we present a simple example to illustrate this behaviour. We will finish with some concluding remarks in section 4. ## 2 Microcanonical entropy and its nonanalyticities We consider classical Hamiltonian systems with degrees of freedom, with Hamiltonian function of the form H(p,q)=12N∑i=1p2i+V(q) (1) with some potential energy . We denote by the -dimensional configuration space and by its cotangent bundle, i.e., the phase space. We shall usually denote configurations as and phase space points as . The fundamental quantity of the microcanonical ensemble is the microcanonical entropy as a function of the energy (per degree of freedom) , sN(ε)=1NlnωN(ε), (2) where ωN(ε)=∫ΛN\rmdp\rmdqδ[H(p,q)−Nε] (3) is the density of states and denotes the Dirac distribution2. A related quantity is the configurational microcanonical entropy as a function of the potential energy (per degree of freedom) , scN(v)=1NlnωcN(v), (4) where ωcN(v)=∫ΓN\rmdqδ[V(q)−Nv] (5) is the configurational density of states. The configurational entropy equals the entropy when the Hamiltonian just consists of a configuration-dependent potential energy, . This is often the case when studying discrete spin systems where a definition of conjugate momenta is difficult. Alternatively, one can define the integrated density of states ΩN(ε)=∫ΛN\rmdp\rmdqΘ[Nε−H(p,q)], (6) where is the Heaviside step function, and the corresponding entropy function is σN(ε)=1NlnΩN(ε). (7) Again, the configurational counterparts can be defined, where ΩcN(v)=∫ΓN\rmdqΘ[Nv−V(q)] (8) and the entropy is given by σcN(v)=1NlnΩcN(v). (9) The density of states (3) and the integrated density of states (6) are related by ωN(ε)=\rmdΩN\rmdε, (10) and an analogous relation holds for the configurational quantities and . Under suitable conditions on the Hamiltonian the difference between the entropies vanishes in the thermodynamic limit, s∞≡limN→∞sN=limN→∞σN,sc∞≡limN→∞scN=limN→∞σcN (11) (see section 3.3.14 of [13] for details). Unless explicitly noted, the general properties of entropy functions considered in the following hold true for both definitions (2) and (7). As mentioned before, the microcanonical entropy is not necessarily an analytic function, neither for finite nor in the thermodynamic limit. We will point out in the following that nonanalyticities of the entropy originate from stationary points of the Hamiltonian [8]. A stationary point of a function is a point such that , and the value is called a stationary value of . When stationary points are non-degenerate, i.e., the Hessian matrix of is nonsingular at all stationary points , the function is called a Morse function. In this case all the stationary points are isolated. The index of the stationary point is the number of negative eigenvalues of the Hessian at . Minima and maxima are stationary points corresponding to and , respectively. In the following we will assume the potential energy to be a Morse function unless explicitly stated otherwise. Conceptually, this is an insignificant restriction, since Morse functions on some manifold form an open dense subset in the space of smooth functions on [14], and are therefore generic. Hence, if the potential is not a Morse function, we can deform it into a Morse function by adding an arbitrarily small perturbation.3 If the Hamiltonian is of standard form (1), its stationary points are of the form (pc,qc)=(0,qc), (12) where is a stationary point of the potential energy . Hence the stationary value of the Hamiltonian coincides with the stationary value of the potential energy, i.e., . Let us first recall some results on the relation between the nonanalyticities of the configurational entropy and the stationary points of . ### 2.1 Nonanalyticities of the configurational entropy The Morse property of the potential ensures that its stationary points are isolated, so that we can safely restrict our attention to a single stationary point . We denote by the corresponding stationary value per degree of freedom. It has been shown in [8] that in this situation the configurational density of states (5) can be written, to the leading order, as ωcN(v)=P(v−vc)+hN,j(v−vc)√∣∣det[HessV(qc)]∣∣+o[(v−vc)N/2−δ] (13) with some , where is a polynomial of degree smaller than in , and hN,j(x)=⎧⎪⎨⎪⎩(−1)j/2x(N−2)/2Θ(x)\mbox{for j even},(−1)(j+1)/2x(N−2)/2π−1ln|x|\mbox{for N even, j odd},(−1)(N−j)/2(−x)(N−2)/2Θ(−x)\mbox{for N, j odd}. (14) If there are further stationary points, the configurational density of states is given by the sum of the contributions of each stationary point. A proof (of an even stronger result including higher order terms) is given in [8]. This result can be rephrased as follows: 1. Every stationary point of gives rise to a nonanalyticity of the configurational entropy at the corresponding stationary value . 2. The order of this nonanalyticity is , i.e., is precisely -times differentiable at .4 Since the integrated density of states is obtained from by integration, the nonanalyticity of the entropy following from definition (7) is slightly weaker. In this case, statement (ii) has to be replaced by: 1. The configurational entropy is precisely -times differentiable at . ### 2.2 The role of kinetic energy We have pointed out in equation (12) that, for standard Hamiltonians of the form (1), if is a stationary point of the potential energy then is a stationary point of and vice versa. Hence the kinetic energy is zero at stationary points and for all stationary points of . As a consequence, for all finite the nonanalyticities of the configurational entropy—which we have traced back to stationary points in the previous section—show up at the very same stationary values as those of the entropy. As we shall see below, the presence of a kinetic energy term has a twofold effect on these nonanalyticities. 1. At any finite , the order of the nonanalyticity is increased by the presence of a kinetic energy term. 2. More surprisingly, in the thermodynamic limit those of the nonanalyticities which survive jump to a different value of the energy. Both these results follow from the fact that for Hamiltonians of the class (1) the density of states can be written as a convolution [13]. Defining a kinetic density of states as ωkN(γ)=∫\mathbbmRN\rmdpδ(12N∑i=1p2i−Nγ), (15) we can write the density of states as ωN(ε)=∫∞0\rmdγωkN(γ)ωcN(ε−γ)=∫ε−∞\rmdγωkN(ε−γ)ωcN(γ), (16) where the configurational density of states is given by equation (5). Since equations (15) and (16) hold in the same form also for the integrated densities of states , all the following results will be valid also for integrated densities of states and entropies defined as in equation (7). We will now discuss the finite- case and the thermodynamic limit separately in the following two subsections. #### The finite-N case. At any finite , the effect of the kinetic energy term on the order of the nonanalyticities of the entropy can be computed explicitly from the convolution integral (16). Such a calculation is reported in A, and the only additional input used is that—in accordance with the results from section 2.1—the nonanalyticities of the configurational density of states are of algebraic type. Alternatively, the result of that calculation can be anticipated via an intuitive argument: We know that is smooth, and we assume for simplicity that the configurational density of states is nonanalytic only at and analytic elsewhere. As long as , it is evident from the right-hand side of (16) that the nonanalyticity of at is never reached in the integration. Hence the convolution integral (16) is the integral over the product of two analytic functions and yields an analytic function. As soon as , the nonanalyticity of is inside the range of integration and induces a nonanalyticity in the convolution integral. As a simple example consider the convolution integral (16) with the choices and with . Performing the integration yields ωN(ε)=12(ε−a)2Θ(ε−a). (17) As expected from the above reasoning, this function inherits the nonanalyticity at from . This is in agreement with our previous observation that the stationary values of the Hamiltonian coincide with those of the potential energy. In this example, is a 0-times differentiable function, whereas is 1-times differentiable. The calculation in A shows that in general adding a kinetic energy increases the order of the nonanalyticity. The effect of a standard kinetic energy term on the nonanalyticities of the entropy at finite , as computed in A, can be summarized as follows: 1. If the configurational density of states is nonanalytic at , then the density of states and the entropy are nonanalytic at ; 2. The density of states and the entropy at are differentiable -times more often than the configurational density of states . As far as the entropy is concerned, statements (i) and (ii) above hold if is in the interior of the domain of the entropy. Both statements hold also for the integrated densities of states and and for the entropy . #### Thermodynamic limit. We will assume in the following that the thermodynamic limit of the configurational entropy exists, i.e., that increases exponentially with asymptotically for large . From definition (15) it follows that the kinetic density of states is related to the volume of an -dimensional sphere with radius , ωkN(γ)=1√2Nγ∫\mathbbmRN\rmdpδ(√∑p2i−√2Nγ)=aNγN/2−1 (18) with aN=2πN/2Γ(N/2)(2N)N/2−1. (19) As a consequence, this quantity likewise grows exponentially in and the thermodynamic limit of the kinetic entropy exist. Under these conditions, the thermodynamic limit is known to exist also for the entropy (see section 3.4.1 of [13]) and is given, apart from irrelevant constants, by s∞(ε)=limN→∞1NlnωN(ε)=limN→∞1Nlnmaxγ⩾0[ωkN(γ)ωcN(ε−γ)]. (20) Denoting by the value of that realizes the extremum in the right-hand side of equation (20), this expression can be rewritten as s∞(ε)=sk∞[~γ(ε)]+sc∞[ε−~γ(ε)]. (21) From the above equation it is apparent that must be equal to the microcanonical average of the kinetic energy per degree of freedom, (22) which allows us to rewrite equation (21) in the form s∞(ε)=sk∞[ε−⟨v⟩(ε)]+sc∞[⟨v⟩(ε)]. (23) If is nonanalytic at , then will be nonanalytic at , where is defined implicitly by ⟨v⟩(ε∗)=vc. (24) Apparently differs from the value of , unless the average kinetic energy (22) vanishes at . Hence, despite their common origin from the nonanalyticity of at , nonanalyticities of jump from their finite- value of to a different value in the thermodynamic limit. These properties of the nonanalyticities of the microcanonical entropy at finite and infinite had previously been studied for a simple model system, the mean-field spherical model. The results reported in Refs. [6, 12]—surprising at the time—are all in agreement with and satisfactorily explained by the results reported in the present article. To further illustrate the predictions, we will discuss in the next section an even simpler, analytically solvable model which is of pedagogical value. ## 3 A simple example: the hypercubic model The hypercubic model, introduced in [16], can be seen as an -dimensional generalization of a particle in a one-dimensional potential made up of two square wells separated by a finite barrier. The Hamiltonian is of standard form (1), i.e., is given by a standard kinetic term plus a potential energy . To define the potential, we consider a hypercube of side length , centered at the origin, as well as two disjoint hypercubes , of side length , symmetrically arranged with respect to the hyperplane (see sketch in Fig. 1). The potential of the hypercubic model is then defined as V(q)=⎧⎪⎨⎪⎩0\mbox{for q∈{A+∪A−},}Nvc\mbox{for q∈B∖{A+∪A−},}∞\mbox{for q∈\mathbbmRN∖B}. (25) The integrated density of states of this model can be computed analytically for arbitrary , and the effect of a kinetic energy term will nicely illustrate the general results of the previous section. ### 3.1 Finite-N integrated microcanonical entropy The integrated density of states is calculated for the hypercubic model by inserting the potential (25) into definition (6), ΩN(ε)=∫\mathbbmRN\rmdp∫A+∪A−\rmdqΘ(Nε−N∑i=1p2i2) +∫\mathbbmRN\rmdp∫B∖{A+∪A−}\rmdqΘ[N(ε−vc)−N∑i=1p2i2] =2aN∫\mathbbmRN\rmdpΘ(Nε−N∑i=1p2i2)+(bN−2aN)∫\mathbbmRN\rmdpΘ[N(ε−vc)−N∑i=1p2i2]. (26) The remaining integrations of the form ∫\mathbbmRN\rmdpΘ(γ22−N∑i=1p2i2)=CNγNwithCN=πN/2Γ(N/2+1) (27) yield the volumes of -dimensional balls of radii and , respectively. Inserting this formula into and making use of (7), the integrated microcanonical entropy σN(ε)=⎧⎨⎩(1N+12)ln2+lna+1NlnCN+12(lnN+lnε)\mbox{for 0<ε can be computed. For the integrated configurational microcanonical entropy, a similar calculation yields σcN(v)={lna+1Nln2\mbox{for 0 From these equations we see that and each have one nonanalyticity in the interior of their domains , located at argument in both cases. The integrated configurational microcanonical entropy is a piecewise constant function with a discontinuity5 at , whereas the functional form of the integrated microcanonical entropy in the vicinity of is , in agreement with the general results of section 2.2.1. ### 3.2 Thermodynamic limit From the entropies (28) and (29), the corresponding thermodynamic limit values σ∞(ε)=limN→∞σN(ε)andσc∞(ε)=limN→∞σcN(ε) (30) can be computed. Since limN→∞1NlnCN=12(1+lnπ+ln2−lnN), (31) for we find σ∞(ε)=lna+12lnε+12lnπ+ln2+12, (32) whereas for the case we can write σ∞(ε) = 12(1+lnπ+ln2−lnN) + limN→∞1Nln{2aN(2Nε)N/2+(bN−2aN)[(2N)N/2(ε−vc)N/2]}. The argument of the logarithm in equation (3.2) is the sum of two terms. Both terms are exponentially large in , so that in the limit only the larger one survives. There is a value ε∗=vc1−(ab)2, (34) of for which the two terms are equal, and apparently . For , the first term in the argument of the logarithm survives, yielding the same functional form for as in the case . For , however, the second term wins, and we obtain as a final result σ∞(ε)={12+ln2+lna+12lnπ+12lnε\mbox{for 0<ε<ε∗},12+ln2+lnb+12lnπ+12ln(ε−vc)% \mbox{for ε⩾ε∗}. (35) This expression shows that the entropy is analytic at but nonanalytic at . Therefore, although the entropy is nonanalytic at for all finite , the nonanalyticity jumps to a different energy value in the thermodynamic limit. Furthermore, at this value the statistical average of the potential energy per degree of freedom equals the value of the finite- nonanalyticity, ⟨v⟩(ε∗)=vc. (36) A plot of the entropy for various finite values of as well as in the thermodynamic limit is shown in Fig. 2. Again, the results for the hypercubic model confirm the general reasoning of section 2.2.2. From equation (35) the microcanonical temperature T(ε)=(\rmdsN\rmdε)−1={2ε\mbox{for ε<ε∗},2(ε−vc)\mbox{for ε⩾ε∗}. (37) is easily computed. The jump in temperature at is a signal of a microcanonical first-order phase transition. ## 4 Concluding remarks Previously, nonanalyticities of the microcanonical density of states and of the microcanonical entropy usually had been investigated for configurational quantities. This choice was mostly motivated by the belief that the effect of a standard kinetic energy term, i.e., a quadratic form in the momenta, is trivial anyway. However, we have shown in the present article that the effect of a standard kinetic energy term on the nonanalyticities of the microcanonical entropy is quite remarkable, both for finite and for infinite . If the configurational microcanonical entropy has a nonanalyticity at , then the microcanonical entropy has a nonanalyticity at the same value of its argument for any finite value of the number of degrees of freedom . The presence of the kinetic energy weakens the nonanalyticities. More precisely, if the configurational entropy is times differentiable, then the entropy is -times differentiable. In the thermodynamic limit, however, the behaviour is very different: The nonanalyticities do not longer occur at the same values of the arguments, but the nonanalyticity of the microcanonical entropy is shifted to a larger energy as given by equation (24). These results give a general explanation of what had previously been observed for the mean-field spherical model [6]. With the hypercubic model we have provided a further example for which both, the configurational microcanonical entropy and the microcanonical entropy can be computed analytically. Due to its simplicity, this model is of pedagogical value and nicely illustrates the general behaviour of nonanalyticities of the entropy. The study of nonanalyticities of the configurational microcanonical entropy, their relation to stationary points of the potential energy, and their connection with topology changes of the constant-potential surfaces in configuration space, has attracted some interest recently and has proved useful to obtain a deeper understanding of phase transitions (see [17, 9, 10] for reviews). With the present work, we add another piece to this understanding by providing the relevant “translation rules” between configurational quantities and their total-energy counterparts. ## Acknowledgments L.C. thanks Physikalisches Institut, Universität Bayreuth for hospitality during the early stage of this work. ## Appendix A Finite-N nonanalyticities of the density of states: explicit calculation We now explicitly calculate the convolution integral (16) at finite . For the sake of simplicity, let the configurational density of states have just one nonanalyticity at (this choice of the stationary value will ease the notation). According to the general results reported in [8] and recalled in section 2, we assume, apart from irrelevant multiplicative constants, ωcN(v)=Θ(v−vmin)[~ωcN(v)+ω±(v)], (38) where the distribution forces to vanish when the potential energy per degree of freedom is smaller than its minimum6 , is smooth and is the nonanalytic part; more precisely, to the leading order, ω±(v)={c−|v|p\mbox{for v<0},c+|v|p\mbox{for v>0}, (39) where and . With the the explicit expression (18) for the kinetic density of states the convolution integral (16) becomes ωN(ε)=~ωN(ε)+aN∫∞0\rmdγγN/2−1ω±(ε−γ)Θ(ε−γ−vmin) (40) with some smooth function . We are interested in the integral containing the nonanalyticity, I=∫∞0\rmdγγN/2−1ω±(ε−γ)Θ(ε−γ−vmin), (41) and we can write with I1 = ∫∞0\rmdγγN/2−1c+|ε−γ|pΘ(ε−γ), (42) I2 = ∫∞0\rmdγγN/2−1c−|ε−γ|pΘ(γ−ε)Θ(ε−γ−vmin). (43) With the substitution we obtain I1=c+εp+N/2Θ(ε)∫10\rmdxxN/2−1(1−x)p=c+εp+N/2B(N/2,1+p)Θ(ε), (44) where is the beta function. For the calculation of it is convenient to treat the cases and separately. For , I2=c−∫ε−vminε\rmdγγN/2−1(γ−ε)p, (45) and with the change of variables one gets I2=−c−∫vmin0\rmdx(−x)p(ε−x)N/2−1. (46) The substitution yields I2 =−c−(−1)pεp+N/2∫vmin/ε0\rmdyyp(1−y)N/2−1 (47) =−c−(−1)pεp+N/2Bvmin/ε(1+p,N/2), where denotes the incomplete beta function.7 For , a similar calculation gives I2=c−(−1)−N/2(−ε)p+N/2B1−vmin/ε(N/2,1+p)Θ(ε−vmin). (48) Assembling the pieces together we finally obtain for the nonanalytic part of the density of states ωN(ε)−~ωN(ε)∝|ε|p+N/2[b−(ε)Θ(−ε)+b+(ε)Θ(ε)], (49) where b−(ε) = c−(−1)−N/2B1−vmin/ε(N/2,1+p)Θ(ε−vmin), (50) b+(ε) = c+B(N/2,1+p)−c−(−1)pBvmin/ε(1+p,N/2). (51) The function given by equation (49), and thus the entropy , is nonanalytic at . At this point, is -times differentiable, whereas the configurational entropy we started out with is only -times differentiable. Without any modifications the calculation can be carried over to the integrated density of states . ## References ### Footnotes 1. Departing slightly from the standard definition, we use the notion of analyticity in the sense of a real function being infinitely-many times differentiable. This property typically, but not always coincides with the standard definition via the existence of a Taylor series. 2. We define all thermodynamic functions per degree of freedom, which accounts for the factor in the definitions. The Boltzmann constant is set to unity. 3. In case is not a Morse function due to a continuous symmetry, Morse-Bott theory [15] should allow to carry over essentially all the results from the theory of standard Morse functions. 4. With we denote the largest integer smaller than . 5. Since the potential (25) is not a Morse function, it is no surprise that the nonanalyticity is not of the generic type given in equations (13) and (14). 6. Since we assume , we have . 7. The incomplete beta function has a branch cut discontinuity in the complex plane running along the negative real axis. This implies that, when expressing in terms of , the correct branch has to be used. The given integral representations, however, are unambiguous. ### References 1. E. G. D. Cohen. George E. Uhlenbeck and statistical mechanics. Am. J. Phys., 58:619–625, 1990. 2. R. B. Griffiths. Rigorous results and theorems. In C. Domb and M. S. Green, editors, Phase Transitions and Critical Phenomena, volume 1. Academic Press, London, 1972. 3. M. Kastner and O. Schnetz. On the mean-field spherical model. J. Stat. Phys., 122:1195–1214, 2006. 4. J. Dunkel and S. Hilbert. Phase transitions in small systems: Microcanonical vs. canonical ensembles. Physica A, 370:390–406, 2006. 5. S. Hilbert and J. Dunkel. Nonanalytic microscopic phase transitions and temperature oscillations in the microcanonical ensemble: An exactly solvable one-dimensional model for evaporation. Phys. Rev. E, 74:011120, 2006. 6. L. Casetti and M. Kastner. Nonanalyticities of entropy functions of finite and infinite systems. Phys. Rev. Lett., 97:100602, 2006. 7. M. Kastner, S. Schreiber, and O. Schnetz. Phase transitions from saddles of the potential energy landscape. Phys. Rev. Lett., 99:050601, 2007. 8. M. Kastner, O. Schnetz, and S. Schreiber. Nonanalyticities of the entropy induced by saddle points of the potential energy landscape. J. Stat. Mech. Theory Exp., 2008:P04025, 2008. 9. M. Kastner. Phase transitions and configuration space topology. Rev. Mod. Phys., 80:167–187, 2008. 10. M. Pettini. Geometry and Topology in Hamiltonian Dynamics and Statistical Mechanics, volume 33 of Interdisciplinary Applied Mathematics. Springer, 2007. 11. M. Kastner and O. Schnetz. Phase transitions induced by saddle points of vanishing curvature. Phys. Rev. Lett., 100:160601, 2008. 12. L. Casetti and M. Kastner. Partial equivalence of statistical ensembles and kinetic energy. Physica A, 384:318–334, 2007. 13. D. Ruelle. Statistical Mechanics: Rigorous Results. Benjamin, Reading, 1969. 14. M. Demazure. Bifurcations and Catastrophes: Geometry of Solutions to Nonlinear Problems. Springer, 2000. 15. T. Frankel. The Geometry of Physics: An Introduction. Cambridge University Press, Cambridge, 2nd edition, 2004. 16. F. Baroni and L. Casetti. Topological conditions for discrete symmetry breaking and phase transitions. J. Phys. A: Math. Gen., 39:529–545, 2006. 17. L. Casetti, M. Pettini, and E. G. D. Cohen. Geometric approach to Hamiltonian dynamics and statistical mechanics. Phys. Rep., 337:237–341, 2000. You are adding the first comment! How to quickly get a good reply: • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made. • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements. • Your comment should inspire ideas to flow and help the author improves the paper. The better we are at sharing our knowledge with each other, the faster we move forward. The feedback must be of minumum 40 characters
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9785050749778748, "perplexity": 508.05122554012553}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376823322.49/warc/CC-MAIN-20181210101954-20181210123454-00399.warc.gz"}
https://worldwidescience.org/topicpages/r/rhic+cryogenic+system.html
#### Sample records for rhic cryogenic system 1. Cryogenic systems for proof of the principle experiment of coherent electron cooling at RHIC Huang, Yuenian; Belomestnykh, Sergey; Brutus, Jean Clifford; Lederle, Dewey; Orfin, Paul; Skaritka, John; Soria, Victor; Tallerico, Thomas; Than, Roberto [Collider Accelerator Department, Brookhaven National Laboratory, Upton, NY 11973 (United States) 2014-01-29 The Coherent electron Cooling (CeC) Proof of Principle (PoP) experiment is proposed to be installed in the Relativistic Heavy Ion Collider (RHIC) to demonstrate proton and ion beam cooling with this new technique that may increase the beam luminosity in certain cases, by as much as tenfold. Within the scope of this project, a 112 MHz, 2MeV Superconducting Radio Frequency (SRF) electron gun and a 704 MHz 20MeV 5-cell SRF cavity will be installed at IP2 in the RHIC ring. The superconducting RF electron gun will be cooled in a liquid helium bath at 4.4 K. The 704 MHz 5-cell SRF cavity will be cooled in a super-fluid helium bath at 2.0 K. This paper discusses the cryogenic systems designed for both cavities. For the 112 MHz cavity cryogenic system, a condenser/boiler heat exchanger is used to isolate the cavity helium bath from pressure pulses and microphonics noise sources. For the 704 MHz 5-cell SRF cavity, a heat exchanger is also used to isolate the SRF cavity helium bath from noise sources in the sub-atmospheric pumping system operating at room temperature. Detailed designs, thermal analyses and discussions for both systems will be presented in this paper. 2. VIBRATION MEASUREMENTS IN A RHIC QUADRUPOLE AT CRYOGENIC TEMPERATURES. JAIN, A.; AYDIN, S.; HE, P.; ANERELLA, M.; GANETIS, G.; HARRISON, M.; PARKER, B.; PLATE, S. 2005-10-17 One of the concerns in using compact superconducting magnets in the final focus region of the ILC is the influence of the cryogen flow on the vibration characteristics. As a first step towards characterizing such motion at nanometer levels, a project was undertaken at BNL to measure the vibrations in a spare RHIC quadrupole under cryogenic conditions. Given the constraints of cryogenic operation, and limited space available, it was decided to use a dual head laser Doppler vibrometer for this work. The performance of the laser vibrometer was tested in a series of room temperature tests and compared with results from Mark L4 geophones. The laser system was then used to measure the vibration of the cold mass of the quadrupole with respect to the outside warm enclosure. These measurements were carried out both with and without the flow of cold helium through the magnet. The results indicate only a minor increase in motion in the horizontal direction (where the cold mass is relatively free to move). 3. The Relativistic Heavy Ion Collider (rhic) Cryogenic System at Brookhaven National Laboratory: Review of the Modifications and Upgrades Since 2002 and Planned Improvements Than, R.; Tuozzolo, J.; Sidi-Yekhlef, A.; Ganni, V.; Knudsen, P.; Arenius, D. 2008-03-01 Brookhaven National Laboratory continues its multi-year program to improve the operational efficiency, reliability, and stability of the cryogenic system, which also resulted in an improved beam availability of the Relativistic Heavy Ion Collider (RHIC). This paper summarizes the work and changes made after each phase over the past four years to the present, as well as proposed future improvements. Power usage dropped from an initial 9.4 MW to the present 5.1 MW and is expected to drop below 5 MW after the completion of the remaining proposed improvements. The work proceeded in phases, balancing the Collider's schedule of operation, time required for the modifications and budget constraints. The main changes include process control, compressor oil removal and management, elimination of the use of cold compressors and two liquid-helium storage tanks, insulation of the third liquid-helium storage tank, compressor-bypass flow reduction and the addition of a load turbine (Joule-Thomson expander) with associated heat exchangers at the cold end of the plant. Also, liquid helium pumps used for forced circulation of the sub-cooled helium through the magnet loops were eliminated by an accelerator supply flow reconfiguration. Planned future upgrades include the resizing of expanders 5 and 6 to increase their efficiencies. 4. Cryogenic Systems Hosoyama, Kenji 2002-02-01 In this lecture we discuss the principle of method of cooling to a very low temperature, i.e. cryogenic. The "gas molecular model" will be introduced to explain the mechanism cooling by the expansion engine and the Joule-Thomson expansion valve. These two expansion processes are normally used in helium refrigeration systems to cool the process gas to cryogenic temperature. The reverse Carnot cycle will be discussed in detail as an ideal refrigeration cycle. First the fundamental process of liquefaction and refrigeration cycles will be discussed, and then the practical helium refrigeration system. The process flow of the system and the key components; -compressor, expander, and heat exchanger- will be discussed. As an example of an actual refrigeration system, we will use the cryogenic system for the KEKB superconducting RF cavity. We will also discuss the liquid helium distribution system, which is very important, especially for the cryogenic systems used in accelerator applications. 1 Principles of Cooling and Fundamental Cooling Cycle 2 Expansion engine, Joule-Thomson expansion, kinetic molecular theory, and enthalpy 3 Liquefaction Systems 4 Refrigeration Systems 5 Practical helium liquefier/refrigeration system 6 Cryogenic System for TRISTAN Superconducting RF Cavity 5. RHIC 10 Hz global orbit feedback system Michnoff, R.; Arnold, L.; Carboni, L.; Cerniglia, P; Curcio, A.; DeSanto, L.; Folz, C.; Ho, C.; Hoff, L.; Hulsart, R.; Karl, R.; Luo, Y.; Liu, C.; MacKay, W.; Mahler, G.; Meng, W.; Mernick, K.; Minty, M.; Montag, C.; Olsen, R.; Piacentino, J.; Popken, P.; Przybylinski, R.; Ptitsyn, V.; Ritter, J.; Schoenfeld, R.; Thieberger, P.; Tuozzolo, J.; Weston, A.; White, J.; Ziminski, P.; Zimmerman, P. 2011-03-28 Vibrations of the cryogenic triplet magnets at the Relativistic Heavy Ion Collider (RHIC) are suspected to be causing the horizontal beam perturbations observed at frequencies around 10 Hz. Several solutions to counteract the effect have been considered in the past, including a local beam feedback system at each of the two experimental areas, reinforcing the magnet base support assembly, and a mechanical servo feedback system. However, the local feedback system was insufficient because perturbation amplitudes outside the experimental areas were still problematic, and the mechanical solutions are very expensive. A global 10 Hz orbit feedback system consisting of 36 beam position monitors (BPMs) and 12 small dedicated dipole corrector magnets in each of the two 3.8 km circumference counter-rotating rings has been developed and commissioned in February 2011. A description of the system architecture and results with beam will be discussed. 6. RHIC BPM System Modifications and Performance Satogata, Todd; Cameron, Peter; Cerniglia, Phil; Cupolo, John; Curcio, Anthony J; Dawson, William C; Degen, Christopher; Gullotta, Justin; Mead, Joe; Michnoff, Robert; Russo, Thomas; Sikora, Robert 2005-01-01 The RHIC beam position monitor (BPM) system provides independent average orbit and turn-by-turn (TBT) position measurements. In each ring, there are 162 measurement locations per plane (horizontal and vertical) for a total of 648 BPM planes in the RHIC machine. During 2003 and 2004 shutdowns, BPM processing electronics were moved from the RHIC tunnel to controls alcoves to reduce radiation impact, and the analog signal paths of several dozen modules were modified to eliminate gain-switching relays and improve signal stability. This paper presents results of improved system performance, including stability for interaction region and sextupole beam-based alignment efforts. We also summarize performance of improved million-turn TBT acquisition channels for nonlinear dynamics and echo studies. 7. SNS Cryogenic Systems Commissioning Hatfield, D.; Casagrande, F.; Campisi, I.; Gurd, P.; Howell, M.; Stout, D.; Strong, H.; Arenius, D.; Creel, J.; Dixon, K.; Ganni, V.; Knudsen, P. 2006-04-01 The Spallation Neutron Source (SNS) is under construction at Oak Ridge National Laboratory. The cold section of the Linac consists of 81 superconducting radio frequency cavities cooled to 2.1K by a 2400 watt cryogenic refrigeration system. The major cryogenic system components include warm helium compressors with associated oil removal and gas management, 4.5K cold box, 7000L liquid helium dewar, 2.1K cold box (consisting of 4 stages of cold compressors), gaseous helium storage, helium purification and gas impurity monitoring system, liquid nitrogen storage and the cryogenic distribution transfer line system. The overall system commissioning and future plans will be presented. 8. SNS Cryogenic Systems Commissioning D. Hatfield; F. Casagrande; I. Campisi; P. Gurd; M. Howell; D. Stout; H. Strong; D. Arenius; J. Creel; K. Dixon; V. Ganni; and P. Knudsen 2005-08-29 The Spallation Neutron Source (SNS) is under construction at Oak Ridge National Laboratory. The cold section of the Linac consists of 81 superconducting radio frequency cavities cooled to 2.1K by a 2400 watt cryogenic refrigeration system. The major cryogenic system components include warm helium compressors with associated oil removal and gas management, 4.5K cold box, 7000L liquid helium dewar, 2.1K cold box (consisting of 4 stages of cold compressors), gaseous helium storage, helium purification and gas impurity monitoring system, liquid nitrogen storage and the cryogenic distribution transfer line system. The overall system commissioning and future plans will be presented. 9. ARTUS: The tune measurement system at RHIC Drees, A.; Brennan, M.; Connolly, R.; Michnoff, R.; DeLong, J. 2000-11-01 The super-conducting Relativistic Heavy Ion Collider (RHIC) with two separate rings and six combined interaction regions will provide collisions between equal and unequal heavy ion species up to Au ions in typically 60 bunches. The betatron tunes of the two beams are among the most important parameters to be measured. The tunes have to be acquired at any moment during accelerator operation and in particular during the acceleration process. At RHIC the tune measurement device (ARTUS) consists of a fast horizontal and vertical kicker magnet and a dedicated beam position monitor in each ring. The system layout is described and first experiences from operation is reported. 10. Upgrade scenario for the RHIC collimation system Robert-Demolaize, G.; Drees, A. 2012-01-19 The RHIC collimation system is used to reduce background levels in both STAR and PHENIX detectors. With a push for higher luminosity in the near future, it becomes critical to check if and how the level of performance of the collimators can be improved. The following reviews a proposal for additional collimators placed further downstream of the current system and designed to intercept the tertiary halo coming out of the IR8 insertion before it can reach the triplet quadrupoles in either STAR or PHENIX. Simulations have been peformed to quantify the efficiency of additional collimator jaws in RHIC. Each figure presented in this article clearly shows that the additional mask collimators provide the expected reduction in losses around the machine, and especially to the incoming triplet to the STAR experiment (IP6), for the Yellow beam as much as for the Blue beam. Looking at compiled statistics for all three working point cases studied, proton losses around the machine are reduced by roughly one order of magnitude: at most a factor 30 for magnet losses, and at most a factor 40 for losses in spaces between magnets. 11. Upgrade scenario for the RHIC collimation system Robert-Demolaize, G.; Drees, A. 2012-01-19 The RHIC collimation system is used to reduce background levels in both STAR and PHENIX detectors. With a push for higher luminosity in the near future, it becomes critical to check if and how the level of performance of the collimators can be improved. The following reviews a proposal for additional collimators placed further downstream of the current system and designed to intercept the tertiary halo coming out of the IR8 insertion before it can reach the triplet quadrupoles in either STAR or PHENIX. Simulations have been peformed to quantify the efficiency of additional collimator jaws in RHIC. Each figure presented in this article clearly shows that the additional mask collimators provide the expected reduction in losses around the machine, and especially to the incoming triplet to the STAR experiment (IP6), for the Yellow beam as much as for the Blue beam. Looking at compiled statistics for all three working point cases studied, proton losses around the machine are reduced by roughly one order of magnitude: at most a factor 30 for magnet losses, and at most a factor 40 for losses in spaces between magnets. 12. Cryogenic Control System Goloborod' ko, S.; /Fermilab 1989-02-27 The control system (CS) for the cryogenic arrangement of the DO Liquid Argon Calorimeter consists of a Texas instruments 560/565 Programmable Logical Controller (PLC), two remote bases with Remote Base Controllers and a corresponding set of input/output (I/O) modules, and a PC AST Premium 286 (IBM AT Compatible). The PLC scans a set of inputs and provides a set of outputs based on a ladder logic program and PID control loops. The inputs are logic or analog (current, voltage) signals from equipment status switches or transducers. The outputs are logic or analog (current or voltage) signals for switching solenoids and positioning pneumatic actuators. Programming of the PLC is preformed by using the TISOFT2/560/565 package, which is installed in the PC. The PC communicates to the PLC through a serial RS232 port and provides operator interface to the cryogenic process using Xpresslink software. 13. A number of upgrades on RHIC power supply system Mi, C. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Bruno, D. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Drozd, J. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Nolan, T. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Orsatti, F. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Heppener, G. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Di Lieto, A. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Schultheiss, C. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Samms, T. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Zapasek, R. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Sandberg, J. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept. 2015-05-03 This year marks the 15th run for the Relativistic Heavy Ion Collider (RHIC). Operation of a reliable superconducting magnet power supply system is a key factor of an accelerator’s performance. Over the past 15 years, the RHIC power supply group has made many improvements to increase the machine availability and reduce failures. During these past 15 years of operating RHIC a lot of problems have been solved or addressed. In this paper some of the essential upgrades/improvements are discussed. 14. Cryogenic Systems and Superconductive Power The report defines, investigates, and experimentally evaluates the key elements of a representative crogenic turborefrigerator subsystem suitable for providing reliable long-lived cryogenic refrigeration for a superconductive ship propulsion system. 15. Preparing Accelerator Systems for the RHIC Sextant Commissioning Trbojevic, D.; Pilat, F.; Ahrens, L.; Barton, D.; Clifford, T.; Connoly, R.; Fischer, W.; Harrison, M.; Mackay, W.; Olsen, B.; Peggs, S.; Satogata, T.; Tepikian, S.; Thompson, P.; Trahern, C.; Witkover, R. 1997-05-01 The Relativistic Heavy Ion Collider (RHIC) construction is progressing steadily towards the beginning of the 1999 when beams will first be circulated in both collider rings. One of the major tests of the RHIC project is the commissioning of the first sextant with gold ion beams. This is a report on the preparation of the RHIC accelerator systems during the first sextant test, including beam position monitors, timing, injection correction through the magnetic septum and kickers, current transformers, flags'' and the ionization beam profile monitors, beam loss monitors, beam and quench permit link system, power supply controls, and the CYBASE data base system. The software and hardware development and coordination of the different systems before commissioning were regularly checked during bi-weekly, and (later) weekly, progress report meetings. 16. Power systems for the RHIC first sextant test Schultheiss, C.; Bruno, D.; Feng, P.K. [and others 1997-07-01 The first sextant test of the RHIC project is an opportunity to evaluate the many systems that must work together for the accelerator to operate. For the main dipole string, the actual main quadrupole power supply with its DSP regulator and output circuit compartment will be used. Temporary supplies will be used for the main quadrupole string, quadrupole offset, and quadrupole shunt supplies. This will let the authors both measure the performance of the main supply as well as determine the interaction among other power elements in the circuit. Correction elements will also be powered. The actual gamma-T power supplies will be used, as well as temporary supplies for the dipole correctors and sextupole supplies. Some of these units are required for beam to be transported, others are to be operated without beam to measure their performance, and how they interact with their superconducting loads. The power supply equipment, and that of other systems, required an infrastucture of AC power and output cable distribution in the RHIC tunnel, outlying service buildings, and interconnecting the tunnel to the service buildings. This note will describe the performance of the RHIC power supply systems during the sextant test, and the experience gained from this exercise. 17. Performance of the RHIC Injection Line Instrumentation Systems Shea, T. J.; Witkover, R. L.; Cameron, P.; Connolly, R.; Ryan, W. A.; Smith, G.; Zitvogel, E. 1997-05-01 The beam injection line from the Alternating Gradient Synchrotron (AGS) to the Relativistic Heavy Ion Collider (RHIC) transports proton and heavy ion bunches. This line and the RHIC first sextant currently contain thefollowing complement of beam instrumentation: stripline position monitors, ionization loss monitors, video profile monitors, and commercial current transformers. Over several years, these systems have been designed and bench tested to assure a desired performance level. The design criteria will be briefly reviewed. Then, using data from laboratory tests and the recent single pass beam tests, desired performance and attained performance will be compared. Finally, experience from the beam based tests will be applied to the design criteria for the future collider ring instrumentation. 18. The RHIC status update Ozaki, S. [Brookhaven National Lab., Upton, NY (United States) 1995-07-15 The construction of the Relativistic Heavy Ion Collider (RHIC) began in 1991, with the completion date originally scheduled for 1997. Significant reduction of the funding levels in FY 1993 and 1994, and the funding level cap for FY 1995 and later years caused a 19-month stretchout of the construction period to the second quarter of FY 1999, and an increase of the total estimated cost (TEC) to 475 M. The Project, therefore, is now at the halfway mark of the construction period with actual cost and schedule performance tracking close to the DOE-approved baseline. Construction funding through FY 1994 reached close to 60% of the TEC. Incidentally, if one adds the current value of preexisting facilities which will be incorporated into RHIC, such as the injection system (Tandem Van de Graaff - the Booster - the AGS), the esixting 3.8 km tunnel, the 24 kW helium refrigerator, etc., the total value of the RHIC facility, when completed, will reach one billion dollars, if not more. The accelerator lattice design was finalized in 1992 after an intensive study was made to optimize the collider design for performance, operational flexibility, and value engineering. The civil construciton, including the collider enclosure, magnet access ports to the ring tunnel, and six service buildings for accelerator power supplies and cryogenic control boxes was completed. 19. The cryogenic control system of BEPCⅡ LI Gang; WANG Ke-Xiang; ZHAO Ji-Jiu; YUE Ke-Juan; DAI Ming-Sui; HUANG Yi-Ling; JIANG Bo 2008-01-01 A superconducting cryogenic system has been designed and deployed in the Beijing Electron-Positron Collider Upgrade Project(BEPCⅡ).The system consists of a Siemens PLC(ST-PLC,Programmable Logic Controller)for the compressor control,an Allen Bradley(AB)PLC for the cryogenic equipments,and the Experimental Physics and Industrial Control System(EPICS)that integrates the PLCs.The system fully automates the superconducting cryogenic control with process control,PID(Proportional-Integral-Differential)control loops,real-time data access and data storage,alarm handler and human machine interface.It is capable of automatic recovery as well.This paper describes the BEPCⅡ cryogenic control system,data communication between ST-PLC and EPICS Input/Output Controllers(IOCs),and the integration of the flow control,the low level interlock,the AB-PLC,and EPICS. 20. The cryogenic control system of BEPCII Li, Gang; Wang, Ke-Xiang; Zhao, Ji-Jiu; Yue, Ke-Juan; Dai, Ming-Hui; Huang, Yi-Ling; Jiang, Bo 2008-04-01 A superconducting cryogenic system has been designed and deployed in the Beijing Electron- Positron Collider Upgrade Project (BEPCII). The system consists of a Siemens PLC (S7-PLC, Programmable Logic Controller) for the compressor control, an Allen Bradley (AB) PLC for the cryogenic equipments, and the Experimental Physics and Industrial Control System (EPICS) that integrates the PLCs. The system fully automates the superconducting cryogenic control with process control, PID (Proportional-Integral-Differential) control loops, real-time data access and data storage, alarm handler and human machine interface. It is capable of automatic recovery as well. This paper describes the BEPCII cryogenic control system, data communication between S7-PLC and EPICS Input/Output Controllers (IOCs), and the integration of the flow control, the low level interlock, the AB-PLC, and EPICS. 1. Cryogenic cooling system for HTS cable Yoshida, Shigeru [Taiyo Nippon Sanso, Tsukuba (Japan) 2017-06-15 Recently, Research and development activity of HTS (High Temperature Superconducting) power application is very progressive worldwide. Especially, HTS cable system and HTSFCL (HTS Fault current limiter) system are proceeding to practical stages. In such system and equipment, cryogenic cooling system, which makes HTS equipment cooled lower than critical temperature, is one of crucial components. In this article, cryogenic cooling system for HTS application, mainly cable, is reviewed. Cryogenic cooling system can be categorized into conduction cooling system and immersion cooling system. In practical HTS power application area, immersion cooling system with sub-cooled liquid nitrogen is preferred. The immersion cooling system is besides grouped into open cycle system and closed cycle system. Turbo-Brayton refrigerator is a key component for closed cycle system. Those two cooling systems are focused in this article. And, each design and component of the cooling system is explained. 2. Flight Performance of the AKARI Cryogenic System Nakagawa, Takao; Hirabayashi, Masayuki; Kaneda, Hidehiro; Kii, Tsuneo; Kimura, Yoshiyuki; Matsumoto, Toshio; Murakami, Hiroshi; Murakami, Masahide; Narasaki, Katsuhiro; Narita, Masanao; Ohnishi, Akira; Tsunematsu, Shoji; Yoshida, Seiji 2007-01-01 We describe the flight performance of the cryogenic system of the infrared astronomical satellite AKARI, which was successfully launched on 2006 February 21 (UT). AKARI carries a 68.5 cm telescope together with two focal plane instruments, Infrared Cameras (IRC) and Far Infrared Surveyor (FIS), all of which are cooled down to cryogenic temperature to achieve superior sensitivity. The AKARI cryogenic system is a unique hybrid system, which consists of cryogen (liquid helium) and mechanical coolers (2-stage Stirling coolers). With the help of the mechanical coolers, 179 L (26.0 kg) of super-fluid liquid helium can keep the instruments cryogenically cooled for more than 500 days. The on-orbit performance of the AKARI cryogenics is consistent with the design and pre-flight test, and the boil-off gas flow rate is as small as 0.32 mg/s. We observed the increase of the major axis of the AKARI orbit, which can be explained by the thrust due to thermal pressure of vented helium gas. 3. Cryogenic System for the Spallation Neutron Source Arenius, D.; Chronis, W.; Creel, J.; Dixon, K.; Ganni, V.; Knudsen, P. 2004-06-01 The Spallation Neutron Source (SNS) is a neutron-scattering facility being built at Oak Ridge, TN for the US Department of Energy. The SNS accelerator linac consists of superconducting radio-frequency (SRF) cavities in cryostats (cryomodules). The linac cryomodules are cooled to 2.1 K by a 2300 watt cryogenic refrigeration system. As an SNS partner laboratory, Jefferson Lab is responsible for the installed integrated cryogenic system design for the SNS linac accelerator consisting of major subsystem equipment engineered and procured from industry. Jefferson Lab's work included developing the major vendor subsystem equipment procurement specifications, equipment procurement, and the integrated system engineering support of the field installation and commissioning. The major cryogenic system components include liquid nitrogen storage, gaseous helium storage, cryogen distribution transfer line system, 2.1-K cold box consisting of four stages of cold compressors, 4.5-K cold box, warm helium compressors with its associated oil removal, gas management, helium purification, gas impurity monitoring systems, and the supportive utilities of electrical power, cooling water and instrument air. The system overview, project organization, the important aspects, and the capabilities of the cryogenic system are described. 4. Process Identification through Test on Cryogenic System Pezzetti, M; Chadli, M; Coppier, H 2008-01-01 UNICOS (UNified Industrial Control System) is the CERN object-based control standard for the cryogenics of the LHC and its experiments. It includes a variety of embedded functions, dedicated to the specific cryogenic processes. To enlarge the capabilities of the standard it is proposed to integrate the parametrical identification step in the control system of large scale cryogenic plants. Different methods of parametrical identification have been tested and the results were combined to obtain a better model. The main objective of the work is to find a compromise between an easy-to-use solution and a good level of process identification model. The study focuses on identification protocol for large delayed system, the measurement consistency and correlation between different inputs and outputs. Furthermore the paper describes in details, the results and the tests carried out on parametrical identification investigations with large scale systems. 5. Robust Multilayer Insulation for Cryogenic Systems Fesmire, J. E.; Scholtens, B. F.; Augustynowicz, S. D. 2007-01-01 New requirements for thermal insulation include robust Multilayer insulation (MU) systems that work for a range of environments from high vacuum to no vacuum. Improved MLI systems must be simple to install and maintain while meeting the life-cycle cost and thermal performance objectives. Performance of actual MLI systems has been previously shown to be much worse than ideal MLI. Spacecraft that must contain cryogens for both lunar service (high vacuum) and ground launch operations (no vacuum) are planned. Future cryogenic spacecraft for the soft vacuum environment of Mars are also envisioned. Industry products using robust MLI can benefit from improved cost-efficiency and system safety. Novel materials have been developed to operate as excellent thermal insulators at vacuum levels that are much less stringent than the absolute high vacuum requirement of current MLI systems. One such robust system, Layered Composite Insulation (LCI), has been developed by the Cryogenics Test Laboratory at NASA Kennedy Space Center. The experimental testing and development of LCI is the focus of this paper. LCI thermal performance under cryogenic conditions is shown to be six times better than MLI at soft vacuum and similar to MLI at high vacuum. The experimental apparent thermal conductivity (k-value) and heat flux data for LCI systems are compared with other MLI systems. 6. The Cryogenic Supervision System in NSRRC Li, Hsing-Chieh; Chiou, Wen-Song; Hsiao, Feng-Zone; Tsai, Zong-Da 2005-01-01 The helium cryogenic system in NSRRC is a fully automatic PLC system using the Siemens SIMATIC 300 controller. Modularization in both hardware and software makes it easy in the program reading, the system modification and the problem debug. Based on the Laview program we had developed a supervision system taking advantage of the Internet technology to get system's real-time information in any place. The functions of this supervision system include the real-time data accessing with more than 300 digital/analog signals, the data restore, the history trend display, and the human machine interface. The data is accessed via a Profibus line connecting the PLC system and the supervision system with a maximum baud rate 1.5 Mbit/s. Due to this supervision system, it is easy to master the status of the cryogenic system within a short time and diagnose the problem. 7. Testing the LHC magnet cryogenic systems Laurent Guiraud 1999-01-01 The magnets in the LHC will be cooled to 1.9 K (- 270.3°C). To keep this 27 km long machine at such a low temperatures requires one of the largest refrigeration systems in the world. These pictures show the cryogenics plant in the testing area. 8. Cryogenic Propellant Boil-Off Reduction System Plachta, D. W.; Christie, R. J.; Carlberg, E.; Feller, J. R. 2008-03-01 Lunar missions under consideration would benefit from incorporation of high specific impulse propellants such as LH2 and LO2, even with their accompanying boil-off losses necessary to maintain a steady tank pressure. This paper addresses a cryogenic propellant boil-off reduction system to minimize or eliminate boil-off. Concepts to do so were considered under the In-Space Cryogenic Propellant Depot Project. Specific to that was an investigation of cryocooler integration concepts for relatively large depot sized propellant tanks. One concept proved promising—it served to efficiently move heat to the cryocooler even over long distances via a compressed helium loop. The analyses and designs for this were incorporated into NASA Glenn Research Center's Cryogenic Analysis Tool. That design approach is explained and shown herein. Analysis shows that, when compared to passive only cryogenic storage, the boil-off reduction system begins to reduce system mass if durations are as low as 40 days for LH2, and 14 days for LO2. In addition, a method of cooling LH2 tanks is presented that precludes development issues associated with LH2 temperature cryocoolers. 9. Upgrade of RHIC Vacuum Systems for High Luminosity Operation Hseuh Hsiao Chaun; Smart, Loralie; Todd, Robert J; Weiss, Daniel 2005-01-01 With increasing ion beam intensity during recent RHIC operations, pressure rises of several decades were observed at most room temperature sections and at a few cold sections. The pressure rises are associated with electron multi-pacting, electron stimulated desorption and beam ion induced desorption and have been one of the major intensity and luminosity limiting factors for RHIC. Improvement of the warm sections has been carried out in the last few years. Extensive in-situ bakes, additional UHV pumping, anti-grazing ridges and beam tube solenoids have been implemented. Several hundred meters of NEG coated beam pipes have been installed and activated. Vacuum monitoring and interlock were enhanced to reduce premature beam aborts. Preliminary measures, such as pumping before cool down to reduce monolayer condensates, were also taken to suppress the pressure rises in the cold sections. The effectiveness of these measures in reducing the pressure rises during machine studies and during physics runs are discussed... 10. Design Tool for Cryogenic Thermal Insulation Systems Demko, Jonathan A [ORNL; Fesmire, J. E. [NASA Kennedy Space Center, Kennedy Space Center, Florida; Augustynowicz, S. D. [Sierra Lobo Inc., Kennedy Space Center, Florida 2008-01-01 Thermal isolation of low-temperature systems from ambient environments is a constant issue faced by practitioners of cryogenics. For energy-efficient systems and processes to be realized, thermal insulation must be considered as an integrated system, not merely an add-on element. A design tool to determine the performance of insulation systems for comparative trade-off studies of different available material options was developed. The approach is to apply thermal analysis to standard shapes (plane walls, cylinders, spheres) that are relatively simple to characterize with a one-dimensional analytical or numerical model. The user describes the system hot and cold boundary geometry and the operating environment. Basic outputs such as heat load and temperature profiles are determined. The user can select from a built-in insulation material database or input user defined materials. Existing information has been combined with the new experimental thermal conductivity data produced by the Cryogenics Test Laboratory for cryogenic and vacuum environments, including high vacuum, soft vacuum, and no vacuum. Materials in the design tool include multilayer insulation, aerogel blankets, aerogel bulk-fill, foams, powders, composites, and other insulation system constructions. A comparison of the design tool to a specific composite thermal insulation system is given. 11. Progress of the FAIR Cryogenic System Kauschke, M.; Kollmus, H.; Martinez-Lopez, M. 2017-02-01 The planning revision of the cryogenic system for the Facility of Antiproton and Ion Research (FAIR, Darmstadt, Germany) resulted in the choice of a single universal plant, which should provide a wide range of cryogenic operation modes, as refrigeration capacity at 4.4K, liquefaction or intermediate temperature levels. The adaptation to the FAIR specific requirements will be done later by adding a second plant. One major demand for the plant is the short term adaptation to variations in the load requirements in the system. An exemplary integration into the overall FAIR system will be shown with the experiments in the Compressed Baryonic Matter (CBM) cave. The CBM cave will house an already existing magnet, HADES, and a new magnet for the CBM experiment, which is still under design. The scheduling of the different operation modes related to the operation of the main consumers as SIS100 or SuperFRS is shown. 12. Cryogenics system: strategy to achieve nominal performance and reliable operation Bremer, J; Casas, J; Claudet, S; Delikaris, D; Delruelle, N; Ferlin, G; Fluder, C; Perin, A; Perinic, G; Pezzetti, M; Pirotte, O; Tavian, L; Wagner, U 2012-01-01 During the LHC operation in 2010 and 2011, the cryogenic system has achieved an availability level fulfilling the overall requirement. To reach this level, the cryogenic system has profited like many other beam-dependent systems from the reduced beam parameters. Therefore, impacts of some failures occurred during the LHC operation were mitigated by using the overcapacity margin, the existing built-in redundancy in between adjacent sector cryogenic plants and the "cannibalization" of spares on two idle cryogenic plants. These two first years of operation were also crucial to identify the weaknesses of the present cryogenic maintenance plan and new issues like SEUs. After the LS1, nominal beam parameters are expected and the mitigated measures will be less effective or not applicable at all. Consequently, a consolidation plan to improve the MTBF and the MTTR of the LHC cryogenic system is under definition. Concerning shutdown periods, the present cryogenic sectorization imposes some restrictions in the type of ... 13. Cryogenic Boil-Off Reduction System Plachta, David W.; Guzik, Monica C. 2014-03-01 A computational model of the cryogenic boil-off reduction system being developed by NASA as part of the Cryogenic Propellant Storage and Transfer technology maturation project has been applied to a range of propellant storage tanks sizes for high-performing in-space cryogenic propulsion applications. This effort focuses on the scaling of multi-layer insulation (MLI), cryocoolers, broad area cooling shields, radiators, solar arrays, and tanks for liquid hydrogen propellant storage tanks ranging from 2 to 10 m in diameter. Component scaling equations were incorporated into the Cryogenic Analysis Tool, a spreadsheet-based tool used to perform system-level parametric studies. The primary addition to the evolution of this updated tool is the integration of a scaling method for reverse turbo-Brayton cycle cryocoolers, as well as the development and inclusion of Self-Supporting Multi-Layer Insulation. Mass, power, and sizing relationships are traded parametrically to establish the appropriate loiter period beyond which this boil-off reduction system application reduces mass. The projected benefit compares passive thermal control to active thermal control, where active thermal control is evaluated for reduced boil-off with a 90 K shield, zero boil-off with a single heat interception stage at the tank wall, and zero boil-off with a second interception stage at a 90 K shield. Parametric studies show a benefit over passive storage at loiter durations under one month, in addition to showing a benefit for two-stage zero boil-off in terms of reducing power and mass as compared to single stage zero boil-off. Furthermore, active cooling reduces the effect of varied multi-layer insulation performance, which, historically, has been shown to be significant. 14. Design of the NIF Cryogenic Target System Gibson, C; Baltz, J; Malsbury, T; Atkinson, D; Brugmann, V; Coffield, F; Edwards, O; Haid, B; Locke, S; Shiromizu, S; Skulina, K 2008-06-10 The United States Department of Energy has embarked on a campaign to conduct credible fusion ignition experiments on the National Ignition Facility (NIF) at the Lawrence Livermore National Laboratory in 2010. The target assembly specified for this campaign requires the formation of a deuterium/tritium (DT) fuel ice layer in a 2 mm diameter capsule at the center of a 9 mm long by 5 mm diameter cylinder, called a hohlraum. The ice layer must be formed and maintained at temperatures below 20 K. At laser shot time, the target is positioned at the center of the NIF target chamber, aligned to the laser beams and held stable to less than 7 {micro}m rms. We have completed the final design of the Cryogenic Target System and are integrating the devices necessary to create, characterize and position the cryogenic target for ignition experiments. These designs, with supporting analysis and prototype test results, will be presented. 15. TUNE FEEDBACK AT RHIC CAMERON,P.; CERNIGLIA,P.; CONNOLLY,R.; CUPOLO,J.; DAWSON,W.C.; DEGEN,C.; DELLAPENNA,A.; DELONG,J.; DREES,A.; HUHN,A.; KESSELMAN,M.; MARUSIC,A.; OERTER,B.; MEAD,J.; SCHULTHEISS,C.; SIKORA,R.; VAN ZEIJTS,J. 2001-06-18 Preliminary phase-locked loop betatron tune measurement results were obtained during RHIC 2000 with a resonant Beam Position Monitor. These results suggested the possibility of incorporating PLL tune measurement into a tune feedback system for RHIC 2001. Tune feedback is useful in a superconducting accelerator, where the machine cycle time is long and inefficient acceleration due to resonance crossing is not comfortably tolerated. This is particularly true with the higher beam intensities planned for RHIC 2001. We present descriptions of a PLL tune measurement system implemented in the DSP/FPGA environment of a RHIC BPM electronics module and the feedback system into which the measurement is incorporated to regulate tune. In addition, we present results from the commissioning of this system during RHIC 2001. 16. Cryogenic Propellant Feed System Analytical Tool Development Lusby, Brian S.; Miranda, Bruno M.; Collins, Jacob A. 2011-01-01 The Propulsion Systems Branch at NASA s Lyndon B. Johnson Space Center (JSC) has developed a parametric analytical tool to address the need to rapidly predict heat leak into propellant distribution lines based on insulation type, installation technique, line supports, penetrations, and instrumentation. The Propellant Feed System Analytical Tool (PFSAT) will also determine the optimum orifice diameter for an optional thermodynamic vent system (TVS) to counteract heat leak into the feed line and ensure temperature constraints at the end of the feed line are met. PFSAT was developed primarily using Fortran 90 code because of its number crunching power and the capability to directly access real fluid property subroutines in the Reference Fluid Thermodynamic and Transport Properties (REFPROP) Database developed by NIST. A Microsoft Excel front end user interface was implemented to provide convenient portability of PFSAT among a wide variety of potential users and its ability to utilize a user-friendly graphical user interface (GUI) developed in Visual Basic for Applications (VBA). The focus of PFSAT is on-orbit reaction control systems and orbital maneuvering systems, but it may be used to predict heat leak into ground-based transfer lines as well. PFSAT is expected to be used for rapid initial design of cryogenic propellant distribution lines and thermodynamic vent systems. Once validated, PFSAT will support concept trades for a variety of cryogenic fluid transfer systems on spacecraft, including planetary landers, transfer vehicles, and propellant depots, as well as surface-based transfer systems. The details of the development of PFSAT, its user interface, and the program structure will be presented. 17. Cryogenic system for a superconducting spectrometer Porter, J. 1983-03-01 The Heavy Ion Spectrometer System (HISS) relies upon superconducting coils of cryostable, pool boiling design to provide a maximum particle bending field of 3 tesla. The cryogenic facility including helium refrigeration, gas management, liquid nitrogen system, and the overall control strategy are described. The system normally operates with a 4 K heat load of 150 watts; the LN2 circuits absorb an additional 4000 watts. The 80K intercept control is by an LSI 11 computer. Total available refrigeration at 4K is 400 watts using reciprocating expanders at the 20K and 4K level. The minicomputer has the capability of optimizing overall utility input cost by varying operating points. A hybrid of pneumatic, analog, and digital control is successful in providing full time unattended operation. The 7m diameter magnet/cryostat assembly is rotatable through 180 degrees to provide a variety of spectrometer orientations. 18. The dipole corrector magnets for the RHIC fast global orbit feedback system Thieberger, P.; Arnold, L.; Folz, C.; Hulsart, R.; Jain, A.; Karl, R.; Mahler, G.; Meng, W.; Mernick, K.; Michnoff, R.; Minty, M.; Montag, C.; Ptitsyn, V.; Ritter, J.; Smart, L.; Tuozzolo, J.; White, J. 2011-03-28 The recently completed RHIC fast global orbit feedback system uses 24 small 'window-frame' horizontal dipole correctors. Space limitations dictated a very compact design. The magnetic design and modelling of these laminated yoke magnets is described as well as the mechanical implementation, coil winding, vacuum impregnation, etc. Test procedures to determine the field quality and frequency response are described. The results of these measurements are presented and discussed. A small fringe field from each magnet, overlapping the opposite RHIC ring, is compensated by a correction winding placed on the opposite ring's magnet and connected in series with the main winding of the first one. Results from measurements of this compensation scheme are shown and discussed. 19. Status Of Superconducting Radiofrequency Separator Cryogenic System Ageyev, A; Kashtanov, E; Kozub, S; Muraviev, M; Orlov, A; Pimenov, P; Polkovnikov, K; Slabodchikov, P; Sytnik, V V; Zintchenko, S 2004-01-01 The OKA experimental complex proposing to use the technique of RF beam separation to produce a Kaon beam is under construction at IHEP. Two deflecting superconducting niobium cavities operating at 1.8 K are the basic elements of the separator. To provide cooling at this temperature commercially available 500 W, 4.5 K helium refrigerator is used to cool liquid helium bath of the satellite refrigerator. The last one is actually a big warm up heat exchanger with flow imbalance and very low pressure drop. Vacuum group consists of two stages of roots blowers and one stage of rotary slide valve pumps. Pump stages are separated by intermediate gas coolers. The schematic, thermodynamics, design capacity and current construction status of the cryogenic system are presented. 20. Status of the Cryogenic System Commissioning at SNS Casagrande, Fabio; Campisi, Isidoro E; Creel, Jonathan; Dixon, Kelly; Ganni, Venkatarao; Gurd, Pamela; Hatfield, Daniel; Howell, Matthew; Knudsen, Peter; Stout, Daniel; Strong, William 2005-01-01 The Spallation Neutron Source (SNS) is under construction at Oak Ridge National Laboratory. The cold section of the Linac consists of 81 superconducting radio frequency cavities cooled to 2.1K by a 2400 Watt cryogenic refrigeration system. The major cryogenic system components include warm helium compressors with associated oil removal and gas management, 4.5K cold box, 7000L liquid helium dewar, 2.1K cold box (consisting of 4 stages of cold compressors), gaseous helium storage, helium purification and gas impurity monitoring system, liquid nitrogen storage and the cryogenic distribution transfer line system. The overall system commissioning strategy and status will be presented. 1. Status of the Cryogenic System Commissioning at SNS F. Casagrande; I.E. Campisi; P.A. Gurd; D.R. Hatfield; M.P. Howell; D. Stout; W.H. Strong; D. Arenius; J.C. Creel; K. Dixon; V. Ganni; P.K. Knudsen 2005-05-16 The Spallation Neutron Source (SNS) is under construction at Oak Ridge National Laboratory. The cold section of the Linac consists of 81 superconducting radio frequency cavities cooled to 2.1K by a 2400 Watt cryogenic refrigeration system. The major cryogenic system components include warm helium compressors with associated oil removal and gas management, 4.5K cold box, 7000L liquid helium dewar, 2.1K cold box (consisting of 4 stages of cold compressors), gaseous helium storage, helium purification and gas impurity monitoring system, liquid nitrogen storage and the cryogenic distribution transfer line system. The overall system commissioning strategy and status will be presented. 2. Hard scattering of partons as a probe of collisions at RHIC using the STAR detector system Christie, W.B. [Brookhaven National Lab., Upton, NY (United States) 1995-07-15 Presented here is the current state of the authors investigations into the use of hard probes to study pp, pA, and AA collisions at the Relativistic Heavy Ion Collider (RHIC) being built at Brookhaven National Laboratory. The overall goal of the RHIC program is the discovery and study of the Quark-Gluon Plasma (QGP), which is predicted to be formed at the high energy densities reached at RHIC in high energy AA collisions. The term {open_quotes}Hard probes{close_quotes} as used in this document includes those particles whose origin is the result of a direct hard parton scatter (i.e qq, qg, or gg). The final states of these hard parton scatters which the author proposes to study include dijets, gamma-jet coincidences, and inclusive high P{sub t} particle spectra. A brief discussion of the physics objectives is given in section 1. This is followed by an introduction to the STAR detector system in section 2, with particular details given for the proposed STAR Electromagnetic Calorimeter (EMC). The present simulation studies and results are given in section 3. The author concludes with a summary and a discussion of future plans in section 4. 3. Cryogenic system operating experience review for fusion applications Cadwallader, L.C. 1992-01-01 This report presents a review of cryogenic system operating experiences, from particle accelerator, fusion experiment, space research, and other applications. Safety relevant operating experiences and accident information are discussed. Quantitative order-of-magnitude estimates of cryogenic component failure rates and accident initiating event frequencies are presented for use in risk assessment, reliability, and availability studies. Safety concerns with cryogenic systems are discussed, including ozone formation, effects of spills, and modeling spill behavior. This information should be useful to fusion system designers and safety analysts, such as the team working on the International Thermonuclear Experimental Reactor design. 4. Evaluation of Stirling cooler system for cryogenic CO2 capture Song, Chun Feng; Kitamura, Yutaka; Li, Shu Hong 2012-01-01 In previous research, a cryogenic system based on Stirling coolers has been developed. In this work, the novel system was applied on CO2 capture from post-combustion flue gas and different process parameters (i.e. flow rate of feed gas, temperature of Stirling cooler and operating condition) were investigated to obtain the optimal performance (CO2 recovery and energy consumption). From the extensive experiments, it was concluded that the cryogenic system could realize CO2 capture without solv... 5. Cryogenic and Vacuum Compatible Metrology Systems Project National Aeronautics and Space Administration — In this Phase I SBIR project for NASA, Flexure Engineering of Greenbelt, MD will leverage the work we did in our current SBIR project entitled: Cryogenic Optical... 6. FRIB Cryogenic Distribution System and Status Ganni, Venkatarao [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Dixon, Kelly D. [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Laverdure, Nathaniel A. [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Yang, Shuo [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Nellis, Timothy [Michigan State Univ., East Lansing, MI (United States); Jones, S. [Michigan State Univ., East Lansing, MI (United States); Casagrande, Fabio [Michigan State Univ., East Lansing, MI (United States) 2015-12-01 The MSU-FRIB cryogenic distribution system supports the 2 K primary, 4 K primary, and 35 - 55 K shield operation of more than 70 loads in the accelerator and the experimental areas. It is based on JLab and SNS experience with bayonet-type disconnects between the loads and the distribution system for phased commissioning and maintenance. The linac transfer line, which features three separate transfer line segments for additional independence during phased commissioning at 4 K and 2 K, connects the folded arrangement of 49 cryomodules and 4 superconducting dipole magnets and a fourth transfer line supports the separator area cryo loads. The pressure reliefs for the transfer line process lines, located in the refrigeration room outside the tunnel/accelerator area, are piped to be vented outdoors. The transfer line designs integrate supply and return flow paths into a combined vacuum space. The main linac distribution segments are produced in a small number of standard configurations; a prototype of one such configuration has been fabricated at Jefferson Lab and has been installed at MSU to support testing of a prototype FRIB cryomodule. 7. Cold Helium Gas Pressurization For Spacecraft Cryogenic Propulsion Systems Morehead, Robert L.; Atwell. Matthew J.; Hurlbert, Eric A.; Melcher, J. C. 2017-01-01 To reduce the dry mass of a spacecraft pressurization system, helium pressurant may be stored at low temperature and high pressure to increase mass in a given tank volume. Warming this gas through an engine heat exchanger prior to tank pressurization both increases the system efficiency and simplifies the designs of intermediate hardware such as regulators, valves, etc. since the gas is no longer cryogenic. If this type of cold helium pressurization system is used in conjunction with a cryogenic propellant, though, a loss in overall system efficiency can be expected due to heat transfer from the warm ullage gas to the cryogenic propellant which results in a specific volume loss for the pressurant, interpreted as the Collapse Factor. Future spacecraft with cryogenic propellants will likely have a cold helium system, with increasing collapse factor effects as vehicle sizes decrease. To determine the collapse factor effects and overall implementation strategies for a representative design point, a cold helium system was hotfire tested on the Integrated Cryogenic Propulsion Test Article (ICPTA) in a thermal vacuum environment at the NASA Glenn Research Center Plum Brook Station. The ICPTA vehicle is a small lander-sized spacecraft prototype built at NASA Johnson Space Center utilizing cryogenic liquid oxygen/liquid methane propellants and cryogenic helium gas as a pressurant to operate one 2,800lbf 5:1 throttling main engine, two 28lbf Reaction Control Engines (RCE), and two 7lbf RCEs (Figure 1). This vehicle was hotfire tested at a variety of environmental conditions at NASA Plum Brook, ranging from ambient temperature/simulated high altitude, deep thermal/high altitude, and deep thermal/high vacuum conditions. A detailed summary of the vehicle design and testing campaign may be found in Integrated Cryogenic Propulsion Test Article Thermal Vacuum Hotfire Testing, AIAA JPC 2017. 8. Workshop Goals: Mini-workshop on RHIC Rf system Hahn, H. [Brookhaven National Lab. (BNL), Upton, NY (United States) 1988-07-11 The objectives of this study are to: review rf system requirements; review cavity and amplifier hardware; establish basis for revised CDR text and cost estimate; and provide a written summary and technical notes. 9. Post Mortem System - Playback of the RHIC Collider Laster, J S; D'Ottavio, T; Marusic, A; Skelly, J F 2001-01-01 A Post Mortem System was developed for the Relativistic Heavy Ion Collider at Brookhaven National Laboratory to provide a playback of the collider state at the time of a beam abort, quench, or other failure event. Post Mortem data is used to provide diagnostics about the failure and to improve future stores. This data is read from hardware buffers and is written directly to the main file system by Accelerator Device Objects in the front-end computers. The Post Mortem System has facilitated analysis of loss monitor and power supply data, such as beam loss during magnet quenches, dump kicker misfires and power supply malfunctions. System details and recent operating experience will be discussed. 10. Tank System Integrated Model: A Cryogenic Tank Performance Prediction Program Bolshinskiy, L. G.; Hedayat, A.; Hastings, L. J.; Sutherlin, S. G.; Schnell, A. R.; Moder, J. P. 2017-01-01 Accurate predictions of the thermodynamic state of the cryogenic propellants, pressurization rate, and performance of pressure control techniques in cryogenic tanks are required for development of cryogenic fluid long-duration storage technology and planning for future space exploration missions. This Technical Memorandum (TM) presents the analytical tool, Tank System Integrated Model (TankSIM), which can be used for modeling pressure control and predicting the behavior of cryogenic propellant for long-term storage for future space missions. Utilizing TankSIM, the following processes can be modeled: tank self-pressurization, boiloff, ullage venting, mixing, and condensation on the tank wall. This TM also includes comparisons of TankSIM program predictions with the test data andexamples of multiphase mission calculations. 11. SynapSense Wireless Environmental Monitoring System of the RHIC & ATLAS Computing Facility at BNL Casella, K.; Garcia, E.; Hogue, R.; Hollowell, C.; Strecker-Kellogg, W.; Wong, A.; Zaytsev, A. 2014-06-01 RHIC & ATLAS Computing Facility (RACF) at BNL is a 15000 sq. ft. facility hosting the IT equipment of the BNL ATLAS WLCG Tier-1 site, offline farms for the STAR and PHENIX experiments operating at the Relativistic Heavy Ion Collider (RHIC), the BNL Cloud installation, various Open Science Grid (OSG) resources, and many other small physics research oriented IT installations. The facility originated in 1990 and grew steadily up to the present configuration with 4 physically isolated IT areas with the maximum rack capacity of about 1000 racks and the total peak power consumption of 1.5 MW. In June 2012 a project was initiated with the primary goal to replace several environmental monitoring systems deployed earlier within RACF with a single commercial hardware and software solution by SynapSense Corporation based on wireless sensor groups and proprietary SynapSense™ MapSense™ software that offers a unified solution for monitoring the temperature and humidity within the rack/CRAC units as well as pressure distribution underneath the raised floor across the entire facility. The deployment was completed successfully in 2013. The new system also supports a set of additional features such as capacity planning based on measurements of total heat load, power consumption monitoring and control, CRAC unit power consumption optimization based on feedback from the temperature measurements and overall power usage efficiency estimations that are not currently implemented within RACF but may be deployed in the future. 12. CONFIGURATION MANUAL POLARIZED PROTON COLLIDER AT RHIC. ROSER,T.; MACKAY,W.W.; ALEKSEEV,I.; BAI,M.; BROWN,K.; BUNCE,G.; CAMERON,P.; COURANT,E.; ET AL. 2001-03-01 In this report, the authors present their design to accelerate and store polarized protons in RHIC, with the level of polarization, luminosity, and control of systematic errors required by the approved RHIC spin physics program. They provide an overview of the physics to be studied using RHIC with polarized proton beams, and a brief description of the accelerator systems required for the project. 13. Configuration Manual Polarized Proton Collider at RHIC Alekseev, I.; Allgower, C.; Bai, M.; Batygin, Y.; Bozano, L.; Brown, K.; Bunce, G.; Cameron, P.; Courant, E.; Erin, S.; Escallier, J.; Fischer, W.; Gupta, R.; Hatanka, K.; Huang, H.; Imai, K.; Ishihara, M.; Jain, A.; Kanavets, V.; Katayama, T.; Kawaguchi, T.; Kelly, E.; Kurita, K.; Lee, S. Y.; Luccio, A.; MacKay, W. W.; Mahler, G.; Makdisi, Y.; Mariam, F.; McGahern, W.; Morgan, G.; Muratore, J.; Okamura, M.; Peggs, S.; Pilat, F.; Ptitsin, V.; Ratner, L.; Roser, T.; Saito, N.; Satoh, H.; Shatunov, Y.; Spinka, H.; Svirida, D.; Syphers, M.; Tepikian, S.; Tominaka, T.; Tsoupas, N.; Underwood, D.; Vasiliev, A.; Wanderer, P.; Willen, E.; Wu, H.; Yokosawa, A.; Zelenski, A. 2006-01-01 In this report we present our design to accelerate and store polarized protons in RHIC, with the level of polarization, luminosity, and control of systematic errors required by the approved RHIC spin physics program. We provide an overview of the physics to be studied using RHIC with polarized proton beams, and a brief description of the accelerator systems required for the project. 14. System Size Dependence of Transverse Momentum Correlations at RHIC Adamczyk, L; Agakishiev, G; Aggarwal, M M; Ahammed, Z; Alekseev, I; Alford, J; Anson, C D; Aparin, A; Arkhipkin, D; Aschenauer, E; Averichev, G S; Balewski, J; Banerjee, A; Barnovska, Z; Beavis, D R; Bellwied, R; Betancourt, M J; Betts, R R; Bhasin, A; Bhati, A K; Bhattarai,; Bichsel, H; Bielcik, J; Bielcikova, J; Bland, L C; Bordyuzhin, I G; Borowski, W; Bouchet, J; Brandin, A V; Brovko, S G; Bruna, E; Bültmann, S; Bunzarov, I; Burton, T P; Butterworth, J; Caines, H; Sánchez, M Calderón de la Barca; Cebra, D; Cendejas, R; Cervantes, M C; Chaloupka, P; Chang, Z; Chattopadhyay, S; Chen, H F; Chen, J H; Chen, J Y; Chen, L; Cheng, J; Cherney, M; Chikanian, A; Christie, W; Chung, P; Chwastowski, J; Codrington, M J M; Corliss, R; Cramer, J G; Crawford, H J; Cui, X; Das, S; Leyva, A Davila; De Silva, L C; Debbe, R R; Dedovich, T G; Deng, J; de Souza, R Derradi; Dhamija, S; di Ruzza, B; Didenko, L; Dilks,; Ding, F; Dion, A; Djawotho, P; Dong, X; Drachenberg, J L; Draper, J E; Du, C M; Dunkelberger, L E; Dunlop, J C; Efimov, L G; Elnimr, M; Engelage, J; Engle, K S; Eppley, G; Eun, L; Evdokimov, O; Fatemi, R; Fazio, S; Fedorisin, J; Fersch, R G; Filip, P; Finch, E; Fisyak, Y; Flores, C E; Gagliardi, C A; Gangadharan, D R; Garand, D; Geurts, F; Gibson, A; Gliske, S; Grebenyuk, O G; Grosnick, D; Guo, Y; Gupta, A; Gupta, S; Guryn, W; Haag, B; Hajkova, O; Hamed, A; Han, L-X; Haque, R; Harris, J W; Hays-Wehle, J P; Heppelmann, S; Hirsch, A; Hoffmann, G W; Hofman, D J; Horvat, S; Huang, B; Huang, H Z; Huck, P; Humanic, T J; Igo, G; Jacobs, W W; Jena, C; Judd, E G; Kabana, S; Kang, K; Kauder, K; Ke, H W; Keane, D; Kechechyan, A; Kesich, A; Kikola, D P; Kiryluk, J; Kisel, I; Kisiel, A; Koetke, D D; Kollegger, T; Konzer, J; Koralt, I; Korsch, W; Kotchenda, L; Kravtsov, P; Krueger, K; Kulakov, I; Kumar, L; Kycia, R A; Lamont, M A C; Landgraf, J M; Landry, K D; LaPointe, S; Lauret, J; Lebedev, A; Lednicky, R; Lee, J H; Leight, W; LeVine, M J; Li, C; Li, W; Li, X; Li, Y; Li, Z M; Lima, L M; Lisa, M A; Liu, F; Ljubicic, T; Llope, W J; Longacre, R S; Luo, X; Ma, G L; Ma, Y G; Don, D M M D Madagodagettige; Mahapatra, D P; Majka, R; Margetis, S; Markert, C; Masui, H; Matis, H S; McDonald, D; McShane, T S; Mioduszewski, S; Mitrovski, M K; Mohammed, Y; Mohanty, B; Mondal, M M; Munhoz, M G; Mustafa, M K; Naglis, M; Nandi, B K; Nasim, Md; Nayak, T K; Nelson, J M; Nogach, L V; Novak, J; Odyniec, G; Ogawa, A; Oh, K; Ohlson, A; Okorokov, V; Oldag, E W; Oliveira, R A N; Olson, D; Pachr, M; Page, B S; Pal, S K; Pan, Y X; Pandit, Y; Panebratsev, Y; Pawlak, T; Pawlik, B; Pei, H; Perkins, C; Peryt, W; Pile, P; Planinic, M; Pluta, J; Plyku, D; Poljak, N; Porter, J; Poskanzer, A M; Powell, C B; Pruneau, C; Pruthi, N K; Przybycien, M; Pujahari, P R; Putschke, J; Qiu, H; Ramachandran, S; Raniwala, R; Raniwala, S; Ray, R L; Riley, C K; Ritter, H G; Roberts, J B; Rogachevskiy, O V; Romero, J L; Ross, J F; Roy, A; Ruan, L; Rusnak, J; Sahoo, N R; Sahu, P K; Sakrejda, I; Salur, S; Sandacz, A; Sandweiss, J; Sangaline, E; Sarkar, A; Schambach, J; Scharenberg, R P; Schmah, A M; Schmidke, B; Schmitz, N; Schuster, T R; Seger, J; Seyboth, P; Shah, N; Shahaliev, E; Shao, M; Sharma, B; Sharma, M; Shen, W Q; Shi, S S; Shou, Q Y; Sichtermann, E P; Singaraju, R N; Skoby, M J; Smirnov, D; Smirnov, N; Solanki, D; Sorensen, P; deSouza, U G; Spinka, H M; Srivastava, B; Stanislaus, T D S; Stevens, J R; Stock, R; Strikhanov, M; Stringfellow, B; Suaide, A A P; Suarez, M C; Sumbera, M; Sun, X M; Sun, Y; Sun, Z; Surrow, B; Svirida, D N; Symons, T J M; de Toledo, A Szanto; Takahashi, J; Tang, A H; Tang, Z; Tarini, L H; Tarnowsky, T; Thomas, J H; Timmins, A R; Tlusty, D; Tokarev, M; Trentalange, S; Tribble, R E; Tribedy, P; Trzeciak, B A; Tsai, O D; Turnau, J; Ullrich, T; Underwood, D G; Van Buren, G; van Nieuwenhuizen, G; Vanfossen,, J A; Varma, R; Vasconcelos, G M S; Vertesi, R; Videbæk, F; Viyogi, Y P; Vokal, S; Voloshin, S A; Vossen, A; Wada, M; Walker, M; Wang, F; Wang, G; Wang, H; Wang, J S; Wang, Q; Wang, X L; Wang, Y; Webb, G; Webb, J C; Westfall, G D; Wieman, H; Wissink, S W; Witt, R; Wu, Y F; Xiao, Z; Xie, W; Xin, K; Xu, H; Xu, N; Xu, Q H; Xu, W; Xu, Y; Xu, Z; Yan,; Yang, C; Yang, Y; Yepes, P; Yi, L; Yip, K; Yoo, I-K; Zawisza, Y; Zbroszczyk, H; Zha, W; Zhang, J B; Zhang, S; Zhang, X P; Zhang, Y; Zhang, Z P; Zhao, F; Zhao, J; Zhong, C; Zhu, X; Zhu, Y H; Zoulkarneeva, Y; Zyzak, M 2013-01-01 We present a study of the average transverse momentum (p_t$) fluctuations and$p_t$correlations for charged particles produced in Cu+Cu collisions at midrapidity for$\\sqrt{s_{NN}} =$62.4 and 200 GeV. These results are compared with those published for Au+Au collisions at same energies, to explore the system size dependence. In addition to the collision energy and system size dependence, the$p_t$correlations results have been studied as functions of the collision centralities, the ranges in$p_t$, the pseudo-rapidity$\\eta$, and the azimuthal angle$\\phi$, for which the correlations are measured. The square root of the measured$p_t$correlations when scaled by mean-$p_t$are found to be independent of both colliding beam energy and system size studied. The transport based model calculations are found to have a better quantitative agreement with the measurements compared to models which incorporate only jet-like correlations. 15. Fast Automated Decoupling at RHIC Beebe-Wang, Joanne 2005-01-01 Coupling correction is essential for the operational performance of RHIC. The independence of the transverse degrees of freedom makes diagnostics and tune control easier, and it is advantageous to operate an accelerator close to the coupling resonance to minimize nearby nonlinear sidebands. An automated decoupling application has been developed at RHIC for coupling correction during routine operations. The application decouples RHIC globally by minimizing the tune separation through finding the optimal settings of two orthogonal skew quadrupole families. The program provides options of automatic, semi-automatic and manual decoupling operations. It accesses tune information from all RHIC tune measurement systems: the PLL (Phase Lock Loop), the high frequency Schottky system, and the tune meter. It also supplies tune and skew quadrupole scans, finding the minimum tune separation, display the real time results and interface with the RHIC control system. We summarize the capabilities of the decoupling application... 16. SRF Test Areas Cryogenic System Controls Graphical User Interface DeGraff, B.D.; Ganster, G.; Klebaner, A.; Petrov, A.D.; Soyars, W.M.; /Fermilab 2011-06-09 Fermi National Accelerator Laboratory has constructed a superconducting 1.3 GHz cavity test facility at Meson Detector Building (MDB) and a superconducting 1.3 GHz cryomodule test facility located at the New Muon Lab Building (NML). The control of these 2K cryogenic systems is accomplished by using a Synoptic graphical user interface (GUI) to interact with the underlying Fermilab Accelerator Control System. The design, testing and operational experience of employing the Synoptic client-server system for graphical representation will be discussed. Details on the Synoptic deployment to the MDB and NML cryogenic sub-systems will also be discussed. The implementation of the Synoptic as the GUI for both NML and MDB has been a success. Both facilities are currently fulfilling their individual roles in SCRF testing as a result of successful availability of the cryogenic systems. The tools available for creating Synoptic pages will continue to be developed to serve the evolving needs of users. 17. Loss maps of RHIC Robert-Demolaize,G. 2007-10-01 State-of-the-art tracking tools were recently developed at CERN to study the cleaning efficiency of the Large Hadron Collider (LHC) collimation system [1]. These tools are fully transportable, meaning that any accelerator lattice that includes a collimation system can be simulated. Each of the two Relativistic Heavy Ion Collider (RHIC) [2] beam lines features a multi-stage collimation system, therefore dedicated datasets from RHIC operations with proton beams can be used to benchmark the tracking codes and assess the accuracy of the predicted hot spots along the LHC. 18. Recent results on collective effects in small systems from PHENIX at RHIC Campbell, Sarah 2016-01-01 Collisions of simple systems, such as$p$+$p$, or$p$+Nucleus have been used as benchmarks for our understanding of heavy ion collsions, since it was assumed they would be free of the effects from hot nuclear matter. Recently long range correlations and anisotropies of momentum spectra have been seen in such collisions, challenging this assumption. Such phenomena have been understood to be the result of the collective motion, which can best be described by hydrodynamics, whose initial conditions are set by the geometry of the colliding systems, together with their fluctuations. This talk will discuss the recent results from the PHENIX experiment at RHIC using a variety of colliding species ($p$+Au,$d$+Au,$^{3}$He+Au) that give a better understanding of the origin of the observed correlations and anisotropies, thus providing insight as to whether a quark gluon plasma is formed in these simple systems. 19. Cryogenic compressive properties of basic epoxy resin systems Markley, F.W.; Hoffman, J.A.; Muniz, D.P. 1985-09-01 The compressive properties of short cylindrical samples of many different epoxy resin systems have been measured at ambient temperature and at 77/sup 0/K. These are pure resin systems of known chemistry, without the inorganic fillers or fibrous reinforcements needed in final cryogenic systems. Of course, chemically incorporated modifiers such as flexibilizing resins have been included. This data should make possible inferences about cryogenic properties from molecular structures and provide specific data useful to formulators and end users. Measurements on some other plastics such as PTFE, Polyimides, and UHMWPE have been made for comparison purposes. 20. Advanced monitoring, fault diagnostics, and maintenance of cryogenic systems Girone, Mario; Pezzetti, Marco In this Thesis, advanced methods and techniques of monitoring, fault diagnostics, and predictive maintenance for cryogenic processes and systems are described. In particular, in Chapter 1, mainstreams in research on measurement systems for cryogenic processes are reviewed with the aim of dening key current trends and possible future evolutions. Then, in Chapter 2, several innovative methods are proposed. A transducer based on a virtual ow meter is presented for monitoring helium distribution and consumption in cryogenic systems for particle accelerators [1]. Furthermore, a comprehensive metrological analysis of the proposed transducer for verifying the metrological performance and pointing out most critical uncertainty sources is described [2]. A model-based method for fault detection and early-stage isolation, able to work with few records of Frequency Response Function (FRF) on an unfaulty compressor, is then proposed [3]. To enrich the proposal, a distributed diagnostic procedure, based on a micro-genetic... 1. Cryogenic Liquid Sample Acquisition System for Remote Space Applications Mahaffy, Paul; Trainer, Melissa; Wegel, Don; Hawk, Douglas; Melek, Tony; Johnson, Christopher; Amato, Michael; Galloway, John 2013-01-01 There is a need to acquire autonomously cryogenic hydrocarbon liquid sample from remote planetary locations such as the lakes of Titan for instruments such as mass spectrometers. There are several problems that had to be solved relative to collecting the right amount of cryogenic liquid sample into a warmer spacecraft, such as not allowing the sample to boil off or fractionate too early; controlling the intermediate and final pressures within carefully designed volumes; designing for various particulates and viscosities; designing to thermal, mass, and power-limited spacecraft interfaces; and reducing risk. Prior art inlets for similar instruments in spaceflight were designed primarily for atmospheric gas sampling and are not useful for this front-end application. These cryogenic liquid sample acquisition system designs for remote space applications allow for remote, autonomous, controlled sample collections of a range of challenging cryogenic sample types. The design can control the size of the sample, prevent fractionation, control pressures at various stages, and allow for various liquid sample levels. It is capable of collecting repeated samples autonomously in difficult lowtemperature conditions often found in planetary missions. It is capable of collecting samples for use by instruments from difficult sample types such as cryogenic hydrocarbon (methane, ethane, and propane) mixtures with solid particulates such as found on Titan. The design with a warm actuated valve is compatible with various spacecraft thermal and structural interfaces. The design uses controlled volumes, heaters, inlet and vent tubes, a cryogenic valve seat, inlet screens, temperature and cryogenic liquid sensors, seals, and vents to accomplish its task. 2. Vacuum and cryogenic system for the MUSE detectors Lizon, J. L.; Accardo, M.; Gojak, Domingo; Reiss, Roland; Kern, Lothar 2012-09-01 MUSE with its 24 detectors distributed over an eight square meter vertical area was requiring a well engineered and extremely reliable cryogenic system. The solution should also use a technology proven to be compatible with the very high sensitivity of the VLT interferometer. A short introduction reviews the various available technologies to cool these 24 chips down to 160 K. The first part of the paper presents the selected concept insisting on the various advantages offered by LN2. In addition to the purely vacuum and cryogenic aspects we highlight some of the most interesting features given by the control system based on a PLC. 3. Large Cryogenics Systems at 1.8 K Tavian, L 2000-01-01 Cryogenics is now widely present in large accelerator projects using applied superconductivity. Economical considerations permanently require an increase of the performance of superconducting devices. One way to do this consists to lower their operating temperature and to cool them with superfluid helium. For this purpose, large cryogenic systems at 1.8 K producing refrigeration capacity in the kW range have to be developed and implemented. These cryogenic systems require large pumping capacity at very low pressure based on integral cold compression or mixed cold-warm compression. This paper describes and compares the different cooling methods with saturated or pressurised superfluid helium, gives the present status of the available process machinery with their practical performance, and reviews the different thermodynamical cycles for producing refrigeration below 2 K, with emphasis on their operational compliance. 4. Designing of epoxy resin systems for cryogenic use Ueki, T.; Nishijima, S.; Izumi, Y. 2005-02-01 The mechanical and thermal properties of several types of epoxy systems were designed based on the chemical structure, network structure and morphology aiming at cryogenic application. In this research di-epoxies or multifunctional epoxies were cured by several kinds of hardeners such as anhydride, amine or phenol and were blended with polycarbonate, carboxyl-terminated butadiene acrylonitrile copolymer or phenoxy. The mechanical properties and thermal properties of these cured epoxies were measured at room and liquid nitrogen temperature. It was found that the two-dimensional network structured linear polymer shows high performance even at cryogenic temperature. It was concluded that the controls of the structures are very important to optimize epoxy systems for cryogenic application. 5. RHIC prefire protection masks Drees, A.; Biscardi, C.; Curcio, T.; Gassner, D.; DeSanto, L.; Fu, W.; Liaw, C. J.; Montag, C.; Thieberger, P.; Yip, K. 2015-01-07 The protection of the RHIC experimental detectors from damage due to beam hitting close upstream elements in cases of abort kicker prefires requires some dedicated precautionary measures with two general options: to bring the beam close to a limiting aperture (i.e. the beam pipe wall), as far upstream of the detector components as possible or, alternatively, to bring a limiting aperture close to the circulating beam. Spontaneous and random prefires of abort kicker modules (Pulse Forming Network, PFN) have a history as long as RHIC is being operated. The abort system consist of 5 kickers in per ring, each of them equipped with its own dedicated PFN. 6. RHIC physics Pajares, C 1999-01-01 A brief review of the hadronic phase transitions is presented by emphasizing the physical ideas and the main signatures of the transition in relation to the most significant results of the SPS experiments and the description of the RHIC experiments. (77 refs). 7. Beam injection into RHIC Fischer, W.; Hahn, H.; MacKay, W.W.; Satogata, T.; Tsoupas, N.; Zhang, W. 1997-07-01 During the RHIC sextant test in January 1997 beam was injected into a sixth of one of the rings for the first time. The authors describe the injection zone and its bottlenecks. They report on the commissioning of the injection system, on beam based measurements of the kickers and the application program to steer the beam. 8. G2 Autonomous Control for Cryogenic Delivery Systems Dito, Scott J. 2014-01-01 The Independent System Health Management-Autonomous Control (ISHM-AC) application development for cryogenic delivery systems is intended to create an expert system that will require minimal operator involvement and ultimately allow for complete autonomy when fueling a space vehicle in the time prior to launch. The G2-Autonomous Control project is the development of a model, simulation, and ultimately a working application that will control and monitor the cryogenic fluid delivery to a rocket for testing purposes. To develop this application, the project is using the programming language/environment Gensym G2. The environment is an all-inclusive application that allows development, testing, modeling, and finally operation of the unique application through graphical and programmatic methods. We have learned G2 through training classes and subsequent application development, and are now in the process of building the application that will soon be used to test on cryogenic loading equipment here at the Kennedy Space Center Cryogenics Test Laboratory (CTL). The G2 ISHM-AC application will bring with it a safer and more efficient propellant loading system for the future launches at Kennedy Space Center and eventually mobile launches from all over the world. 9. Low heat-gain cryogenic-liquid transfer system Hows, G. E.; Wright, B. J. 1970-01-01 Cryogenic-liquid transfer system, containing a ring structure with tensioned small diameter, high strength wires, provides adequate physical support for the piping, minimizes the conductive heat paths between the piping and jacket, and allows for thermal expansion and contraction of the piping. 10. Spiral 2 Cryogenic System for The Superconducting LINAC Ghribi, A.; Bernaudin, P.-E.; Bert, Y.; Commeaux, C.; Houeto, M.; Lescalié, G. 2017-02-01 SPIRAL 21 is a rare isotope accelerator dedicated to the production of high intensity beams (E = 40 MeV, I = 5 mA). The driver is a linear accelerator (LINAC) that uses bulk Niobium made quarter wave RF cavities. 19 cryomodules inclose one or two cavities respectively for the low and the high energy sections. To supply the 1300 W at 4.2 K required to cool down the LINAC, a cryogenic system has been set up. The heart of the latter is a 3 turbines geared HELIAL®LF (ALAT2) cold box that delivers both the liquid helium for the cavities and the 60 K Helium gaz for the thermal screens. 19 valve-boxes insure cryogenic fluid distribution and management. Key issues like cool down speed or cavity RF frequency stability are closely linked to the cryogenic system management. To overcome these issues, modelling and simulation efforts are being undertaken prior to the first cool down trials. In this paper, we present a status update of the Spiral 2 cryogenic system and the cool down strategy considered for its commissioning. 11. Feasibility Study of Cryogenic Cutting Technology by Using a Computer Simulation and Manufacture of Main Components for Cryogenic Cutting System Kim, Sung Kyun; Lee, Dong Gyu; Lee, Kune Woo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of); Song, Oh Seop [Chungnam National University, Deajeon (Korea, Republic of) 2009-06-15 Cryogenic cutting technology is one of the most suitable technologies for dismantling nuclear facilities due to the fact that a secondary waste is not generated during the cutting process. In this paper, the feasibility of cryogenic cutting technology was investigated by using a computer simulation. In the computer simulation, a hybrid method combined with the SPH (smoothed particle hydrodynamics) method and the FE (finite element) method was used. And also, a penetration depth equation, for the design of the cryogenic cutting system, was used and the design variables and operation conditions to cut a 10 mm thickness for steel were determined. Finally, the main components of the cryogenic cutting system were manufactures on the basis of the obtained design variables and operation conditions. 12. The integrated cryogenic system for the atmospheric vertical interferometric detector on FY-4 satellite Wu, Yinong; Liu, EnGuang; Jiang, Zhenhua; Yang, Baoyu; Mu, Yongbin 2016-05-01 The cryogenic system for the atmospheric vertical interferometric detector on FY-4 satellite includes a Stirling cryocooler, a radiant cooler, a cryogenic heat pipe and some flexible thermal links as well. These cryogenic elements were integrated together in order to decrease the background radiation and maximize the sensitivity with high efficiency and high reliability. This paper summarizes the cryogenic integration design, technical challenges, and the results of thermal and performance testing. 13. Mobile remote surveillance system for the CERN LHC cryogenic system Torbjørn, Houge 2006-01-01 This thesis documents the research, planning and partial implementation of a remote surveillance system for use in the CERN LHC machine tunnel. The system is planned to provide surveillance of the cryogenic system in the LHC, eliminating the need for the personnel to go personally to look at a piece of possible faulty equipment. For this project, the complete system is planned. The system will be controlled via an Ethernet connection. This is fed into a 400V power line as a powerline communication signal, and picked up by the surveillance system. Then it is decoded into an Ethernet signal again, and sent to a camera with an on board web server. The power is transported by the same powered rail as used for communication, so that the system can take power and communication along the whole tunnel. The thesis describes the goals of the system and explains the requirements it needs to meet. Several solutions, especially technologies for communication, are considered, and details about them are described. A solutio... 14. Cryogenic system for VECC K500 superconducting cyclotron Pal, G; Bhattacharyya, T K; Bhandari, R K 2009-01-01 VEC Centre, Kolkata in India is at an advanced stage of commissioning a K500 superconducting cyclotron. The superconducting coil of the magnet for cyclotron is cooled by liquid helium. Three liquid helium cooled cryopanels, placed inside the Dees of the radiofrequency system, maintain the vacuum in the acceleration region of the superconducting cyclotron. The cryogenic system for magnet for cyclotron has been tested by cooling the coil and energizing the magnet. The cryogenic system for cryopanels has also been tested. Heater and temperature sensor were placed on the liquid helium cold head for cryopanel. The temperature of the cold head was observed to be below 20 K upto a heat load of 11.7 watt. 15. Validation and Performance of the LHC Cryogenic System through Commissioning of the First Sector Serio, L; Casas-Cubillos, J; Chakravarty, A; Claudet, S; Gicquel, F; Gomes, P; Kumar, M; Kush, PK; Millet, F; Perin, A; Rabehl, R; Singh, MR; Soubiran, M; Tavian, L 2008-01-01 The cryogenic system [1] for the Large Hadron Collider accelerator is presently in its final phase of commissioning at nominal operating conditions. The refrigeration capacity for the LHC is produced using eight large cryogenic plants and eight 1.8 K refrigeration units installed on five cryogenic islands. Machine cryogenic equipment is installed in a 26.7-km circumference ring deep underground tunnel and are maintained at their nominal operating conditions via a distribution system consisting of transfer lines, cold interconnection boxes at each cryogenic island and a cryogenic distribution line. The functional analysis of the whole system during all operating conditions was established and validated during the first sector commissioning in order to maximize the system availability. Analysis, operating modes, main failure scenarios, results and performance of the cryogenic system are presented. 16. Conference on Refrigeration for Cryogenic Sensors and Electronic Systems Sullivan, D B; McCarthy, S E; Cryogenic Refrigeration Conference; International Cryocooler Conference; Cryocoolers 1 1981-01-01 This proceedings documents the output of a meeting of refrigeration specialists held at the National Bureau of Standards, Boulder, CO, on October 6 and 7, 1980. Building on an earlier invitation-only meeting in 1977, the purpose of this first open meeting was to discuss progress in the development of refrigeration systems to cool cryogenic sensors and electronic systems in the temperature range below 20 K and with required cooling capacities below 10 W. The meeting was jointly sponsored by the International Institute of Refrigeration - Commission A1/2, the Office of Naval Research, the Naval Research Laboratory, the Cryogenic Engineering Conference, and the National Bureau of Standards. This first open cryocooler conference consisted of 23 papers presented by representatives of industry, government, and academia. The conference proceedings reproduced here was published by the National Bureau of Standards in Boulder, Colorado as NBS Special Publication #607. Subsequent meetings would become known as the Intern... 17. Test results after refurbish of cryogenic system for smiles Otsuka, Kiyomi; Tsunematsu, Shoji; Okabayashi, Akinobu; Narasaki, Katsuhiro; Satoh, Ryota 2010-09-01 Superconducting Sub-millimeter-wave Limb-Emission Sounder (SMILES) is to be operated aboard the Japanese Experiment Module (JEM) of the International Space Station (ISS) in 2009. SMILES uses two superconductor-insulator-superconductor (SIS) mixers for sub-millimeter-wave atmospheric observation and they are cooled to 4 K levels by a cryogenic system with a two-stage Stirling cooler, a Joule-Thomson (JT) cycle cooler and a cryostat composed of three stages. Two-stage Stirling cooler precools the JT circuit and also cools radiation shields in the cryostat. JT circuit has three tube-in-tube type heat exchangers and an orifice for JT expansion in the cryostat. The cryogenic system is built, tested and delivered. 18. Dynamic simulations of the cryogenic system of a tokamak Cirillo, R.; Hoa, C.; Michel, F.; Poncet, J. M.; Rousset, B. 2015-12-01 Power generation in the next decades could be provided by thermo-nuclear fusion reactors like tokamaks. There inside, the fusion reaction takes place thanks to the generation of plasmas at hundreds of millions of degrees that must be confined magnetically with superconductive coils, cooled down to 4.4K. The plasma works cyclically and the coil system is subjected to pulsed heat load which has to be handled by the refrigerator. By smoothing the variable loads, the refrigerator capacity can be set close to the average power; optimizing investment and operational costs. Within the “Broader Approach agreement” related to ITER project, CEA (Commissariat a l'Energie Atomique et aux Energies Alternatives) is in charge of providing the cryogenic system for the Japanese tokamak (JT-60SA), that is currently under construction in Naka. The system has been designed to handle the pulsed heat loads. To prepare the acceptance tests of the cryogenic system foreseen in 2016, both dynamic modelling and experimental tests on a scaled down mock-up are of high interest for assessing pulsed load smoothing control. After explaining HELIOS (HElium Loop for hIgh lOad Smoothing) operating modes, a dynamic model is presented, with results on the pulsed heat load scenarios. All the simulations have been performed with EcosimPro® and the associated cryogenic library CRYOLIB. 19. Array controller system with cryogenic pre-amplifiers for MIMIZUKU Okada, K.; Sako, S.; Miyata, T.; Kamizuka, T.; Ohsawa, R.; Uchiyama, M. S.; Mori, K.; Yamaguchi, J.; Asano, K.; Uchiyama, M. 2016-07-01 MIMIZUKU is a mid-infrared imager and spectrograph being developed for the University of Tokyo Atacama Observatory (TAO) 6.5-m telescope (PI: Y. Yoshii). To fully utilize a high atmospheric transmission of the Chajnantor site, MIMIZUKU covers a wide wavelength range from 2 to 38 μm with three array detectors: a HAWAII-1RG HgCdTe 1024 × 1024 array with a 5 μm cutoff manufactured by Teledyne, an Aquarius Si:As IBC 1024 × 1024 array by Raytheon, and a MF-128 Si:Sb BIB 128 × 128 array by DRS. We have newly developed an array controller system to operate these multiple arrays. A sampling rate higher than 0.5 MHz is required to prevent from saturation of their wells in broad-band imaging observations with MIMIZUKU due to high thermal background flux. Such high speed signals are dulled when passing through lines from the arrays to readout circuits. To overcome this problem, we have developed high-speed cryogenic buffer pre-amplifier circuits with commercial GaAs MESFETs, instead of Si JFETs, which are generally used in buffer amplifiers at cryogenic temperatures. The cryogenic buffer circuits are installed on an outer wall of the optical bench of MIMIZUKU at 20 K. We have measured readout noises of the array controller system including the cryogenic buffers in a test cryostat and room temperature circuits and confirmed that input referred noises of the system are lower than the specification value of the readout noise of the Aquarius array. 20. Experience with the String2 Cryogenic Instrumentation and Control System Gomes, P; Blanco-Viñuela, E; Casas-Cubillos, J; Pelletier, S; Rodríguez, M A; Serio, L; Suraci, A; Vauthier, N 2005-01-01 String2 was a 120 m full-scale model of a regular cell of the LHC accelerator arc. It was composed of eight superconducting main magnets, fed by a separate cryogenic distribution line (QRL); an electrical feed box (DFB) which supported the superconducting current leads that powered the magnets [1]. Bearing an intensive experimental programme, String2 was heavily instrumented. The cryogenic instrumentation and control system, whose complexity was close to a full 3.3 km LHC sector, have already been described in [2]. String2 was a useful facility to validate design choices and to gain knowledge on installation and commissioning procedures. This paper reports on the experience of four years of designing, installation, commissioning and maintenance, and outlines the lessons learned for the LHC. 1. The Control System for the Cryogenics in the LHC Tunnel Gomes, P; Antoniotti, F; Avramidou, R; Balle, Ch; Blanco-Viñuela, E; Carminati, Ch; Casas-Cubillos, J; Ciechanowski, M; Dragoneas, A; Dubert, P; Fampris, X; Fluder, C; Fortescue, E; Gaj, W; Gousiou, E; Jeanmonod, N; Jodłowski, P; Karagiannis, F; Klisch, M; López, A; Macuda, P; Malinowski, P; Molina, E; Paiva, S; Patsouli, A; Penacoba, G; Sosin, M; Soubiran, M; Suraci, A; Tovar, A; Vauthier, N; Wolak, T; Zwalinski, L 2009-01-01 The Large Hadron Collider makes extensive use of superconductors, in magnets for bending and focusing the particles, and in RF cavities for accelerating them, which are operated at 1.9 K and 4.5 K. The process automation for the cryogenic distribution around the accelerator circumference is based on 16 Programmable Logic Controllers, each running 250 control loops, 500 alarms and interlocks, and a phase sequencer. Spread along 27 km and under ionizing radiation, 15 000 cryogenic sensors and actuators are accessed through industrial field networks. We describe the main hardware and software components of the control system, their deployment and commissioning, together with the project organization, challenges faced, and solutions found. 2. The Detector Calibration System for the CUORE cryogenic bolometer array Cushman, J S; Davis, C J; Ejzak, L; Lenz, D; Lim, K E; Heeger, K M; Maruyama, R H; Nucciotti, A; Sangiorgio, S; Wise, T 2016-01-01 The Cryogenic Underground Observatory for Rare Events (CUORE) is a ton-scale cryogenic experiment designed to search for neutrinoless double-beta decay of$^{130}$Te and other rare events. The CUORE detector consists of 988 TeO$_2bolometers operated underground at 10~mK in a dilution refrigerator at the Laboratori Nazionali del Gran Sasso. Candidate events are identified through a precise measurement of their energy. The absolute energy response of the detectors is established by the regular calibration of each individual bolometer using gamma sources. The close-packed configuration of the CUORE bolometer array combined with the extensive shielding surrounding the detectors requires the placement of calibration sources within the array itself. The CUORE Detector Calibration System is designed to insert radioactive sources into and remove them from the cryostat while respecting the stringent heat load, radiopurity, and operational requirements of the experiment. This paper describes the design, commissioning... 3. Rectification of the OPAL Cold Neutron Source Cryogenic System Lu, Weijian [Australian Nuclear Science and Technology Organisation, Sydney (Australia) 2013-07-01 The Cold Neutron Source (CNS) at ANSTO's OPAL Reactor had experienced repeated outages since 2009 due to failures in the cryogenic system. An extensive root cause analysis was initiated in May 2012, led by an ANSTO team that also involved knowledgeable external experts. At the conclusion of the investigation, a set of recommendations was released to address the identified contributing causes. A rectification program was established to implement the solutions. Cryogenic operation of the CNS, providing end users with cold neutrons, successfully returned to service in July 2013. Thanks to the unique stand-by operation mode of the CNS, irradiation activities at the reactor, as well as thermal neutron availability, had not been affected during the year-long investigation/rectification process. Some technical and operational aspects of the investigation, testing and engineering modifications are discussed in this presentation. 4. Cryogenic systems advanced monitoring, fault diagnostics, and predictive maintenance Arpaia, Pasquale 2017-01-01 Cryogenics, the study and technology of materials and systems at very low temperature, is widely used for sensors and instruments requiring very highly precise measurements with low electrical resistance, especially for measurements of materials and energies at a very small scale. Thus, the need to understand how instruments operate and perform over time at temperatures below -2920 F (-1800 C) is critical, for applications from Magnetic Resonance Imaging (MRI) to Nuclear Magnetic Resonance Spectroscopy to instrumentation for particle accelerators of all kinds. This book brings to the reader guidance learned from work at the European Laboratory for Nuclear Research (CERN), and its large scale particle accelerator in Switzerland to help engineers and technicians implement best practices in instrumentation at cryogenic temperatures, including a better understanding of fault detection and predictive maintenance. Special problems with devices like flow meters, pressure gauges, and temperature gauges when operating... 5. Cryogenic system for the MYRRHA superconducting linear accelerator Chevalier, Nicolas R.; Junquera, Tomas [Accelerators and Cryogenic Systems, 86, rue de Paris, 91400 Orsay (France); Thermeau, Jean-Pierre [Institut de Physique Nucléaire, Université Paris Sud, 91400 Orsay (France); Romão, Luis Medeiros; Vandeplassche, Dirk [SCK-CEN, Boeretang 200, 2400 Mol (Belgium) 2014-01-29 SCK⋅CEN, the Belgian Nuclear Research Centre, is designing MYRRHA, a flexible fast spectrum research reactor (80 MW{sub th}), conceived as an accelerator driven system (ADS), able to operate in sub-critical and critical modes. It contains a continuous-wave (CW) superconducting (SC) proton accelerator of 600 MeV, a spallation target and a multiplying core with MOX fuel, cooled by liquid lead-bismuth (Pb-Bi). From 17 MeV onward, the SC accelerator will consist of 48 β=0.36 spoke-loaded cavities (352 MHz), 34 β=0.47 elliptical cavities (704 MHz) and 60 β=0.65 elliptical cavities (704 MHz). We present an analysis of the thermal loads and of the optimal operating temperature of the cryogenic system. In particular, the low operating frequency of spoke cavities makes their operation in CW mode possible both at 4.2 K or at 2 K. Our analysis outlines the main factors that determine at what temperature the spoke cavities should be operated. We then present different cryogenic fluid distribution schemes, important characteristics (storage, transfer line, etc.) and the main challenges offered by MYRRHA in terms of cryogenics. 6. Molecular absorption cryogenic cooler for liquid hydrogen propulsion systems Klein, G. A.; Jones, J. A. 1982-01-01 A light weight, long life molecular absorption cryogenic cooler (MACC) system is described which can use low temperature waste heat to provide cooling for liquid hydrogen propellant tanks for interplanetary spacecraft. Detailed tradeoff studies were made to evaluate the refrigeration system component interactions in order to minimize the mass of the spacecraft cooler system. Based on this analysis a refrigerator system mass of 31 kg is required to provide the .48 watts of cooling required by a 2.3 meter diameter liquid hydrogen tank. 7. Autonomous Systems: Autonomous Cryogenic Loading Operations Project National Aeronautics and Space Administration — The main objectives are to develop and integrate Integrated Systems Health Management (ISHM) tools and component technologies into a seamless health management... 8. NSLS-II RF Cryogenic System Rose, J.; Dilgen, T.; Gash, B.; Gosman, J.; Mortazavi, P.; Papu, J.; Ravindranath, V.; Sikora, R.; Sitnikov, A.; Wilhelm, H.; Jia, Y.; Monroe, C. 2015-05-03 The National Synchrotron Light Source II is a 3 GeV X-ray user facility commissioned in 2014. A new helium refrigerator system has been installed and commissioned to support the superconducting RF cavities in the storage ring. Special care was taken to provide very stable helium and LN2 pressures and flow rates to minimize microphonics and thermal effects at the cavities. Details of the system design along with commissioning and early operations data will be presented. 9. Cryogenic System for a High Temperature Superconducting Power Transmission Cable Demko, J.A.; Gouge, M.J.; Hughey, R.L.; Lue, J.W.; Martin, R.; Sinha, U.; Stovall, J.P. 1999-07-12 High-temperature superconducting (HTS) cable systems for power transmission are under development that will use pressurized liquid nitrogen to provide cooling of the cable and termination hardware. Southwire Company and Oak Ridge National Laboratory have been operating a prototype HTS cable system that contains many of the typical components needed for a commercial power transmission application. It is being used to conduct research in the development of components and systems for eventual commercial deployment. The cryogenic system was built by Air Products and Chemicals, Allentown, Pennsylvania, and can circulate up to 0.35 kg/s of liquid nitrogen at temperatures as low as 67 K at pressures of 1 to 10 bars. Sufficient cooling is provided for testing a 5-m-long HTS transmission cable system that includes the terminations required for room temperature electrical connections. Testing of the 5-m HTS transmission cable has been conducted at the design ac conditions of 1250 A and 7.5 kV line to ground. This paper contains a description of the essential features of the HTS cable cryogenic system and performance results obtained during operation of the system. The salient features of the operation that are important in large commercial HTS cable applications will be discussed. 10. ATLAS magnet common cryogenic, vacuum, electrical and control systems Miele, P; Delruelle, N; Geich-Gimbel, C; Haug, F; Olesen, G; Pengo, R; Sbrissa, E; Tyrvainen, H; ten Kate, H H J 2004-01-01 The superconducting Magnet System for the ATLAS detector at the LHC at CERN comprises a Barrel Toroid, two End Cap Toroids and a Central Solenoid with overall dimensions of 20 m diameter by 26 m length and a stored energy of 1.6 GJ. Common proximity cryogenic and electrical systems for the toroids are implemented. The Cryogenic System provides the cooling power for the 3 toroid magnets considered as a single cold mass (600 tons) and for the CS. The 21 kA toroid and the 8 kA solenoid electrical circuits comprise both a switch-mode power supply, two circuit breakers, water cooled bus bars, He cooled current leads and the diode resistor ramp-down unit. The Vacuum System consists of a group of primary rotary pumps and sets of high vacuum diffusion pumps connected to each individual cryostat. The Magnet Safety System guarantees the magnet protection and human safety through slow and fast dump treatment. The Magnet Control System ensures control, regulation and monitoring of the operation of the magnets. The update... 11. The refrigeration and cryogenic distribution system for the shortpulse x-ray source Green, Michael A.; Corlett, John N. 2002-10-20 This report describes the essential elements of the cryogenic system. The cryogenic distribution system starts at the level of the linac superconducting RF cavities [1] and moves out through the cryogenic piping to the liquid helium refrigeration plant that will be used to cool the RF cavities and the undulator magnets. For this report, the cryogenic distribution system and cryogenic refrigerator includes the following elements: (1) The piping within the linac cryogenic modules will influence the heat transfer through the super-fluid helium from the outer surface of the TESLA niobium cavity and the liquid to gas interface within the horizontal header pipe where the superfluid helium boils. This piping determines the final design of the linac cryogenic module. (2) The acceptable pressure drops determine the supply and return piping dimensions. (3) The helium distribution system is determined by the need to cool down and warm up the various elements in the light source. (4) The size of the cryogenic plant is determined by the heat loads and the probable margin of error on those heat loads. Since the final heat loads are determined by the acceleration gradient in the cavities, a linac with five cryogenic modules will be compared to a linac with only four cryogenic modules. The design assumes that all cryogenic elements in the facility will be cooled using a common cryogenic plant. To minimize vibration effects on the beam lines, this plant is assumed to be located some distance from the synchrotron light beam lines. All of the cryogenic elements in the facility will be attached to the helium refrigeration system through cryogenic transfer lines. The largest single cryogenic load is the main linac, which consists of four or five cryogenic modules depending on the design gradient for the cavities in the linac section. The second largest heat load comes from the cryogenic modules that contain the transverse deflecting RF cavities. The injector linac is the third largest 12. Beam Induced Pressure Rise at RHIC Zhang, S Y; Bai, Mei; Blaskiewicz, Michael; Cameron, Peter; Drees, Angelika; Fischer, Wolfram; Gullotta, Justin; He, Ping; Hseuh Hsiao Chaun; Huang, Haixin; Iriso, Ubaldo; Lee, Roger C; Litvinenko, Vladimir N; MacKay, William W; Nicoletti, Tony; Oerter, Brian; Peggs, Steve; Pilat, Fulvia Caterina; Ptitsyn, Vadim; Roser, Thomas; Satogata, Todd; Smart, Loralie; Snydstrup, Louis; Thieberger, Peter; Trbojevic, Dejan; Wang, Lanfa; Wei, Jie; Zeno, Keith 2005-01-01 Beam induced pressure rise in RHIC warm sections is currently one of the machine intensity and luminosity limits. This pressure rise is mainly due to electron cloud effects. The RHIC warm section electron cloud is associated with longer bunch spacings compared with other machines, and is distributed non-uniformly around the ring. In addition to the countermeasures for normal electron cloud, such as the NEG coated pipe, solenoids, beam scrubbing, bunch gaps, and larger bunch spacing, other studies and beam tests toward the understanding and counteracting RHIC warm electron cloud are of interest. These include the ion desorption studies and the test of anti-grazing ridges. For high bunch intensities and the shortest bunch spacings, pressure rises at certain locations in the cryogenic region have been observed during the past two runs. Beam studies are planned for the current 2005 run and the results will be reported. 13. Fermilab Muon Campus g-2 Cryogenic Distribution Remote Control System Pei, L; Klebaner, A; Soyars, W; Bossert, R 2015-01-01 The Muon Campus (MC) is able to measure Muon g-2 with high precision and comparing its value to the theoretical prediction. The MC has four 300 KW screw compressors and four liquid helium refrigerators. The centerpiece of the Muon g-2 experiment at Fermilab is a large, 50-foot-diameter superconducting muon storage ring. This one-of-a-kind ring, made of steel, aluminum and superconducting wire, was built for the previous g-2 experiment at Brookhaven. Due to each subsystem has to be far away from each other and be placed in the distant location, therefore, Siemens Process Control System PCS7-400, Automation Direct DL205 & DL05 PLC, Synoptic and Fermilab ACNET HMI are the ideal choices as the MC g-2 cryogenic distribution real-time and on-Line remote control system. This paper presents a method which has been successfully used by many Fermilab distribution cryogenic real-time and On-Line remote control systems. 14. Fermilab Muon Campus g-2 Cryogenic Distribution Remote Control System Pei, L.; Theilacker, J.; Klebaner, A.; Soyars, W.; Bossert, R. 2015-11-05 The Muon Campus (MC) is able to measure Muon g-2 with high precision and comparing its value to the theoretical prediction. The MC has four 300 KW screw compressors and four liquid helium refrigerators. The centerpiece of the Muon g-2 experiment at Fermilab is a large, 50-foot-diameter superconducting muon storage ring. This one-of-a-kind ring, made of steel, aluminum and superconducting wire, was built for the previous g-2 experiment at Brookhaven. Due to each subsystem has to be far away from each other and be placed in the distant location, therefore, Siemens Process Control System PCS7-400, Automation Direct DL205 & DL05 PLC, Synoptic and Fermilab ACNET HMI are the ideal choices as the MC g-2 cryogenic distribution real-time and on-Line remote control system. This paper presents a method which has been successfully used by many Fermilab distribution cryogenic real-time and On-Line remote control systems. 15. Layered composite thermal insulation system for nonvacuum cryogenic applications Fesmire, J. E. 2016-03-01 A problem common to both space launch applications and cryogenic propulsion test facilities is providing suitable thermal insulation for complex cryogenic piping, tanks, and components that cannot be vacuum-jacketed or otherwise be broad-area-covered. To meet such requirements and provide a practical solution to the problem, a layered composite insulation system has been developed for nonvacuum applications and extreme environmental exposure conditions. Layered composite insulation system for extreme conditions (or LCX) is particularly suited for complex piping or tank systems that are difficult or practically impossible to insulate by conventional means. Consisting of several functional layers, the aerogel blanket-based system can be tailored to specific thermal and mechanical performance requirements. The operational principle of the system is layer-pairs working in combination. Each layer pair is comprised of a primary insulation layer and a compressible radiant barrier layer. Vacuum-jacketed piping systems, whether part of the ground equipment or the flight vehicle, typically include numerous terminations, disconnects, umbilical connections, or branches that must be insulated by nonvacuum means. Broad-area insulation systems, such as spray foam or rigid foam panels, are often the lightweight materials of choice for vehicle tanks, but the plumbing elements, feedthroughs, appurtenances, and structural supports all create "hot spot" areas that are not readily insulated by similar means. Finally, the design layouts of valve control skids used for launch pads and test stands can be nearly impossible to insulate because of their complexity and high density of components and instrumentation. Primary requirements for such nonvacuum thermal insulation systems include the combination of harsh conditions, including full weather exposure, vibration, and structural loads. Further requirements include reliability and the right level of system breathability for thermal 16. A new method for flow measurement in cryogenic systems Grohmann, S. 2014-03-01 A new method for mass flow measurement of fluids in pipes is presented; its novelty lies in the capability for intrinsic calibration. The method is founded on a concept, where two independent analytic expressions for the flow rate are formed from the same direct measurement readings (input parameters). If the input parameters were error-free, the two expressions would yield identical results, by definition. This fact can be used as goal function in a minimization routine that removes systematic errors of the inherently error-prone input parameters. The uncertainty of the mass flow measurement is then only influenced by statistical effects and is typically less than 1% with regard to the measured value. The new method is explained by a proof-of-principle that is based on measurements in a large-scale cryogenic system. The intrinsic calibrations can be executed in situ at any moment during operation of a plant, and with no need for a reference standard. While the new method is applicable in any system involving single-phase fluid flow, it offers particular advantages in cryogenic application. 17. The detector calibration system for the CUORE cryogenic bolometer array Cushman, Jeremy S.; Dally, Adam; Davis, Christopher J.; Ejzak, Larissa; Lenz, Daniel; Lim, Kyungeun E.; Heeger, Karsten M.; Maruyama, Reina H.; Nucciotti, Angelo; Sangiorgio, Samuele; Wise, Thomas 2017-02-01 The Cryogenic Underground Observatory for Rare Events (CUORE) is a ton-scale cryogenic experiment designed to search for neutrinoless double-beta decay of 130Te and other rare events. The CUORE detector consists of 988 TeO2 bolometers operated underground at 10 mK in a dilution refrigerator at the Laboratori Nazionali del Gran Sasso. Candidate events are identified through a precise measurement of their energy. The absolute energy response of the detectors is established by the regular calibration of each individual bolometer using gamma sources. The close-packed configuration of the CUORE bolometer array combined with the extensive shielding surrounding the detectors requires the placement of calibration sources within the array itself. The CUORE Detector Calibration System is designed to insert radioactive sources into and remove them from the cryostat while respecting the stringent heat load, radiopurity, and operational requirements of the experiment. This paper describes the design, commissioning, and performance of this novel source calibration deployment system for ultra-low-temperature environments. 18. Integration and characterization of the cryogenic system of MEGARA Ferrusca R., D.; Cisneros G., M. E.; Velázquez, M.; Zenteno H., J. A.; Gil de Paz, A.; Gallego, J.; Carrasco, E.; Sánchez-Moreno, F. C.; Iglesias-Páramo, J. 2016-08-01 MEGARA (Multi-Espectrógrafo en GTC de Alta Resolución para Astronomía) is an optical Integral-Field Unit and Multi-Object Spectrograph designed for the GTC (Gran Telescopio de Canarias) 10.4m telescope in La Palma, it is expected that the spectrograph will be delivered to GTC towards the end of 2016. MEGARA includes an open cycle cryostat which harbors the scientific CCD of the instrument at an operating temperature of 153 K, this cryogenic system has been designed and integrated by the "Astronomical Instrumentation Lab for Millimeter Wavelengths" at the Instituto Nacional de Astrofísica, Óptica y Electrónica (INAOE) in Mexico. Early this year the cryostat has finished its fabrication and now it is on AIV phases, in this paper we report the cryostat CCD-head and dewar back integration, vacuum and cryogenic test results are also reported. The final integration of the cryostat with the other components of the instrument is taking place at LICA lab at the Universidad Complutense de Madrid. 19. The XRS Low Temperature Cryogenic System: Ground Performance Test Results Breon, Susan; Sirron, Peter; Boyle, Robert; Canavan, Ed; DiPirro, Michael; Serlemitsos, Aristides; Tuttle, James; Whitehouse, Paul 1998-01-01 The X-Ray Spectrometer (XRS) instrument is part of the Astro-E mission scheduled to launch early in 2000. Its cryogenic system is required to cool a 32-element square array of x-ray microcalorimeters to 60-65 mK over a mission lifetime of at least 2 years. This is accomplished using an adiabatic demagnetization refrigerator (ADR) contained within a two-stage superfluid helium/solid neon cooler. Goddard Space Flight Center is providing the ADR and helium dewar. The flight system was assembled in Sept. 1997 and subjected to extensive thermal performance tests. This paper presents test results at both the system and component levels. In addition, results of the low temperature topoff performed in Japan with the engineering unit neon and helium dewars are discussed. 20. Cryogenic control system of the large COMPASS polarized target Gautheron, F; Baum, G; Berglund, P; Doshita, N; Görtz, S; Gustafsson, K K; Horikawa, N; Kisselev, Yu V; Koivuniemi, J H; Kondo, K; Meyer, Werner T; Reicherz, G 2004-01-01 The dilution refrigerator used to cool the large COMPASS polarized target is monitored through a PC running LabVIEW trademark 6.1 under Windows 2000 trademark . About 60 parameters of the target (temperatures, pressures, flow rates) are continuously plotted and checked. They are periodically recorded in an Oracle trademark database and in a data file. An alarm for every parameter can be individually activated and optionally connected to a GSM (Global System for Mobile Communication) delivery message system. A web server receives and publishes the online status of the target with online tables and graphics on a dedicated COMPASS polarized target information web site. A Siemens programmable logic controller (PLC) powered by an uninterruptable source keeps the cryogenic system safe and stable during the long beam periods by controlling valves and interlocks. This safety feature protects the dilution refrigerator against potential damages in case of power failure. 1. Liquid-Vapor Equilibrium of Multicomponent Cryogenic Systems Thompson, W. Reid; Calado, Jorge C. G.; Zollweg, John A. 1990-01-01 Liquid-vapor and solid-vapor equilibria at low to moderate pressures and low temperatures are important in many solar system environments, including the surface and clouds of Titan, the clouds of Uranus and Neptune, and the surfaces of Mars and Triton. The familiar cases of ideal behavior are limiting cases of a general thermodynamic representation for the vapor pressure of each component in a homogeneous multicomponent system. The fundamental connections of laboratory measurements to thermodynamic models are through the Gibbs-Duhem relation and the Gibbs-Helmholtz relation. Using laboratory measurements of the total pressure, temperature, and compositions of the liquid and vapor phases at equilibrium, the values of these parameters can be determined. The resulting model for vapor-liquid equilibrium can then conveniently and accurately be used to calculate pressures, compositions, condensation altitudes, and their dependencies on changing climatic conditions. A specific system being investigated is CH4-C2H6-N2, at conditions relevant to Titan's surface and atmosphere. Discussed are: the modeling of existing data on CH4-N2, with applications to the composition of Titan's condensate clouds; some new measurements on the CH4-C2H6 binary, using a high-precision static/volumetric system, and on the C2H6-N2 binary, using the volumetric system and a sensitive cryogenic flow calorimeter; and describe a new cryogenic phase-equilibrium vessel with which we are beginning a detailed, systematic study of the three constituent binaries and the ternary CH4-C2H6-N2 system at temperatures ranging from 80 to 105 K and pressures from 0.1 to 7 bar. 2. High density matter at RHIC Thomas S Ullrich 2004-02-01 QCD predicts a phase transition between hadronic matter and a quark-gluon plasma at high energy density. The relativistic heavy ion collider (RHIC) at Brookhaven National Laboratory is a new facility dedicated to the experimental study of matter under extreme conditions. Already the first round of experimental results at RHIC indicated that the conditions to create a new state of matter are indeed reached in the collisions of heavy nuclei. Studies of particle spectra and their correlations at low transverse momenta provide evidence of strong pressure gradients in the highly interacting dense medium and hint that we observe a system in thermal equilibrium. Recent runs with high statistics allow us to explore the regime of hard-scattering processes where the suppression of hadrons at large transverse momentum, and quenching of di-jets are observed thus providing further evidence for extreme high density matter created in collisions at RHIC. 3. Control and operation cost optimization of the HISS cryogenic system Porter, J.; Bieser, F.; Anderson, D. 1983-08-01 The Heavy Ion Spectrometer System (HISS) relies upon superconducting coils of cryostable design to provide a maximum particle bending field of 3 tesla. A previous paper describes the cryogenic facility including helium refrigeration and gas management. A control strategy which has allowed full time unattended operation, along with significant nitrogen and power cost reductions is discussed. Reduction of liquid nitrogen consumption was accomplished by using the sensible heat available in the cold exhaust gas. Measured nitrogen throughput agrees with calculations for sensible heat utilization of zero to 70%. Calculated consumption saving over this range is 40 liters per hour for conductive losses to the supports only. It is found that the measured throughput differential for the total system is higher. 4. Control and operation cost optimization of the HISS cryogenic system Porter, J.; Bieser, F.; Anderson, D. 1983-08-01 The Heavy Ion Spectrometer System (HISS) relies upon superconducting coils of cryostable design to provide a maximum particle bending field of 3 tesla. A previous paper describes the cryogenic facility including helium refrigeration and gas management. This paper discusses a control strategy which has allowed full time unattended operation, along with significant nitrogen and power cost reductions. Reduction of liquid nitrogen consumption has been accomplished by making use of the sensible heat available in the cold exhaust gas. Measured nitrogen throughput agrees with calculations for sensible heat utilization of zero to 70%. Calculated consumption saving over this range is 40 liters per hour for conductive losses to the supports only. The measured throughput differential for the total system is higher. 5. Analysis of the cryogenic system behavior for pulsed heat load in EAST Hu, L. B.; Zhuang, M.; Zhou, Z. W.; Xia, G. H. 2014-01-01 EAST is the first full superconducting fusion device. The plasma is confined by the magnetic fields generated from a large set of superconducting magnets which are made of cable in-conduit conductor (CICC). In operation, these magnets suffer heat loads from thermal and nuclear radiation from the surrounding components and plasma as well as the eddy currents and the AC losses generated within the magnets, together with the heat conduction through supports and the resistive heat generated at the current lead transiting to room temperature. The cryogenic system of our EAST consists of a 2kW/4K helium refrigerator and a distribution system for the cooling of poloidal field (PF) and toroidal field (TF) coils, structures, thermal shields, buslines and current leads. Pulsed heat load is the main difference between the cryogenic system of a full superconducting Tokamak system and other large scale cryogenic systems. The cryogenic system operates in a pulsed heat loads mode requiring the helium refrigerator to remove periodically large heat loads in time. At the same time, the cryogenic system parameters such as helium cooling superconducting magnets, helium refrigerator and helium distribution system are changing. In this paper, the variation range of the parameters of superconducting magnets and refrigerator has been analyzed in the typical plasma discharge mode. The control scheme for the pulsed loads characteristics of the cryogenic system has been proposed, the implementation of which helps to smooth the pulse loads and to improve the stability of the operation of the cryogenic system. 6. General purpose multiplexing device for cryogenic microwave systems Chapman, Benjamin J.; Moores, Bradley A.; Rosenthal, Eric I.; Kerckhoff, Joseph; Lehnert, K. W. 2016-05-01 We introduce and experimentally characterize a general purpose device for signal processing in circuit quantum electrodynamics systems. The device is a broadband two-port microwave circuit element with three modes of operation: it can transmit, reflect, or invert incident signals between 4 and 8 GHz. This property makes it a versatile tool for lossless signal processing at cryogenic temperatures. In particular, rapid switching (≤ 15 ns ) between these operation modes enables several multiplexing readout protocols for superconducting qubits. We report the device's performance in a two-channel code domain multiplexing demonstration. The multiplexed data are recovered with fast readout times (up to 400 ns ) and infidelities ≤ 10-2 for probe powers ≥ 7 fW , in agreement with the expectation for binary signaling with Gaussian noise. 7. An overview of Ball Aerospace cryogen storage and delivery systems Marquardt, J.; Keller, J.; Mills, G.; Schmidt, J. 2015-12-01 Starting on the Gemini program in the 1960s, Beech Aircraft (now Ball Aerospace) has been designing and manufacturing dewars for a variety of cryogens including liquid hydrogen and oxygen. These dewars flew on the Apollo, Skylab and Space Shuttle spacecraft providing fuel cell reactants resulting in over 150 manned spaceflights. Since Space Shuttle, Ball has also built the liquid hydrogen fuel tanks for the Boeing Phantom Eye unmanned aerial vehicle. Returning back to its fuel cell days, Ball has designed, built and tested a volume-constrained liquid hydrogen and oxygen tank system for reactant delivery to fuel cells on unmanned undersea vehicles (UUVs). Herein past history of Ball technology is described. Testing has been completed on the UUV specific design, which will be described. 8. RHIC Polarized proton operation Huang, H.; Ahrens, L.; Alekseev, I.G.; Aschenauer, E.; Atoian, G.; Bai, M.; Bazilevsky, A.; Blaskiewicz, M.; Brennan, J.M.; Brown, K.A.; Bruno, D.; Connolly, R.; Dion, A.; D' Ottavio, T.; Drees, K.A.; Fischer, W.; Gardner, C.; Glenn, J.W.; Gu, X.; Harvey, M.; Hayes, T.; Hoff, L.; Hulsart, R.L.; Laster, J.; Liu, C.; Luo, Y.; MacKay, W.W.; Makdisi, Y.; Marr, G.J.; Marusic, A.; Meot, F.; Mernick, K.; Michnoff, R,; Minty, M.; Montag, C.; Morris, J.; Nemesure, S.; Poblaguev, A.; Ptitsyn, V.; Ranjibar, V.; Robert-Demolaize, G.; Roser, T.; J.; Severino, F.; Schmidke, B.; Schoefer, V.; Severino, F.; Smirnov, D.; Smith, K.; Steski, D.; Svirida, D.; Tepikian, S.; Trbojevic, D.; Tsoupas, N.; Tuozzolo, J. Wang, G.; Wilinski, M.; Yip, K.; Zaltsman, A.; Zelenski, A.; Zeno, K.; Zhang, S.Y. 2011-03-28 The Relativistic Heavy Ion Collider (RHIC) operation as the polarized proton collider presents unique challenges since both luminosity(L) and spin polarization(P) are important. With longitudinally polarized beams at the experiments, the figure of merit is LP{sup 4}. A lot of upgrades and modifications have been made since last polarized proton operation. A 9 MHz rf system is installed to improve longitudinal match at injection and to increase luminosity. The beam dump was upgraded to increase bunch intensity. A vertical survey of RHIC was performed before the run to get better magnet alignment. The orbit control is also improved this year. Additional efforts are put in to improve source polarization and AGS polarization transfer efficiency. To preserve polarization on the ramp, a new working point is chosen such that the vertical tune is near a third order resonance. The overview of the changes and the operation results are presented in this paper. Siberian snakes are essential tools to preserve polarization when accelerating polarized beams to higher energy. At the same time, the higher order resonances still can cause polarization loss. As seen in RHIC, the betatron tune has to be carefully set and maintained on the ramp and during the store to avoid polarization loss. In addition, the orbit control is also critical to preserve polarization. The higher polarization during this run comes from several improvements over last run. First we have a much better orbit on the ramp. The orbit feedback brings down the vertical rms orbit error to 0.1mm, much better than the 0.5mm last run. With correct BPM offset and vertical realignment, this rms orbit error is indeed small. Second, the jump quads in the AGS improved input polarization for RHIC. Third, the vertical tune was pushed further away from 7/10 snake resonance. The tune feedback maintained the tune at the desired value through the ramp. To calibrate the analyzing power of RHIC polarimeters at any energy above 9. Conceptual design of the Relativistic Heavy Ion Collider: RHIC Samios, Nicholas P. 1986-05-01 The complete Relativistic Heavy Ion Collider (RHIC) facility will be a complex set of accelerators and beam transfer equipment connecting them. A significant portion of the total facility either exists or is under construction. Two existing Tandem Van de Graaff accelerators will serve for the initial ion acceleration. Ions with a charge of -1 would be accelerated from ground to +15 MV potential, pass through a stripping foil, and accelerate back to ground potential, where they would pass through a second stripping foil. From there the ions will traverse a long transfer line to the AGS tunnel and be injected into the Booster accelerator. The Booster accelerates the ion bunch, and then the ions pass through one more stripper and then enter the Alternating Gradient Synchrotron (AGS), where they are accelerated to the top AGS energy and transferred to the collider. Bending and focusing of ion beams is to be achieved by superconducting magnets. The physics goals behind the RHIC are enumerated, particularly as regards the study of quark matter and the characteristics of high energy nucleus-nucleus collisions. The design of the collider and all its components is described, including the injector, the lattice, magnet system, cryogenic and vacuum systems, beam transfer, injection, and dump, rf system, and beam instrumentation and control system. Also given are cost estimates, construction schedules, and a management plan. (LEW) 10. SCHOTTKY MEASUREMENTS DURING RHIC 2000. CAMERON,P.; CUPOLO,J.; DEGEN,C.; HAMMONS,L.; KESSELMAN,M.; LEE,R.; MEYER,A.; SIKORA,R. 2001-06-18 The 2GHz Schottky system was a powerful diagnostic during RHIC 2000 commissioning. A continuous monitor without beam excitation, it provided betatron tune, chromaticity, momentum spread relative emittance, and synchrotron tune. It was particularly useful during transition studies. In addition, a BPM was resonated at 230MHz for Schottky measurements. 11. Absolute polarimetry at RHIC Okada, H; Bravar, A; Bunce, G; Dhawan, S; Eyser, K O; Gill, R; Haeberli, W; Huang, H; Jinnouchi, O; Makdisi, Y; Nakagawa, I; Nass, A; Saitô, N; Stephenson, E; Sviridia, D; Wise, T; Wood, J; Zelenski, A 2007-01-01 Precise and absolute beam polarization measurements are critical for the RHIC spin physics program. Because all experimental spin-dependent results are normalized by beam polarization, the normalization uncertainty contributes directly to final physics uncertainties. We aimed to perform the beam polarization measurement to an accuracy of\\Delta P_{beam}/P_{beam} < 5%$. The absolute polarimeter consists of Polarized Atomic Hydrogen Gas Jet Target and left-right pairs of silicon strip detectors and was installed in the RHIC-ring in 2004. This system features \\textit{proton-proton} elastic scattering in the Coulomb nuclear interference (CNI) region. Precise measurements of the analyzing power$A_N$of this process has allowed us to achieve$\\Delta P_{beam}/P_{beam} =4.2%$in 2005 for the first long spin-physics run. In this report, we describe the entire set up and performance of the system. The procedure of beam polarization measurement and analysis results from 2004-2005 are described. Physics topics of$A... 12. 900-L liquid xenon cryogenic system operation for the MEG experiment Haruyama, T; Mihara, S; Hisamatsu, Y; Iawamoto, W; Mori, T; Nishiguchi, H; Otani, W; Sawada, R; Uchiyama, Y; Nishitani, T 2009-01-01 A cryogenic system for the MEG (muon rare decay) experiment has started operation at the Paul Sherrer Institute in Zurich. The main part of the MEG detector is the 900-L liquid xenon calorimeter for gamma ray detection, equipped with 850 photo multipliers directly immersed in liquid xenon. A 200 W pulse tube cryocooler enabled LN2-free operation of this calorimeter. A liquid purification system; using a liquid pump and a zero boil-off 1000-L cryogenic buffer dewar is also included in the system. The first entire engineering run was carried out in November-December 2007 and satisfactory cryogenic performances were confirmed. 13. Cryogenics Vision Workshop for High-Temperature Superconducting Electric Power Systems Proceedings Energetics, Inc. 2000-01-01 The US Department of Energy's Superconductivity Program for Electric Systems sponsored the Cryogenics Vision Workshop, which was held on July 27, 1999 in Washington, D.C. This workshop was held in conjunction with the Program's Annual Peer Review meeting. Of the 175 people attending the peer review meeting, 31 were selected in advance to participate in the Cryogenics Vision Workshops discussions. The participants represented cryogenic equipment manufactures, industrial gas manufacturers and distributors, component suppliers, electric power equipment manufacturers (Superconductivity Partnership Initiative participants), electric utilities, federal agencies, national laboratories, and consulting firms. Critical factors were discussed that need to be considered in describing the successful future commercialization of cryogenic systems. Such systems will enable the widespread deployment of high-temperature superconducting (HTS) electric power equipment. Potential research, development, and demonstration (RD and D) activities and partnership opportunities for advancing suitable cryogenic systems were also discussed. The workshop agenda can be found in the following section of this report. Facilitated sessions were held to discuss the following specific focus topics: identifying Critical Factors that need to be included in a Cryogenics Vision for HTS Electric Power Systems (From the HTS equipment end-user perspective) identifying R and D Needs and Partnership Roles (From the cryogenic industry perspective) The findings of the facilitated Cryogenics Vision Workshop were then presented in a plenary session of the Annual Peer Review Meeting. Approximately 120 attendees participated in the afternoon plenary session. This large group heard summary reports from the workshop session leaders and then held a wrap-up session to discuss the findings, cross-cutting themes, and next steps. These summary reports are presented in this document. The ideas and suggestions 14. Progress and status of cryogenic refrigeration system for project Hydra Yuan, J.; Catseman, F.; Tilleman, H.; Henderson, N. 2015-12-01 In the last two decades, HTS cables have been successfully demonstrated around the world, preparing HTS power cables for a full commercial introduction. Among the demonstration projects, circulating subcooled liquid nitrogen to maintain the HTS cable at operating temperature is a widely adopted approach. In this approach, the cooling systems are absolutely critical to the successful operation of the HTS cables. This paper describes the progress and status of the cryogenic refrigeration system designed and manufactured for project Hydra, which is a project jointly funded by the U.S. Department of Homeland Security Science and Technology Directorate, American Superconductor and Consolidated Edison Company of New York, Inc. American Superconductor is leading the team supported by Con Edison, Ultera, Altran Solutions, and DH Industries. The cable is an inherently fault current limiting HTS cable, approximately 200 m long and designed to carry 96 MVA at a distribution level voltage of 13.8 kV. The cable will be installed and energized near New York City. The refrigeration system was designed and manufactured by DH Industries. This paper provides details on the successful factory acceptance testing completed in November 2014. 15. Miniaturized, Low Power Cryogenic Inlet System with Sampling Probes for Titan Project National Aeronautics and Space Administration — Thorleaf Research, Inc. has demonstrated feasibility in Phase 1 and now proposes a Phase 2 effort to develop a miniature, low power cryogenic inlet system with... 16. A Novel Flow Measurement System for Cryogenic Two-Phase Flow Project National Aeronautics and Space Administration — Flow rate measurements for cryogenic propellants are required for spacecraft and space exploration systems. Such a requirement has been hampered by lack of fast and... 17. Modeling Unsteady Cavitation Effects and Dynamic Loads in Cryogenic Systems Project National Aeronautics and Space Administration — There currently are no analytical or CFD tools that can reliably predict unsteady cavitation dynamics in liquid rocket cryogenic systems. Analysis of cavitating... 18. Installation and commissioning of a cryogen distribution system for the TPS project Tsai, H. H.; Hsiao, F. Z.; Li, H. C.; Lin, M. C.; Wang, C.; Liao, W. R.; Lin, T. F.; Chiou, W. S.; Chang, S. H.; Chuang, P. S. D. 2016-07-01 A cryogen distribution system was installed and commissioned to transfer liquid nitrogen (LN2) and liquid helium (LHe) from storage dewars to superconducting radio-frequency (SRF) cavities for the 3-GeV Taiwan Photon Source (TPS) project. The cryogen distribution system comprises one distribution valve box (DVB), four control valve boxes (CVB) and seven sections of multichannel transfer line (MCL). The DVB distributes the LHe and LN2 to the CVB, and then to the SRF cavities through independent vacuum-jacketed transfer lines. The vaporized GHe and GN2 from the cryomodules are collected via the MCL. The cryogen distribution system was installed and commissioned from October 2014 to the end of March 2015. This paper presents the installation, pre-commissioning and commissioning of the cryogen distribution system, and describes the heat load test. Thermal acoustic oscillation (TAO) was found in the GHe process line; this phenomenon and its solution are also presented and discussed. 19. Miniaturized, Low Power Cryogenic Inlet System with Sampling Probes for Titan Project National Aeronautics and Space Administration — Thorleaf Research, Inc. proposes to develop a miniature, low power cryogenic inlet system with sampling probes for Titan. This addresses a key technology gap for... 20. Main Consolidations and Improvements of the Control System and Instrumentation for the LHC Cryogenics Fluder, C; Bremer, J; Bremer, K; Ivens, B; Casas-Cubillos, J; Claudet, S; Gomes, P; Ivens, B; Perin, A; Pezzetti, M; Tovar-Gonzalez, A; Vauthier, N 2013-01-01 Operation of the LHC during 2010 and 2011 with 3.5 TeV beam energy and luminosity up to 3.65x1033 cm-2 s-1, led to radiation-induced failures of micro-electronic devices used in the cryogenic control system. Mitigating actions addressed equipment relocation and corrective patches on electronics and software. Driven by the technical requirements and by feedback from the cryogenic operation team, numerous consolidations and improvements were implemented on-the-fly, enhancing availability and operability of the LHC cryogenics. Furthermore, additional diagnostic tools, test benches, technical procedures and trainings have been provided to strengthen first line support services. 1. CERN experience and strategy for the maintenance of cryogenic plants and distribution systems Serio, L; Claudet, S; Delikaris, D; Ferlin, G; Pezzetti, M; Pirotte, O; Tavian, L; Wagner, U 2015-01-01 CERN operates and maintains the world largest cryogenic infrastructure ranging from ageing installations feeding detectors, test facilities and general services, to the state-of-the-art cryogenic system serving the flagship LHC machine complex. After several years of exploitation of a wide range of cryogenic installations and in particular following the last two years major shutdown to maintain and consolidate the LHC machine, we have analysed and reviewed the maintenance activities to implement an efficient and reliable exploitation of the installations. We report the results, statistics and lessons learned on the maintenance activities performed and in particular the required consolidations and major overhauling, the organization, management and methodologies implemented. 2. A Novel Pre-cooling System for a Cryogenic Pulsating Heat Pipe Xu, Dong; Liu, Huiming; Gong, Linghui; Xu, Xiangdong; Li, Laifeng To reduce the influence of the pipe material on the measurement of effective thermal conductivity, the pipe of a cryogenic pulsating heat pipe is generally made of stainless steel. Because of the low thermal conductivity of stainless steel, the pre-cooling of the evaporator in cryogenic pulsating heat pipe using helium as working fluid at 4.2 K is a problem. We designed a mechanical-thermal switch between the cryocooler and the evaporator, which was on during the pre-cooling process and off during the test process. By using the pre-cooling system, the cool down time of the cryogenic pulsating heat pipe was reduced significantly. 3. CERN experience and strategy for the maintenance of cryogenic plants and distribution systems Serio, L.; Bremer, J.; Claudet, S.; Delikaris, D.; Ferlin, G.; Pezzetti, M.; Pirotte, O.; Tavian, L.; Wagner, U. 2015-12-01 CERN operates and maintains the world largest cryogenic infrastructure ranging from ageing installations feeding detectors, test facilities and general services, to the state-of-the-art cryogenic system serving the flagship LHC machine complex. After several years of exploitation of a wide range of cryogenic installations and in particular following the last two years major shutdown to maintain and consolidate the LHC machine, we have analysed and reviewed the maintenance activities to implement an efficient and reliable exploitation of the installations. We report the results, statistics and lessons learned on the maintenance activities performed and in particular the required consolidations and major overhauling, the organization, management and methodologies implemented. 4. Performance of a proximity cryogenic system for the ATLAS central solenoid magnet Doi, Y; Makida, Y; Kondo, Y; Kawai, M; Aoki, K; Haruyama, T; Kondo, T; Mizumaki, S; Wachi, Y; Mine, S; Haug, F; Delruelle, N; Passardi, Giorgio; ten Kate, H H J 2002-01-01 The ATLAS central solenoid magnet has been designed and constructed as a collaborative work between KEK and CERN for the ATLAS experiment in the LHC project The solenoid provides an axial magnetic field of 2 Tesla at the center of the tracking volume of the ATLAS detector. The solenoid is installed in a common cryostat of a liquid-argon calorimeter in order to minimize the mass of the cryostat wall. The coil is cooled indirectly by using two-phase helium flow in a pair of serpentine cooling line. The cryogen is supplied by the ATLAS cryogenic plant, which also supplies helium to the Toroid magnet systems. The proximity cryogenic system for the solenoid has two major components: a control dewar and a valve unit In addition, a programmable logic controller, PLC, was prepared for the automatic operation and solenoid test in Japan. This paper describes the design of the proximity cryogenic system and results of the performance test. (7 refs). 5. Cryogenic mechanical property testing system directly cooled by G-M cryocooler Huang, R. J.; Liu, Q.; Li, L. F.; Gong, L. H.; Liu, H. M.; Xu, D. 2014-01-01 Cryogenic mechanical properties are generally considered to be some of the most important parameters in cryogenic engineering. Therefore, it is very important to test and investigate mechanical properties at low temperatures. Most systems for cryogenic mechanical property testing are cooled using liquid nitrogen (300 K-77 K) or liquid helium (77 K-4.2 K). As we know, liquid helium is relatively rare and thus expensive. In this study, to attain accurate and stable intermediate temperatures and reduce testing cost, a cryogenic mechanical property testing system cooled by a G-M cryocooler was studied and developed. In this system, the sample can be cooled down to 10.5 K after about 10 hours of running. The tension, bending and compression testing (load range up to 50 kN) can be carried out. 6. Cryogenic characterization of the Planck sorption cooler system flight model Morgante, G; Melot, F; Stassi, P; Terenzi, L; Wilson, P; Hernandez, B; Wade, L; Gregorio, A; Bersanelli, M; Butler, C; Mandolesi, N; 10.1088/1748-0221/4/12/T12016 2009-01-01 This paper is part of the Prelaunch status LFI papers published on JINST: http://www.iop.org/EJ/journal/-page=extra.proc5/1748-0221 Two continuous closed-cycle hydrogen Joule-Thomson (J-T) sorption coolers have been fabricated and assembled by the Jet Propulsion Laboratory (JPL) for the European Space Agency (ESA) Planck mission. Each refrigerator has been designed to provide a total of ~ 1W of cooling power at two instrument interfaces: they directly cool the Planck Low Frequency Instrument (LFI) around 20K while providing a pre-cooling stage for a 4 K J-T mechanical refrigerator for the High Frequency Instrument (HFI). After sub-system level validation at JPL, the cryocoolers have been delivered to ESA in 2005. In this paper we present the results of the cryogenic qualification and test campaigns of the Nominal Unit on the flight model spacecraft performed at the CSL (Centre Spatial de Liege) facilities in 2008. Test results in terms of input power, cooling power, temperature, and temperature fluctuations o... 7. Commissioning of the Cryogenic System for the ATLAS Superconducting Magnets Delruelle, N; Bradshaw, T; Haug, F; ten Kate, H H J; Passardi, Giorgio; Pengo, R; Pezzetti, M; Pirotte, O; Rochford, J 2006-01-01 The paper describes the test results of the helium cryoplant for the superconducting magnets of the ATLAS particle detector at CERN. It consists of two refrigerators used in common by all the magnets and of two proximity cryogenic systems (PCS) interfacing respectively the toroids and the central solenoid. Emphasis is given to the commissioning of the refrigerators: the main unit of 6 kW equivalent capacity at 4.5 K and the thermal shield refrigerator providing 20 kW between 40 K and 80 K. The first unit is used for refrigeration at 4.5 K and for the cooling of three sets of 20 kA current leads, while the second one provides, in addition to the 20 kW refrigeration of the thermal shields, 60 kW for the cool-down to 100 K of the 660 ton cold mass of the magnets. The tests, carried out with the equipment in the final underground configuration, are extended to the PCS that includes the large liquid helium centrifugal pumps (each providing 1.2 kg/s) for forced-flow cooling of the magnets and the complex distributi... 8. Cryogenic characterization of the Planck sorption cooler system flight model Morgante, G; Terenzi, L; Butler, C; Mandolesi, N [INAF - IASF Bologna, via P. Gobetti 101, 40129 Bologna (Italy); Pearson, D; Wilson, P; Hernandez, B; Wade, L [Jet Propulsion Laboratory, 4800 Oak Grove Drive, Pasadena California 91109 (United States); Melot, F; Stassi, P [Laboratoire de Physique Subatomique et de Cosmologie 53 Avenue des Martyrs, 38026 Grenoble Cedex (France); Gregorio, A [Dipartimento di Fisica, Universita degli Studi di Trieste, via Valerio 2 - I-34127 Trieste (Italy); Bersanelli, M, E-mail: morgante@iasfbo.inaf.i [Dipartimento di Fisica, Universita degli Studi di Milano, via Celoria 16, - I20133 Milano (Italy) 2009-12-15 Two continuous closed-cycle hydrogen Joule-Thomson (J-T) sorption coolers have been fabricated and assembled by the Jet Propulsion Laboratory (JPL) for the European Space Agency (ESA) Planck mission. Each refrigerator has been designed to provide a total of {approx} 1W of cooling power at two instrument interfaces: they directly cool the Planck Low Frequency Instrument (LFI) around 20K while providing a pre-cooling stage for a 4 K J-T mechanical refrigerator for the High Frequency Instrument (HFI). After sub-system level validation at JPL, the cryocoolers have been delivered to ESA in 2005. In this paper we present the results of the cryogenic qualification and test campaigns of the Nominal Unit on the flight model spacecraft performed at the CSL (Centre Spatial de Liege) facilities in 2008. Test results in terms of input power, cooling power, temperature, and temperature fluctuations over the flight allowable ranges for these interfaces are reported and analyzed with respect to mission requirements. 9. Longitudinal impedance of RHIC Blaskiewicz, M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Brennan, J. M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Mernick, K. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-05-03 The longitudinal impedance of the two RHIC rings has been measured using the effect of potential well distortion on longitudinal Schottky measurements. For the blue RHIC ring Im(Z/n) = 1.5±0.2Ω. For the yellow ring Im(Z/n) = 5.4±1Ω. 10. Stochastic cooling in RHIC Brennan,J.M.; Blaskiewicz, M. M.; Severino, F. 2009-05-04 After the success of longitudinal stochastic cooling of bunched heavy ion beam in RHIC, transverse stochastic cooling in the vertical plane of Yellow ring was installed and is being commissioned with proton beam. This report presents the status of the effort and gives an estimate, based on simulation, of the RHIC luminosity with stochastic cooling in all planes. 11. Real-Time Model-Based Leak-Through Detection within Cryogenic Flow Systems Walker, M.; Figueroa, F. 2015-01-01 The timely detection of leaks within cryogenic fuel replenishment systems is of significant importance to operators on account of the safety and economic impacts associated with material loss and operational inefficiencies. Associated loss in control of pressure also effects the stability and ability to control the phase of cryogenic fluids during replenishment operations. Current research dedicated to providing Prognostics and Health Management (PHM) coverage of such cryogenic replenishment systems has focused on the detection of leaks to atmosphere involving relatively simple model-based diagnostic approaches that, while effective, are unable to isolate the fault to specific piping system components. The authors have extended this research to focus on the detection of leaks through closed valves that are intended to isolate sections of the piping system from the flow and pressurization of cryogenic fluids. The described approach employs model-based detection of leak-through conditions based on correlations of pressure changes across isolation valves and attempts to isolate the faults to specific valves. Implementation of this capability is enabled by knowledge and information embedded in the domain model of the system. The approach has been used effectively to detect such leak-through faults during cryogenic operational testing at the Cryogenic Testbed at NASA's Kennedy Space Center. 12. Conceptual design report for the University of Rochester cryogenic target delivery system Fagaly, R.L.; Alexander, N.B.; Bourque, R.F.; Dahms, C.F.; Lindgren, J.R.; Miller, W.J. (General Atomics, San Diego, CA (United States)); Bittner, D.N.; Hendricks, C.D. (W.J. Schafer Associates, Livermore, CA (United States)) 1993-05-01 The upgrade of the Omega laser at the University of Rochester's Laboratory for Laser Energetics (UR/LLE) will result in a need for large targets filled with D[sub 2] or Dt and maintained at cryogenic temperatures. This mandates a cryogenic target delivery system capable of filling, layering, characterizing and delivering cryogenic targets to the Omega Upgrade target chamber. The program goal is to design, construct, and test the entire target delivery system by June 1996. When completed (including an operational demonstration), the system will be shipped to Rochester for reassembly and commissioning in time for the Omega Upgrade cryogenic campaign, scheduled to start in 1998. General Atomics has been assigned the task of developing the conceptual design for the cryogenic target delivery system. Design and fabrication activities will be closely coordinated with the University of Rochester, Lawrence Livermore National laboratory (LLNL) and Los Alamos National Laboratory (LANL), drawing upon their knowledge base in fuel layering and cryogenic characterization. The development of a target delivery system for Omega could also benefit experiments at Lawrence Livermore National Laboratory and the other ICF Laboratories in that the same technologies could be applied to NOVA, the National Ignition Facility or the future Laboratory Microfusion Facility. 13. Conceptual design report for the University of Rochester cryogenic target delivery system Fagaly, R.L.; Alexander, N.B.; Bourque, R.F.; Dahms, C.F.; Lindgren, J.R.; Miller, W.J. [General Atomics, San Diego, CA (United States); Bittner, D.N.; Hendricks, C.D. [W.J. Schafer Associates, Livermore, CA (US) 1993-05-01 The upgrade of the Omega laser at the University of Rochesters Laboratory for Laser Energetics (UR/LLE) will result in a need for large targets filled with D{sub 2} or Dt and maintained at cryogenic temperatures. This mandates a cryogenic target delivery system capable of filling, layering, characterizing and delivering cryogenic targets to the Omega Upgrade target chamber. The program goal is to design, construct, and test the entire target delivery system by June 1996. When completed (including an operational demonstration), the system will be shipped to Rochester for reassembly and commissioning in time for the Omega Upgrade cryogenic campaign, scheduled to start in 1998. General Atomics has been assigned the task of developing the conceptual design for the cryogenic target delivery system. Design and fabrication activities will be closely coordinated with the University of Rochester, Lawrence Livermore National laboratory (LLNL) and Los Alamos National Laboratory (LANL), drawing upon their knowledge base in fuel layering and cryogenic characterization. The development of a target delivery system for Omega could also benefit experiments at Lawrence Livermore National Laboratory and the other ICF Laboratories in that the same technologies could be applied to NOVA, the National Ignition Facility or the future Laboratory Microfusion Facility. 14. Summary of the RHIC Retreat 2007 Pilat,F.; Gardner, C.; Montag, C.; Roser, T. 2008-08-01 The RHIC Retreat 2007 took place on July 16-17 2007 at the Foxwoods Resort in CT, about 3 weeks after the end of the RHIC Run-7. The goal of the Retreat is traditionally to plan the upcoming run in the light of the results from the previous one, by providing a snapshot of the present understanding of the machine and a forum for free and frank discussion. A particular attention was paid to the challenge of increasing the time at store, and the related issue of system reliability. An interesting Session covered all new developments aimed to improve the machine performance and luminosity. In Section 2 we summarize the results from Run-7 for RHIC and the injectors and discuss the present objectives of the RHIC program and performance. Sections 3-6 are summaries of the Retreat sessions focused on preparation for deuteron gold and polarized protons, respectively, machine availability and new developments. 15. Consequences of high-$x$ proton size fluctuations in small collision systems at RHIC McGlinchey, D; Perepelitsa, D V 2016-01-01 Recent measurements of jet production rates at large transverse momentum ($p_T$) in the collisions of small projectiles with large nuclei at RHIC and the LHC indicate that they have an unexpected relationship with estimates of the collision centrality. One compelling interpretation of the data is that it captures an $x_p$-dependent decrease in the average interaction strength of the nucleon in the projectile undergoing a hard scattering. A weakly interacting or "shrinking" nucleon in the projectile strikes fewer nucleons in the nucleus, resulting in a particular pattern of centrality-dependent modifications to high-$p_T$ processes. We describe a simple one-parameter geometric implementation of this picture within a modified Monte Carlo Glauber model tuned to $d$$+$Au jet data, and explore two of its major consequences. First, the model predicts a particular projectile-species dependence to the centrality dependence at high-$x_p$, opposite to that expected from an energy loss effect. Second, we find that some ... 16. Construction and commissioning of a hydrogen cryogenic distillation system for tritium recovery at ICIT Rm. Valcea Ana, George, E-mail: george.ana@icsi.ro [Institute for Cryogenic and Isotopic Technologies, Rm. Valcea (Romania); Cristescu, Ion [Karlsruhe Istitute for Technologies, Tritium Laboratory, Eggenstein-Leopoldshaffen (Germany); Draghia, Mirela [ISTECH, Timisoara (Romania); Bucur, Ciprian; Balteanu, Ovidiu; Vijulie, Mihai; Popescu, Gheorghe; Costeanu, Claudiu; Sofilca, Nicolae; Stefan, Iulia; Daramus, Robert; Niculescu, Alina; Oubraham, Anisoara; Spiridon, Ionut; Vasut, Felicia; Moraru, Carmen; Brad, Sebastian [Institute for Cryogenic and Isotopic Technologies, Rm. Valcea (Romania); Pasca, Gheorghe [ISTECH, Timisoara (Romania) 2016-05-15 Highlights: • Cryogenic distillation (CD) process is being employed for tritium separation from tritiated hydrogen mixtures. • Process control and safety phylosophy with the detritiation plant from Rm. Vâlcea. • Tests undertaken prior to commissioning of the CD system from Rm. Vâlcea. • Preliminary experiments with the CD system (non-radiological). - Abstract: Cryogenic distillation (CD) of hydrogen in combination with Liquid Phase Catalytic Exchange (LPCE) or Combined Electrolytic Catalytic Exchange (CECE) process is used for tritium removal/recovery from tritiated water. Tritiated water is being obtained after long time operation of CANDU reactors, or in case of ITER mainly by the Detritiation System (DS). The cryogenic distillation system (CDS) used to remove/recover tritium from a hydrogen stream consists of a cascade of cryogenic distillation columns and a refrigeration unit which provides the cooling capacity for the condensers of CD columns. The columns, together with the condensers and the process heat-exchangers are accommodated in a vacuumed cold box. In the particularly case of the ICIT Plant, the cryogenic distillation cascade consists of four columns with diameters between 100–7 mm and it has been designed to process up to 10 mc/h of tritiated deuterium. This paper will present the steps undertaken for construction and commissioning of a pilot plant for tritium removal/recovery by cryogenic distillation of hydrogen. The paper will show besides preliminary data obtained during commissioning, also general characteristics of the plant and its equipments. 17. CRYOGENIC SYSTEM FOR PRECISE CALIBRATION OF TEMPERATURE SENSORS A. N. Solovyev 2016-09-01 Full Text Available A calibration technique for cryogenic temperature sensors is proposed and implemented. The experimental setup is based on the helium cryogenerator, providing calibration of the temperature sensors of various types in wide temperature range, including cryogenic band (25-100K. A condensation thermometer with hydrogen, neon, argon and xenon as working gases is used as a reference sensor. The experimental setup was successfully used for precise (0.1K precision calibration of platinum resistive temperature detectors (Pt-100 for international nuclear physics experiments MuSun and PolFusion. The setup can also be used for calibration of temperature sensors of the other types. 18. Optical density measurements in a multiphase cryogenic fluid flow system Korman, Valentin; Wiley, John; Gregory, Don A. 2006-05-01 An accurate determination of fluid flow in a cryogenic propulsion environment is difficult under the best of circumstances. The extreme thermal environment increases the mechanical constraints, and variable density conditions create havoc with traditional flow measurement schemes. Presented here are secondary results of cryogenic testing of an all-optical sensor capable of a mass flow measurement by directly interrogating the fluid's density state and a determination of the fluid's velocity. The sensor's measurement basis does not rely on any inherent assumptions as to the state of the fluid flow (density or otherwise). The fluid sensing interaction model will be discussed. Current test and evaluation data and future development work will be presented. 19. Zero Gravity Cryogenic Vent System Concepts for Upper Stages Ravex, Alain; Flachbart, Robin; Holt, Barney The capability to vent in zero gravity without resettling is a technology need that involves practically all uses of sub-critical cryogenics in space. Venting without resettling would extend cryogenic orbital transfer vehicle capabilities. However, the lack of definition regarding liquid/ullage orientation coupled with the somewhat random nature of the thermal stratification and resulting pressure rise rates, lead to significant technical challenges. Typically a zero gravity vent concept, termed a thermodynamic vent system (TVS), consists of a tank mixer to destratify the propellant, combined with a Joule-Thomson (J-T) valve to extract thermal energy from the propellant. Marshall Space Flight Center's (MSFC's) Multipurpose Hydrogen Test Bed (MHTB) was used to test both spray bar and axial jet TVS concepts. The axial jet system consists of a recirculation pump heat exchanger unit. The spray bar system consists of a recirculation pump, a parallel flow concentric tube, heat exchanger, and a spray bar positioned close to the longitudinal axis of the tank. The operation of both concepts is similar. In the mixing mode, the recirculation pump withdraws liquid from the tank and sprays it into the tank liquid, ullage, and exposed tank surfaces. When energy extraction is required, a small portion of the recirculated liquid is passed sequentially through the J-T expansion valve, the heat exchanger, and is vented overboard. The vented vapor cools the circulated bulk fluid, thereby removing thermal energy and reducing tank pressure. The pump operates alone, cycling on and off, to destratify the tank liquid and ullage until the liquid vapor pressure reaches the lower set point. At that point, the J-T valve begins to cycle on and off with the pump. Thus, for short duration missions, only the mixer may operate, thus minimizing or even eliminating boil-off losses. TVS performance testing demonstrated that the spray bar was effective in providing tank pressure control within a 6 20. Concept of a Cryogenic System for a Cryogen-Free 25 T Superconducting Magnet Iwai, Sadanori; Takahashi, Masahiko; Miyazaki, Hiroshi; Tosaka, Taizo; Tasaki, Kenji; Hanai, Satoshi; Ioka, Shigeru; Watanabe, Kazuo; Awaji, Satoshi; Oguro, Hidetoshi A cryogen-free 25 T superconducting magnet using a ReBCO insert coil that generates 11.5 T in a 14 T background field of outer low-temperature superconducting (LTS) coils is currently under development. The AC loss of the insert coil during field ramping is approximately 8.8 W, which is difficult to dissipate at the operating temperature of the LTS coils (4 K). However, since a ReBCO coil can operate at a temperature above 4 K, the ReBCO insert coil is cooled to about 10 K by two GM cryocoolers, and the LTS coils are independently cooled by two GM/JT cryocoolers. Two GM cryocoolers cool a circulating helium gas through heat exchangers, and the gas is transported over a long distance to the cold stage located on the ReBCO insert coil, in order to protect the cryocoolers from the leakage field of high magnetic fields. The temperature difference of the 2nd cold stage of the GM cryocoolers and the insert coil can be reduced by increasing the gas flow rate. However, at the same time, the heat loss of the heat exchangers increases, and the temperature of the second cold stage is raised. Therefore, the gas flow rate is optimized to minimize the operating temperature of the ReBCO insert coil by using a flow controller and a bypass circuit connected to a buffer tank. 1. The Development of the Control System for the Cryogenics in the LHC Tunnel Fluder, C; Casas-Cubillos, J; Dubert, P; Gomes, P; Pezzetti, M; Tovar-Gonzalez, A; Zwalinski, L 2011-01-01 The Large Hadron Collider (LHC) was commissioned at CERN and started operation with beams in 2008. The LHC makes extensive use of superconductors, in magnets, electrical feed boxes and accelerating cavities, which are operated at cryogenic temperatures. The process automation for the cryogenic distribution around the 27 km accelerator circumference is based on 18 Programmable Logic Controllers (PLCs); overall, they handle 4 000 control loops and 8 000 alarms and interlocks; 16 000 cryogenic sensors and actuators are accessed through industrial field networks. This paper reviews the control system architecture and the main hardware and software components; presents the hardware commissioning and software production methodologies; and illustrates some of the problems faced during development, commissioning and nominal cryogenics operation, together with the solutions applied. 2. Structural Design and Analysis of a 150 kJ HTS SMES Cryogenic System Han, Peng; Wu, Yu; Liu, Huajun; Li, Laifeng; Yang, Huihui A 150 kJ high temperature superconducting magnetic energy storage (HTS-SMES) system is under manufacturing in China. This paper focuses on the structural design and analysis of the SMES cryogenic system. The cryogenic system is designed and fabricated to maintain the working temperature. The system includes a vacuum vessel, its thermal radiation shield, its supporting devices, conduction plates, and current leads. Two G-M cryocoolers are used for the system cooling, the main one is connected to the HTS coils and the other is connected to the thermal shield and the lower ends of the current leads. In this study, the 3D models of the SMES cryogenic system were created with CATIA, a 3D model design software, and the analysis of the SMES cryogenic system was done by ANSYS. The mechanical analysis results on the vacuum vessel, suspension devices and supporting devices are presented, particularly the analyses on suspenders and shelf supports are of vital importance since the finished SMES system should meet vehicle-mounted requirements in long time transport. The heat load and the temperature distribution of the thermal shield were analyzed. A cooling experiment of the cryogenic system was made and the thermal shield was cooled down to about 50 K. 3. Cryogenic and non-cryogenic pool calcites indicating permafrost and non-permafrost periods: a case study from the Herbstlabyrinth-Advent Cave system (Germany D. K. Richter 2010-11-01 Full Text Available Weichselian cryogenic calcites collected in what is referred to as the Rätselhalle of the Herbstlabyrinth-Advent Cave system are structurally classified as rhombohedral crystals and spherulitic aggregates. The carbon and oxygen isotopic composition of these precipitates (δ13C = +0.6 to −7.3‰ δ18O = −6.9 to −18.0‰ corresponds to those of known slowly precipitated cryogenic cave calcites under conditions of isotopic equilibrium between water and ice of Central European caves. The carbon and oxygen isotopic composition varies between different caves which is attributed to the effects of cave air ventilation before the freezing started. By petrographic and geochemical comparisons of Weichselian cryogenic calcite with recent to sub-recent precipitates as well as Weichselian non-cryogenic calcites of the same locality, a model for the precipitation of these calcites is proposed. While the recent and sub-recent pool-calcites isotopically match the composition of interglacial speleothems (stalagmites, etc., isotope ratios of Weichselian non-cryogenic pool-calcites reflect cooler conditions. Weichselian cryogenic calcites show a trend towards low δ18O values with higher carbon isotope ratios reflecting slow freezing of the precipitating solution. In essence, the isotope geochemistry of the Weichselian calcites reflects the climate history changing from overall initial permafrost conditions to permafrost-free and subsequently to renewed permafrost conditions. Judging from the data compiled here, the last permafrost stage in the Rätselhalle is followed by a warm period (interstadial and/or Holocene. During this warmer period, the cave ice melted and cryogenic and non-cryogenic Weichselian calcite precipitates were deposited on the cave ground or on fallen blocks, respectively. 4. Design and development of a device management platform for EAST cryogenic system Zhou, Zhiwei, E-mail: zzw@ipp.ac.cn; Lu, Xiaofei, E-mail: xiaofeilu@ipp.ac.cn; Zhuang, Ming, E-mail: zhm@ipp.ac.cn; Hu, Liangbing, E-mail: huliangbing@ipp.ac.cn; Xia, Genhai, E-mail: xgh@ipp.ac.cn 2014-05-15 Highlights: • A device management platform for EAST cryogenic system based on DCS is designed. • This platform enhances the integrity and continuity of system device information. • It can help predictive maintenance and device management decision. - Abstract: EAST cryogenic system is one of the critical sub-systems of the EAST tokamak device. It is a large scale helium cryoplant, which adopts distributed control system to realize monitoring and control of the cryogenic process and devices. However, the maintenance and management of most field devices are still in the corrective maintenance or traditional preventive maintenance stage. Under maintained or over maintained problems widely exist, which could cause devices fault and increase operation costs. Therefore, a device management platform is proposed for a safe and steady operation as well as fault diagnosis and predictive maintenance of EAST cryogenic system. This paper presents the function design and architecture design of the cryogenic device management platform. This platform is developed based on DeltaV DCS and acquires monitoring data through OPC protocol. It consists of three pillars, namely device information management, device condition management, and device performance monitoring. The development and implementation of every pillar are illustrated in detail in this paper. Test results and discussions are presented in the end. 5. First Assessment of Reliability Data for the LHC Accelerator and Detector Cryogenic System Components Perinic, G; Alonso-Canella, I; Balle, C; Barth, K; Bel, J F; Benda, V; Bremer, J; Brodzinski, K; Casas-Cubillos, J; Cuccuru, G; Cugnet, M; Delikaris, D; Delruelle, N; Dufay-Chanat, L; Fabre, C; Ferlin, G; Fluder, C; Gavard, E; Girardot, R; Haug, F; Herblin, L; Junker, S; Klabi , T; Knoops, S; Lamboy, J P; Legrand, D; Metselaar, J; Park, A; Perin, A; Pezzetti, M; Penacoba-Fernandez, G; Pirotte, O; Rogez, E; Suraci, A; Stewart, L; Tavian, L J; Tovar-Gonzalez, A; Van Weelderen, R; Vauthier, N; Vullierme, B; Wagner, U 2012-01-01 The Large Hadron Collider (LHC) cryogenic system comprises eight independent refrigeration and distribution systems that supply the eight 3.3 km long accelerator sectors with cryogenic refrigeration power as well as four refrigeration systems for the needs of the detectors ATLAS and CMS. In order to ensure the highest possible reliability of the installations, it is important to apply a reliability centred approach for the maintenance. Even though large scale cryogenic refrigeration exists since the mid 20th century, very little third party reliability data is available today. CERN has started to collect data with its computer aided maintenance management system (CAMMS) in 2009, when the accelerator has gone into normal operation. This paper presents the reliability observations from the operation and the maintenance side, as well as statistical data collected by the means of the CAMMS system. 6. Cryogenic helium gas circulation system for advanced characterization of superconducting cables and other devices Pamidi, Sastry; Kim, Chul Han; Kim, Jae-Ho; Crook, Danny; Dale, Steinar 2012-04-01 A versatile cryogenic test bed, based on circulating cryogenic helium gas, has been designed, fabricated, and installed at the Florida State University Center for Advanced Power Systems (FSU-CAPS). The test bed is being used to understand the benefits of integrating the cryogenic systems of multiple superconducting power devices. The helium circulation system operates with four sets of cryocooler and heat exchanger combinations. The maximum operating pressure of the system is 2.1 MPa. The efficacy of helium circulation systems in cooling superconducting power devices is evaluated using a 30-m-long simulated superconducting cable in a flexible cryostat. Experiments were conducted at various mass flow rates and a variety of heat load profiles. A 1-D thermal model was developed to understand the effect of the gas flow parameters on the thermal gradients along the cable. Experimental results are in close agreement with the results from the thermal model. 7. Development of a cryogenic target delivery system for HiPER Perin, J.P. [CEA, DRFMC, Serv Basses Temp, F-38054 Grenoble, (France); Perlado, M. [Univ Politecn Madrid, Inst Fus Nucl DENIM, ETSII, Madrid, (Spain); Tolley, M. [Rutherford Appleton Lab, Cent Laser Facil, Rutherford, NJ (United States) 2009-07-01 For the future, we have to develop new sources of energy. These new sources may be based on nuclear fusion with magnetic confinement (as with the ITER experiment) or with a new concept based on inertial confinement. The European community plans to build a facility (HiPER project) which is dedicated to reaching high gain with cryogenic targets, and to test the concepts of target mass production and rep rate shots. The cryogenic system for the first phase experiments in HiPER is based on the cryogenic system developed for the French facility Laser MegaJoule (LMJ). The latter must be modified and upgraded for direct drive targets. In particular the target must be protected from the radiation flux from the vacuum vessel by a thermal shroud In addition, the LMJ system must be equipped with a thermal system to allow layering of the fusion fuel to take place. (authors) 8. RHIC 12x150A current lead temperature controller: design and implementation Mi, C.; Seberg, S.; Ganetis, Hamdi, K.; Louie, W.; Heppner, G.; Jamilkowski, J.; Bruno, D.; DiLieto, A.; Sirio, C.; Tuozzolo, J.; Sandberg, J.; Unger, K. 2011-03-28 There are 60 12 x 150A current leads distributed in six RHIC service buildings; each lead delivers power supply current from room temperature to cryogenic temperature in RHIC. Due to the humid environment, condensation occurs frequently and ice forms quickly during operation, especially during an extensive storage period. These conditions generate warnings and alarms to which personnel must respond and establish temporary solutions to keep the machine operating. In here, we designed a temperature control system to avoid such situations. This paper discusses its design, implementation, and some results. There are six service buildings in the RHIC complex; each building has two valve boxes that transfer room-temperature current cables from the power supplies into superconducting leads, and then transport them into the RHIC tunnel. In there, the transition between the room-temperature lead into superconducting lead is critical and essential; smooth running during the physics store is crucial for the machine's continuing operation. One of the problems that often occurred previously was the icing of these current leads that could result in a potential leakage current onto ground, thereby preventing a continuous supply of physics store. Fig. 1 illustrates a typical example on a power lead. Among the modifications of the design of the valve box, we list below the new requirements for designing the temperature controller to prevent icing occurring: (1) Remotely control, monitor, and record each current lead's temperature in real time. Prevent icing or overheating of a power lead. (2) Include a temperature alarm for the high/low level threshold. In this paper we discuss the design, implementation, upgrades to, and operation of this new system. 9. ELECTRON COOLING FOR RHIC. BEN-ZVI,I. 2001-05-13 The Accelerator Collider Department (CAD) at Brookhaven National Laboratory is operating the Relativistic Heavy Ion Collider (RHIC), which includes the dual-ring, 3.834 km circumference superconducting collider and the venerable AGS as the last part of the RHIC injection chain. CAD is planning on a luminosity upgrade of the machine under the designation RHIC II. One important component of the RHIC II upgrade is electron cooling of RHIC gold ion beams. For this purpose, BNL and the Budker Institute of Nuclear Physics in Novosibirsk entered into a collaboration aimed initially at the development of the electron cooling conceptual design, resolution of technical issues, and finally extend the collaboration towards the construction and commissioning of the cooler. Many of the results presented in this paper are derived from the Electron Cooling for RHIC Design Report [1], produced by the, BINP team within the framework of this collaboration. BNL is also collaborating with Fermi National Laboratory, Thomas Jefferson National Accelerator Facility and the University of Indiana on various aspects of electron cooling. 10. Low cryogen inventory, forced flow Ne cooling system with room temperature compression stage and heat recuperation Shornikov, A; Wolf, A 2014-01-01 We present design and commissioning results of a forced flow cooling system utilizing neon at 30 K. The cryogen is pumped through the system by a room-temperature compression stage. To decouple the cold zone from the compression stage a recuperating counterflow tube-in-tube heat exchanger is used. Commissioning demonstrated successful condensation of neon and transfer of up to 30 W cooling power to the load at 30 K using only 30 g of the cryogen circulating in the system at pressures below 170 kPa. 11. Economics of Large Helium Cryogenic Systems experience from Recent Projects at CERN Claudet, S; Lebrun, P; Tavian, L; Wagner, U 1999-01-01 Large projects based on applied superconductivity, such as particle accelerators, tokamaks or SMES, require powerful and complex helium cryogenic systems, the cost of which represents a significant, if not dominant fraction of the total capital and operational expenditure. It is therefore important to establish guidelines and scaling laws for costing such systems, based on synthetic estimators of their size and performance. Although such data has already been published for many years, the experience recently gathered at CERN with the LEP and LHC projects, which have de facto turned the laboratory into a major world cryogenic center, can be exploited to update this information and broaden the range of application of the scaling laws. We report on the economics of 4.5 K and 1.8 K refrigeration, cryogen distribution and storage systems, and indicate paths towards their cost-to-performance optimisation. 12. New cryogenic system of the next-generation infrared astronomy mission SPICA Ogawa, H.; Nakagawa, T.; Matsuhara, H.; Shinozaki, K.; Goto, K.; Isobe, N.; Kawada, M.; Mizutani, T.; Sato, Y.; Sugita, H.; Takeuchi, S.; Yamawaki, T.; Shibai, H. 2016-07-01 We present the new design of the cryogenic system of the next-generation infrared astronomy mission SPICA under the new framework. The new design employs the V-groove design for radiators, making the best use of the Planck heritage. The new design is based on the ESA-JAXA CDF study (NG-CryoIRTel, CDF-152(A)) with a 2 m telescope, and we modified the CDF design to accommodate the 2.5 m telescope to meet the science requirements of SPICA. The basic design concept of the SPICA cryogenic system is to cool the Science Instrument Assembly (SIA, which is the combination of the telescope and focal-plane instruments) below 8K by the combination of the radiative cooling system and mechanical cryocoolers without any cryogen. 13. Preliminary Design of Control Network for HT-7U Tokamak Cryogenic System Jin Yibin(金毅彬); Zhuang Ming(庄明); Bai Hongyu(白宏宇) 2003-01-01 In the course of the cryoplant modernization, a control network will be set up in order to facilitate the control, the supervision, the centralized data acquisition and the alarm handling of the cryogenic system for HT-7U tokamak. The paper introduces the preliminary design of control network based on the Controller Link Network for HT-7U tokamak cryogenic system. The multi-layer structure mentioned in the paper is the mainstream of automatic control.The control philosophy, the structure of the network and the components for control are also presented. 14. The Integration Of The LHC Cryogenics Control System Data Into The CERN Layout Database Fortescue-Beck, E; Gomes, P 2011-01-01 The Large Hadron Collider’s Cryogenic Control System makes extensive use of several databases to manage data appertaining to over 34,000 cryogenic instrumentation channels. This data is essential for populating the software of the PLCs which are responsible for maintaining the LHC at the appropriate temperature. In order to reduce the number of data sources and the overall complexity of the system, the databases have been rationalised and the automatic tool, that extracts data for the control software, has been simplified. This paper describes the main improvements that have been made and considers the success of the project. 15. Cryogenics a textbook Thipse, S S 2013-01-01 A Textbook covers lucidly various cryogenic applications including cryogenic engines and space and electronic applications. Importance of cryogenic engines in space propulsion, complete thermodynamic analysis of cryogenic systems with special emphasis on cryogenic cycles, Dewar vessels used to store cryogenic fluids and their applications in various industries have also been discussed in detail. Explanation of Superconductivity and its applications with a description of various Cryocoolers used in industry has also been provided with extensive details. Further technical information on cryogens has been specified alongwith the vacuum technology which has been sufficiently described with examples. Science of Cryonics has been elaborated and all aspects of technology related to functioning of cryogenic plants and their construction including valves, pipes has been incorporated in this book. 16. Cryogenic transmission electron microscopy (cryo-TEM) for studying the morphology of colloidal drug delivery systems Kuntsche, Judith; Horst, Jennifer C; Bunjes, Heike 2011-01-01 Cryogenic transmission electron microscopy (cryo-TEM) has evolved into an indispensable tool for the characterization of colloidal drug delivery systems. It can be applied to study the size, shape and internal structure of nanoparticulate carrier systems as well as the overall colloidal composition... 17. Two-circuit cryogenic system for cooling and cryostating a superconductive turbogenerator Vishnev, I.P.; Kalitin, P.P.; Krauze, A.I. 1985-01-01 This paper reports the results of experimentation with a cryogenic system which indicate that the system meets the refrigeration and cryostating requirements of superconductive turbogenerators and the thermal, hydraulic, mechanical and electrical calculation procedures which they have developed and tested and which make it possible to plan similar high-power superconductive electrical devices. 18. Vent System Analysis for the Cryogenic Propellant Storage Transfer Ground Test Article Hedayat, A 2013-01-01 To test and validate key capabilities and technologies required for future exploration elements such as large cryogenic propulsion stages and propellant depots, NASA is leading the efforts to develop and design the Cryogenic Propellant Storage and Transfer (CPST) Cryogenic Fluid Management (CFM) payload. The primary objectives of CPST payload are to demonstrate: 1) in-space storage of cryogenic propellants for long duration applications; and 2) in-space transfer of cryogenic propellants. The Ground Test Article (GTA) is a technology development version of the CPST payload. The GTA consists of flight-sized and flight-like storage and transfer tanks, liquid acquisition devices, transfer, and pressurization systems with all of the CPST functionality. The GTA is designed to perform integrated passive and active thermal storage and transfer performance testing with liquid hydrogen (LH2) in a vacuum environment. The GTA storage tank is designed to store liquid hydrogen and the transfer tank is designed to be 5% of the storage tank volume. The LH2 transfer subsystem is designed to transfer propellant from one tank to the other utilizing pressure or a pump. The LH2 vent subsystem is designed to prevent over-pressurization of the storage and transfer tanks. An in-house general-purpose computer program was utilized to model and simulate the vent subsystem operation. The modeling, analysis, and the results will be presented in the final paper. 19. Stochastic cooling in RHIC Brennan J. M.; Blaskiewicz, M.; Mernick, K. 2012-05-20 The full 6-dimensional [x,x'; y,y'; z,z'] stochastic cooling system for RHIC was completed and operational for the FY12 Uranium-Uranium collider run. Cooling enhances the integrated luminosity of the Uranium collisions by a factor of 5, primarily by reducing the transverse emittances but also by cooling in the longitudinal plane to preserve the bunch length. The components have been deployed incrementally over the past several runs, beginning with longitudinal cooling, then cooling in the vertical planes but multiplexed between the Yellow and Blue rings, next cooling both rings simultaneously in vertical (the horizontal plane was cooled by betatron coupling), and now simultaneous horizontal cooling has been commissioned. The system operated between 5 and 9 GHz and with 3 x 10{sup 8} Uranium ions per bunch and produces a cooling half-time of approximately 20 minutes. The ultimate emittance is determined by the balance between cooling and emittance growth from Intra-Beam Scattering. Specific details of the apparatus and mathematical techniques for calculating its performance have been published elsewhere. Here we report on: the method of operation, results with beam, and comparison of results to simulations. 20. Development of a cryogenic FTIR system for measuring very small attenuation coefficients of infrared materials Kaji, Sayumi; Sarugaku, Yuki; Ikeda, Yuji; Nakanishi, Kenshi; Kobayashi, Naoto; Kondo, Sohei; Arasaki, Takayuki; Kawakita, Hideyo 2016-07-01 We have been working on a long-term project for developing a variety of infrared immersion gratings for near- to mid-infrared wavelengths. The transmittance of material is essential to realize high-efficiency immersion gratings for astronomical applications. For a typical grating, the attenuation coefficient αatt must be 70%. However, as there are few reports of αatt optical materials in the literatures, we performed high-accuracy measurements of αatt for a variety of infrared materials applicable to immersion gratings. We have already reported αatt at room temperature for single-crystal Si, single-crystal Ge, CVD-ZnS, CVDZnSe, and high-resistivity single-crystal CdZnTe (Ikeda et al. 2009[7], Kaji et al. 2014[10], and Sarugaku et al. 2016[9]). Next, we proceeded with the measurements of αatt at cryogenic temperatures of 20-80 K range, which is the typical operational temperatures of infrared instruments, and for which the shifts of the band gap and/or the sharpness of the lattice absorption lines from the corresponding room temperature values are expected. Thus, we developed a new cryogenic FTIR system that enables high-accuracy measurements at cryogenic temperatures. The system has a mechanism with which two sample cells and a reference cell can be easily and quickly switched without any vacuum leak or temperature change. Our preliminary measurement of Ge using this cryogenic FTIR system found that both the cut-on and cut-off wavelengths shift to the shorter (from 2.0 to 1.7 μm) and longer (from 10.6 to 10.9 μm) wavelengths, respectively, when the temperature is decreased from room temperature to the cryogenic temperature (<28 K). We plan to complete cryogenic measurements for a variety of infrared materials by the end of 2016. 1. A method to measure internal stray radiation of cryogenic infrared imaging systems under various ambient temperatures Tian, Qijie; Chang, Songtao; Li, Zhou; He, Fengyun; Qiao, Yanfeng 2017-03-01 2. Cryogen Safety Course 8876 Glass, George [Los Alamos National Lab. (LANL), Los Alamos, NM (United States) 2017-06-13 Cryogenics (from the Greek word κρvoζ, meaning frost or icy cold) is the study of the behavior of matter at very cold temperatures. The purpose of this course is to provide trainees with an introduction to cryogen use, the hazards and potential accidents related to cryogen systems, cryogen safety components, and the requirements that govern the design and use of cryogen systems at Los Alamos National Laboratory (LANL). The knowledge you gain will help you keep your workplace safe for yourself and your coworkers. 3. A 400 MHz Low Noise Amplifier at Cryogenic Temperature for Superconductor Filter System 2007-01-01 A cryogenic low noise amplifier (LNA) using Agilent high electron mobility transistor (HEMT) for 380 MHz to 480 MHz is designed and fabricated, and the excellent cryogenic performance in superconducting receiver front-end for communication system is achieved. A special input impedance matching topology is implemented to provide low noise figure (NF) and good input matching in this cryogenic LNA design. The measurement results show that the NF is within 0.25 dB from the minimum NF of a single transistor, the power gain is above 20 dB, the flatness is within 1 dB, and the maximum input return loss is lower than -20 dB in bandwidth. 4. Vibration Measurements to Study the Effect of Cryogen Flow in Superconducting Quadrupole. He,P.; Anerella, M.; aydin, S.; Ganetis, G. Harrison, M.; Jain, A.; Parker, B. 2007-06-25 The conceptual design of compact superconducting magnets for the International Linear Collider final focus is presently under development. A primary concern in using superconducting quadrupoles is the potential for inducing additional vibrations from cryogenic operation. We have employed a Laser Doppler Vibrometer system to measure the vibrations in a spare RHIC quadrupole magnet under cryogenic conditions. Some preliminary results of these studies were limited in resolution due to a rather large motion of the laser head as well as the magnet. As a first step towards improving the measurement quality, a new set up was used that reduces the motion of the laser holder. The improved setup is described, and vibration spectra measured at cryogenic temperatures, both with and without helium flow, are presented. 5. Cryogenic system with GM cryocooler for krypton, xenon separation from hydrogen-helium purge gas Chu, X. X.; Zhang, D. X.; Qian, Y.; Liu, W. [Shanghai Institute of Applied Physics, Chinese Academy of Sciences, Shanghai, 201800 (China); Zhang, M. M.; Xu, D. [Technical Institute of Physics and Chemistry, Chinese Academy of Sciences, Beijing, 100190 (China) 2014-01-29 In the thorium molten salt reactor (TMSR), fission products such as krypton, xenon and tritium will be produced continuously in the process of nuclear fission reaction. A cryogenic system with a two stage GM cryocooler was designed to separate Kr, Xe, and H{sub 2} from helium purge gas. The temperatures of two stage heat exchanger condensation tanks were maintained at about 38 K and 4.5 K, respectively. The main fluid parameters of heat transfer were confirmed, and the structural heat exchanger equipment and cold box were designed. Designed concentrations after cryogenic separation of Kr, Xe and H{sub 2} in helium recycle gas are less than 1 ppb. 6. Cryogenic system with GM cryocooler for krypton, xenon separation from hydrogen-helium purge gas Chu, X. X.; Zhang, M. M.; Zhang, D. X.; Xu, D.; Qian, Y.; Liu, W. 2014-01-01 In the thorium molten salt reactor (TMSR), fission products such as krypton, xenon and tritium will be produced continuously in the process of nuclear fission reaction. A cryogenic system with a two stage GM cryocooler was designed to separate Kr, Xe, and H2 from helium purge gas. The temperatures of two stage heat exchanger condensation tanks were maintained at about 38 K and 4.5 K, respectively. The main fluid parameters of heat transfer were confirmed, and the structural heat exchanger equipment and cold box were designed. Designed concentrations after cryogenic separation of Kr, Xe and H2 in helium recycle gas are less than 1 ppb. 7. RHIC FY15 pp Run RHIC and AGS polarization analysis Huang, H. [Brookhaven National Lab. (BNL), Upton, NY (United States); Adams, P. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2016-02-20 The polarization information is important for the spin physics program in Relativistic Heavy Ion Collider (RHIC). There are discrepancies between AGS and RHIC polarization measurements. First, the face value of AGS polarization is higher than RHIC ones in general. Second, the measured polarization profile (described by the profile ratio R) is stronger in AGS than in RHIC. This note analyzes the polarization data from FY15 pp run period. The results show that the differences between AGS and RHIC polarization measurements are reasonable, but the R value difference is puzzling. The difference between blue and yellow ring is worth of spin simulation to explain. 8. Thermodynamic Vent System for an On-Orbit Cryogenic Reaction Control Engine Hurlbert, Eric A.; Romig, Kris A.; Jimenez, Rafael; Flores, Sam 2012-01-01 A report discusses a cryogenic reaction control system (RCS) that integrates a Joule-Thompson (JT) device (expansion valve) and thermodynamic vent system (TVS) with a cryogenic distribution system to allow fine control of the propellant quality (subcooled liquid) during operation of the device. It enables zero-venting when coupled with an RCS engine. The proper attachment locations and sizing of the orifice are required with the propellant distribution line to facilitate line conditioning. During operations, system instrumentation was strategically installed along the distribution/TVS line assembly, and temperature control bands were identified. A sub-scale run tank, full-scale distribution line, open-loop TVS, and a combination of procured and custom-fabricated cryogenic components were used in the cryogenic RCS build-up. Simulated on-orbit activation and thruster firing profiles were performed to quantify system heat gain and evaluate the TVS s capability to maintain the required propellant conditions at the inlet to the engine valves. Test data determined that a small control valve, such as a piezoelectric, is optimal to provide continuously the required thermal control. The data obtained from testing has also assisted with the development of fluid and thermal models of an RCS to refine integrated cryogenic propulsion system designs. This system allows a liquid oxygenbased main propulsion and reaction control system for a spacecraft, which improves performance, safety, and cost over conventional hypergolic systems due to higher performance, use of nontoxic propellants, potential for integration with life support and power subsystems, and compatibility with in-situ produced propellants. 9. Conceptual design of the cryogenic system and estimation of the recirculated power for CFETR Liu, Xiaogang; Qiu, Lilong; Li, Junjun; Wang, Zhaoliang; Ren, Yong; Wang, Xianwei; Li, Guoqiang; Gao, Xiang; Bi, Yanfang 2017-01-01 The China Fusion Engineering Test Reactor (CFETR) is the next tokamak in China’s roadmap for realizing commercial fusion energy. The CFETR cryogenic system is crucial to creating and maintaining operational conditions for its superconducting magnet system and thermal shields. The preliminary conceptual design of the CFETR cryogenic system has been carried out with reference to that of ITER. It will provide an average capacity of 75 to 80 kW at 4.5 K and a peak capacity of 1300 kW at 80 K. The electric power consumption of the cryogenic system is estimated to be 24 MW, and the gross building area is about 7000 m2. The relationships among the auxiliary power consumed by the cryogenic system, the fusion power gain and the recirculated power of CFETR are discussed, with the suggestion that about 52% of the electric power produced by CFETR in phase II must be recirculated to run the fusion test reactor. 10. Automatic Management Systems for the Operation of the Cryogenic Test Facilities for LHC Series Superconducting Magnets Tovar-Gonzalez, A; Herblin, L; Lamboy, J P; Vullierme, B 2006-01-01 Prior to their final preparation before installation in the tunnel, the ~1800 series superconducting magnets of the LHC machine shall be entirely tested at reception on modular test facilities. The operation 24 hours per day of the cryogenic test facilities is conducted in turn by 3-operator teams, assisted in real time by the use of the Test Bench Priorities Handling System, a process control application enforcing the optimum use of cryogenic utilities and of the "Tasks Tracking System", a web-based e-traveller application handling 12 parallel 38-task test sequences. This paper describes how such computer-based management systems can be used to optimize operation of concurrent test benches within technical boundary conditions given by the cryogenic capacity, and how they can be used to study the efficiency of the automatic steering of all individual cryogenic sub-systems. Finally, this paper presents the overall performance of the cryomagnet test station for the first complete year of operation at high produ... 11. Design and Implementation of Quench Protection in the EAST Cryogenic Control System ZHOU Zhiwei; BAI Hongyu; ZHUANG Ming; HU Liangbing; XIA Genhai 2009-01-01 An outline of the complete design of the cryogenic system, including the detection of quench signals, the processing flow after a quench and the concrete measures of pressure release for EAST is described in this paper. The hardware and software configurations on DeltaV DCS are illustrated in detail. The results of quench protection testing in cooling experiments are also analyzed. 12. First Operational Experience and Performance Optimization of the ATLAS Magnet Cryogenic System Delruelle, N; Dudarev, A; Passardi, G; Ten Kate, H H J 2012-01-01 The ATLAS magnet system, comprising a superconducting central solenoid and three superconducting toroids, has been successfully ramped up for the first time to the nominal operational current of 20.4 kA on 4th August 2008. Since then, new cryogenic operational challenges have been raised, like the smoothing of steady-state parameters, the enhancing of transient procedures to minimize thermal shocks on the magnet cold masses, the optimization of the complex cryogenic system in order to reduce the compressors electric consumption and finally how to avoid regular clogging of the shield refrigerator by water contamination. This paper presents the heat load identification of the various cryogenic sub-systems done at 4.5 K and how one of these loads was reduced, what was gained - in term of electrical consumption - by tuning the turbines settings of the main refrigerator and finally the first consolidation of the cryogenic system implemented in order to minimize the detector downtime during LHC beam runs. 13. Micromechanics, fracture mechanics and gas permeability of composite laminates for cryogenic storage systems Choi, Sukjoo A micromechanics method is developed to investigate microcrack propagation in a liquid hydrogen composite tank at cryogenic temperature. The unit cell is modeled using square and hexagonal shapes depends on fiber and matrix layout from microscopic images of composite laminates. Periodic boundary conditions are applied to the unit cell. The temperature dependent properties are taken into account in the analysis. The laminate properties estimated by the micromechanics method are compared with empirical solutions using constituent properties. The micro stresses in the fiber and matrix phases based on boundary conditions in laminate level are calculated to predict the formation of microcracks in the matrix. The method is applied to an actual liquid hydrogen storage system. The analysis predicts micro stresses in the matrix phase are large enough to cause microcracks in the composite. Stress singularity of a transverse crack normal to a ply-interface is investigated to predict the fracture behavior at cryogenic conditions using analytical and finite element analysis. When a transverse crack touches a ply-interface of a composite layer with same fiber orientation, the stress singularity is equal to ½. When the transverse crack propagates to a stiffer layer normal to a ply-direction, the singularity becomes less than ½ and vice versa. Finite element analysis is performed to evaluate fracture toughness of a laminated beam subjected to the fracture load measured by the fracture experiment at room and cryogenic temperatures. As results, the fracture load at cryogenic temperature is significantly lower than that at room temperature. However, when thermal stresses are taken into consideration, for both cases of room and cryogenic temperatures, the variation of fracture toughness becomes insignificant. The result indicates fracture toughness is a characteristic property which is independent to temperature changes. The experimental analysis is performed to investigate the 14. The RHIC polarized source upgrade Zelenski, A.; Atoian, G.; Davydenko, V.; Ivanov, A.; Kolmogorov, A.; Ritter, J.; Steski, D.; Zubets, V. 2010-09-27 The RHIC polarized H{sup -} ion source is being upgraded to higher intensity (5-10 mA) and polarization for use in the RHIC polarization physics program at enhanced luminosity RHIC operation. The higher beam peak intensity will allow reduction of the transverse beam emittance at injection to AGS to reduce polarization losses in AGS. There is also a planned RHIC luminosity upgrade by using the electron beam lens to compensate the beam-beam interaction at collision points. This upgrade is also essential for future BNL plans for a high-luminosity electron - proton (ion) Collider eRHIC. 15. Insertion of helical Siberian snakes in RHIC Luccio, A.; Pilat, F. 1995-05-01 Spin rotators and Siberian snakes for RHIC can be built using 4 helical magnets obtained, by twisting, from the cosine dipoles. The authors found that the fringe fields are important. In the calculations they have used a plausible model for the fringe. However, only magnetic measurements on the prototypes presently being built will allow a final optimization. The linear coupling at injection, {Delta}Q{sub min} < 10{sup {minus}2}, is well within the range of the RHIC decoupling system. At storage, the coupling introduced by the devices ({Delta}Q{sub min} < 10{sup {minus}4}) is negligible. 16. BUNCHED BEAM STOCHASTIC COOLING PROJECT FOR RHIC. 2005-09-18 The main performance limitation for RHIC is emittance growth caused by IntraBeam Scattering during the store. We have developed a longitudinal bunched-beam stochastic cooling system in the 5-8 GHz band which will be used to counteract IBS longitudinal emittance growth and prevent de-bunching during the store. Solutions to the technical problems of achieving sufficient kicker voltage and overcoming the electronic saturation effects caused by coherent components within the Schottky spectrum are described. Results from tests with copper ions in RHIC during the FY05 physics run, including the observation of signal suppression, are presented. 17. Production of Quarkonia at RHIC Vertesi, Robert 2015-01-01 The production of different quarkonium states provides unique insight to the hot and cold nuclear matter effects in the strongly interacting medium that is formed in high energy heavy ion collisions. While LHC explores the energy frontier, RHIC has a broad physics program to explore the nuclear modification at different energies in a wide range of systems. Some of the most interesting recent results on $J/\\psi$ and $\\Upsilon$ production in p+p, d+Au and A+A collisions from PHENIX and STAR are summarized in this work. 18. Thermal and structural analysis of a cryogenic conduction cooling system for a HTS NMR magnet In, Se Hwan; Hong, Yong Jun; Yeom, Han Kil; Ko, Hyo Bong; Park, Seong Je [Korea Institute of Machinery and Materials, Daejeon (Korea, Republic of) 2016-03-15 The superconducting NMR magnets have used cryogen such as liquid helium for their cooling. The conduction cooling method using cryocoolers, however, makes the cryogenic cooling system for NMR magnets more compact and user-friendly than the cryogen cooling method. This paper describes the thermal and structural analysis of a cryogenic conduction cooling system for a 400 MHz HTS NMR magnet, focusing on the magnet assembly. The highly thermo-conductive cooling plates between HTS double pancake coils are used to transfer the heat generated in coils, namely Joule heating at lap splice joints, to thermal link blocks and finally the cryocooler. The conduction cooling structure of the HTS magnet assembly preliminarily designed is verified by thermal and structural analysis. The orthotropic thermal properties of the HTS coil, thermal contact resistance and radiation heat load are considered in the thermal analysis. The thermal analysis confirms the uniform temperature distribution for the present thermal design of the NMR magnet within 0.2 K. The mechanical stress and the displacement by the electromagnetic force and the thermal contraction are checked to verify structural stability. The structural analysis indicates that the mechanical stress on each component of the magnet is less than its material yield strength and the displacement is acceptable in comparison with the magnet dimension. 19. LONGITUDINAL SOLUTIONS IN RHIC. 2003-05-12 Stable, coherent, longitudinal oscillations have been observed in the RHIC accelerator. Within the context of perturbation theory, the beam parameters and machine impedance suggest these oscillations should be Landau damped. When nonlinear effects are included, long lived, stable oscillations become possible for low intensity beams. Simulations and theory are compared with data. 20. Passive Capillary Pumped Cryocooling System for Zero-Boil-Off Cryogen Storage Tanks Project National Aeronautics and Space Administration — Significant cost and weight savings of a space mission can be achieved by improving the cryogenic storage technology. Added cryogen mass due to the cryogen boil-off,... 1. Upgrade to the Cryogenic Hydrogen Gas Target Monitoring System Slater, Michael; Tribble, Robert 2013-10-01 The cryogenic hydrogen gas target at Texas A&M is a vital component for creating a secondary radioactive beam that is then used in experiments in the Momentum Achromat Recoil Spectrometer (MARS). A stable beam from the K500 superconducting cyclotron enters the gas cell and some incident particles are transmuted by a nuclear reaction into a radioactive beam, which are separated from the primary beam and used in MARS experiments. The pressure in the target chamber is monitored so that a predictable isotope production rate can be assured. A black box'' received the analog pressure data and sent RS232 serial data through an outdated serial connection to an outdated Visual Basic 6 (VB6) program, which plotted the chamber pressure continuously. The black box has been upgraded to an Arduino UNO microcontroller [Atmel Inc.], which can receive the pressure data and output via USB to a computer. It has been programmed to also accept temperature data for future upgrade. A new computer program, with updated capabilities, has been written in Python. The software can send email alerts, create audible alarms through the Arduino, and plot pressure and temperature. The program has been designed to better fit the needs of the users. Funded by DOE and NSF-REU Program. Hu, Zhongjun; Zhang, Ning; Li, Zhengyu; Li, Q. 2012-06-01 3. A Cryogenic Fluid System Simulation in Support of Integrated Systems Health Management Barber, John P.; Johnston, Kyle B.; Daigle, Matthew 2013-01-01 Simulations serve as important tools throughout the design and operation of engineering systems. In the context of sys-tems health management, simulations serve many uses. For one, the underlying physical models can be used by model-based health management tools to develop diagnostic and prognostic models. These simulations should incorporate both nominal and faulty behavior with the ability to inject various faults into the system. Such simulations can there-fore be used for operator training, for both nominal and faulty situations, as well as for developing and prototyping health management algorithms. In this paper, we describe a methodology for building such simulations. We discuss the design decisions and tools used to build a simulation of a cryogenic fluid test bed, and how it serves as a core technology for systems health management development and maturation. 4. A cryogenic phase locking loop system for a superconducting integrated receiver Khudchenko, A. V.; Koshelets, V. P.; Dmitriev, P. N.; Ermakov, A. B.; Yagoubov, P. A.; Pylypenko, O. M. 2009-08-01 The authors present a new cryogenic device, an ultrawideband cryogenic phase locking loop system (CPLL). The CPLL was developed for phase locking of a flux-flow oscillator (FFO) in a superconducting integrated receiver (SIR) but can be used for any cryogenic terahertz oscillator. The key element of the CPLL is the cryogenic phase detector (CPD), a recently proposed new superconducting element. The CPD is an innovative implementation of a superconductor-insulator-superconductor tunnel junction. All components of the CPLL reside inside a cryostat at 4.2 K, with the loop length of cables 50 cm and the total loop delay 4.5 ns. So small a delay results in a CPLL synchronization bandwidth as wide as 40 MHz and allows phase locking of more than 60% of the power emitted by the FFO, even for FFO linewidths of about 11 MHz. This percentage of phase locked power is three times that achieved with conventional room temperature PLLs. Such an improvement enables reducing the FFO phase noise and extending the SIR operation range. 5. Cryogenic Phase-Locking Loop System Based on SIS Tunnel Junction Khudchenko, A. V.; Koshelets, V. P.; Kalashnikov, K. V. An ultra-wideband cryogenic phase-locking loop (CPLL) system is a new cryogenic device. The CPLL is intended for phase-locking of a Flux-Flow Oscillator (FFO) in a Superconducting Integrated Receiver (SIR) but can be used for any cryogenic terahertz oscillator. The key element of the CPLL is Cryogenic Phase Detector (CPD), a recently proposed new superconducting element. The CPD is an innovative implementation of superconductor-insulator-superconductor (SIS) tunnel junction. All components of the CPLL reside inside a cryostat at 4.2 K, with the loop length of about 50 cm and the total loop delay 5.5 ns. Such a small delay results in CPLL synchronization bandwidth as wide as 40 MHz and allows phase-locking of more than 60% of the power emitted by the FFO even for FFO linewidth of about 10 MHz. This percentage of phase-locked power three times exceeds that achieved with conventional room-temperature PLLs. Such an improvement enables reducing the FFO phase noise and extending the SIR operation range.Another new approach to the FFO phase-locking has been proposed and experimentally verified. The FFO has been synchronized by a cryogenic harmonic phase detector (CHPD) based on the SIS junction. The CHPD operates simultaneously as the harmonic mixer (HM) and phase detector. We have studied the HM based on the SIS junction theoretically; in particular we calculated 3D dependences of the HM output signal power versus the bias voltage and the LO power. Results of the calculations have been compared with experimental measurements. Good qualitative and quantitative correspondence has been achieved. The FFO phase-locking by the CHPD has been demonstrated. Such a PLL system is expected to be extra wideband. This concept is very promising for building of the multi-pixel SIR array. 6. Features of applying systems approach for evaluating the reliability of cryogenic systems for special purposes E. D. Chertov 2016-01-01 Full Text Available Summary. The analysis of cryogenic installations confirms objective regularity of increase in amount of the tasks solved by systems of a special purpose. One of the most important directions of development of a cryogenics is creation of installations for air separation product receipt, namely oxygen and nitrogen. Modern aviation complexes require use of these gases in large numbers as in gaseous, and in the liquid state. The onboard gas systems applied in aircraft of the Russian Federation are subdivided on: oxygen system; air (nitric system; system of neutral gas; fire-proof system. Technological schemes ADI are in many respects determined by pressure of compressed air or, in a general sense, a refrigerating cycle. For the majority ADI a working body of a refrigerating cycle the divided air is, that is technological and refrigerating cycles in installation are integrated. By this principle differentiate installations: low pressure; average and high pressure; with detander; with preliminary chilling. There is also insignificant number of the ADI types in which refrigerating and technological cycles are separated. These are installations with external chilling. For the solution of tasks of control of technical condition of the BRV hardware in real time and estimates of indicators of reliability it is offered to use multi-agent technologies. Multi-agent approach is the most acceptable for creation of SPPR for reliability assessment as allows: to redistribute processing of information on elements of system that leads to increase in overall performance; to solve a problem of accumulating, storage and recycling of knowledge that will allow to increase significantly efficiency of the solution of tasks of an assessment of reliability; to considerably reduce intervention of the person in process of functioning of system that will save time of the person of the making decision (PMD and will not demand from it special skills of work with it. 7. Study on cooling process of cryogenic system for superconducting magnets of BEPCⅡ ZONG Zhan-Guo; LIU Li-Qiang; XIONG Lian-You; LI Shao-Peng; XU Qing-Jin; HE Kun; ZHANG Liang; GAO Jie 2008-01-01 In the upgrade project of the Beijing Electron Positron Collider(BEPCⅡ),three superconducting magnets are employed to realize the goal of two orders of magnitude higher luminosity.A cryogenic system with a total capacity of 0.5 kW at 4.5 K was built at the Institute of High Energy Physics(IHEP)to support the operations of these superconducting devices.For preparing the commissioning of the system,the refrigeration process Was simulated and analyrzed numerically.The numerical model Was based on the latest engineering progress and focused on the normal operation mode.The pressure and temperature profiles of the cryogenic system are achieved with the simulation.The influence of the helium mass flow rates to cool superconducting magnets on the thermodynamic parameters of their normal operation is also studied and discussed in this paper. 8. Installation and pre-commissioning of the cryogenic system of JT-60SA tokamak Hoa, C.; Michel, F.; Roussel, P.; Fejoz, P.; Girard, S.; Goncalves, R.; Lamaison, V.; Natsume, K.; Kizu, K.; Koide, Y.; Yoshida, K.; Cardella, A.; Portone, A.; Verrecchia, M.; Wanner, M.; Beauvisage, J.; Bertholat, F.; Gaillard, G.; Heloin, V.; Langevin, B.; Legrand, J.; Maire, S.; Perrier, J. M.; Pudys, V. 2017-02-01 The cryogenic system for the superconducting tokamak JT-60SA is currently being commissioned in Naka, Japan and shall be ready for operation in summer 2016. This contribution is part of the Broader Approach agreement between Japan and Europe. With an equivalent refrigeration capacity of about 9.5 kW at 4.5 K the cryogenic system will supply cryo-pump panels at 3.7 K, superconducting magnets and their structures at 4.4 K, high temperature superconducting current leads at 50 K and thermal shields between 80 K and 100 K. The system has been specifically designed to handle large pulse loads at 4.4 K during plasma operation. The mechanical and electrical assembly of the cryogenic system has been achieved within six months by October 2015. The main contractor Air Liquide Advanced Technology (AL-aT) have supplied eight parallel working screw compressors with a common oil removal and dryer system, a Refrigeration Cold Box and an Auxiliary Cold box with cold rotating machines. F4E has provided six GHe storage vessels and QST has provided the complete infrastructure and the facilities for the utilities. The paper gives an overview of the main design features, the infrastructure and the status of installation and pre-commissioning. 9. Polarized Proton Collisions at RHIC Bai, Mei; Alekseev, Igor G; Alessi, James; Beebe-Wang, Joanne; Blaskiewicz, Michael; Bravar, Alessandro; Brennan, Joseph M; Bruno, Donald; Bunce, Gerry; Butler, John J; Cameron, Peter; Connolly, Roger; De Long, Joseph; Drees, Angelika; Fischer, Wolfram; Ganetis, George; Gardner, Chris J; Glenn, Joseph; Hayes, Thomas; Hseuh Hsiao Chaun; Huang, Haixin; Ingrassia, Peter; Iriso, Ubaldo; Laster, Jonathan S; Lee, Roger C; Luccio, Alfredo U; Luo, Yun; MacKay, William W; Makdisi, Yousef; Marr, Gregory J; Marusic, Al; McIntyre, Gary; Michnoff, Robert; Montag, Christoph; Morris, John; Nicoletti, Tony; Oddo, Peter; Oerter, Brian; Osamu, Jinnouchi; Pilat, Fulvia Caterina; Ptitsyn, Vadim; Roser, Thomas; Satogata, Todd; Smith, Kevin T; Svirida, Dima; Tepikian, Steven; Tomas, Rogelio; Trbojevic, Dejan; Tsoupas, Nicholaos; Tuozzolo, Joseph; Vetter, Kurt; Wilinski, Michelle; Zaltsman, Alex; Zelenski, Anatoli; Zeno, Keith; Zhang, S Y 2005-01-01 The Relativistic Heavy Ion Collider~(RHIC) provides not only collisions of ions but also collisions of polarized protons. In a circular accelerator, the polarization of polarized proton beam can be partially or fully lost when a spin depolarizing resonance is encountered. To preserve the beam polarization during acceleration, two full Siberian snakes were employed in RHIC to avoid depolarizing resonances. In 2003, polarized proton beams were accelerated to 100~GeV and collided in RHIC. Beams were brought into collisions with longitudinal polarization at the experiments STAR and PHENIX by using spin rotators. RHIC polarized proton run experience demonstrates that optimizing polarization transmission efficiency and improving luminosity performance are significant challenges. Currently, the luminosity lifetime in RHIC is limited by the beam-beam effect. The current state of RHIC polarized proton program, including its dedicated physics run in 2005 and efforts to optimize luminosity production in beam-beam limite... 10. A study of RHIC crystal collimation Trbojevic, D; Harrison, M; Parker, B; Thompson, P; Stevens, A; Mokhov, N V; Drozhdin, A I 1998-01-01 The Relativistic Heavy Ion Collider (RHIC) will experience increasing longitudinal and transverse heavy ion emittances, mostly due to intra-beam scattering (IBS). The experiments in RHIC are expected to not only have reduced luminosities due to IBS but also an unwanted beam halo. Primary betatron collimators will be used to remove the large amplitude particles. The efficiency of the primary collimator in RHIC depends very much on the alignment of the jaws which needs to be within few micro-radians for the best conditions. As proposed by V. Biryukov bent crystals could be used to improve the efficiency of an existing collimation system by installing them upstream of the collimator jaws. Bent crystals have been successfully used in SPS, Protvino and Fermilab for extraction of the beam particles channeled through them. This study examines possible improvements of the primary collimator system for heavy ions at RHIC by use of bent crystals. Bent crystals will reduce the collimator jaws alignment requirement and w... 11. Exergy Analysis of the Cryogenic Helium Distribution System for the Large Hadron Collider (lhc) Claudet, S.; Lebrun, Ph.; Tavian, L.; Wagner, U. 2010-04-01 The Large Hadron Collider (LHC) at CERN features the world's largest helium cryogenic system, spreading over the 26.7 km circumference of the superconducting accelerator. With a total equivalent capacity of 145 kW at 4.5 K including 18 kW at 1.8 K, the LHC refrigerators produce an unprecedented exergetic load, which must be distributed efficiently to the magnets in the tunnel over the 3.3 km length of each of the eight independent sectors of the machine. We recall the main features of the LHC cryogenic helium distribution system at different temperature levels and present its exergy analysis, thus enabling to qualify second-principle efficiency and identify main remaining sources of irreversibility. 12. Exergy Analysis of the Cryogenic Helium Distribution System for the Large Hadron Collider (LHC) Claudet, S; Tavian, L; Wagner, U 2010-01-01 The Large Hadron Collider (LHC) at CERN features the world’s largest helium cryogenic system, spreading over the 26.7 km circumference of the superconducting accelerator. With a total equivalent capacity of 145 kW at 4.5 K including 18 kW at 1.8 K, the LHC refrigerators produce an unprecedented exergetic load, which must be distributed efficiently to the magnets in the tunnel over the 3.3 km length of each of the eight independent sectors of the machine. We recall the main features of the LHC cryogenic helium distribution system at different temperature levels and present its exergy analysis, thus enabling to qualify second-principle efficiency and identify main remaining sources of irreversibility.. 13. IBS simulation with different RF configurations in RHIC Liu, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Fedotov, A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Minty, M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ptitsyn, V. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2016-11-07 It is a crucial task to understand the beam emittance growth during RHIC cycle and the underlying causes. One would benefit not just for the current operation of RHIC, also for the design of eRHIC. This report focuses on the Intra-Beam Scattering (IBS) contribution to the emittance growth of the proton beam with two different configurations of RF system. The answers to these questions will be given in the end of the report; can IBS explain the emittance growth all alone? What’s the difference of IBS growth rates for different RF configurations? 14. RHIC PLANS TOWARDS HIGHER LUMINOSITY FEDOTOV,A. 2007-06-25 The Relativistic Heavy Ion Collider (RHIC) is designed to provide luminosity over a wide range of beam energies and species, including heavy ions, polarized protons, and tric beam collisions. In the first seven years of operation there has been a rapid increase in the achieved peak and average luminosity, substantially exceeding design values. Work is presently underway to achieve the Enhanced Design parameters. Planned major upgrades include the Electron Beam Ion Source (EBIS), RHIC-11, and construction of an electron-ion collider (eRHIC). We review the expected RHIC upgrade performance. Electron cooling and its impact on the luminosity both for heavy ions and protons are discussed in detail. 15. RHIC physics overview Ruan, Lijuan 2010-01-01 The results from data taken during the last several years at the Relativistic Heavy-Ion Collider (RHIC) will be reviewed in the paper. Several selected topics that further our understanding of constituent quark scaling, jet quenching and color screening effect of heavy quarkonia in the hot dense medium will be presented. Detector upgrades will further probe the properties of Quark Gluon Plasma. Future measurements with upgraded detectors will be presented. The discovery perspectives from future measurements will also be discussed. 16. Electromagnetic Signals at RHIC Turbide, S; Turbide, Simon; Gale, Charles 2006-01-01 We calculate the direct photon yield in central and mid-peripheral Au+Au collisions at the Relativistic Heavy-Ion Collider (RHIC). The processes involving the propagation of jets have been convolved with a leading order treatment of jet energy loss in the medium and a one dimensional hydrodynamic expansion. The quark-gluon plasma (QGP) contribution turns out to be important, especially the in-medium conversion of a jet into a photon, for successfully describing recent photon measurements. 17. The performance of a piezoelectric-sensor-based SHM system under a combined cryogenic temperature and vibration environment Qing, Xinlin P.; Beard, Shawn J.; Kumar, Amrita; Sullivan, Kevin; Aguilar, Robert; Merchant, Munir; Taniguchi, Mike 2008-10-01 A series of tests have been conducted to determine the survivability and functionality of a piezoelectric-sensor-based active structural health monitoring (SHM) SMART Tape system under the operating conditions of typical liquid rocket engines such as cryogenic temperature and vibration loads. The performance of different piezoelectric sensors and a low temperature adhesive under cryogenic temperature was first investigated. The active SHM system for liquid rocket engines was exposed to flight vibration and shock environments on a simulated large booster LOX-H2 engine propellant duct conditioned to cryogenic temperatures to evaluate the physical robustness of the built-in sensor network as well as operational survivability and functionality. Test results demonstrated that the developed SMART Tape system can withstand operational levels of vibration and shock energy on a representative rocket engine duct assembly, and is functional under the combined cryogenic temperature and vibration environment. 18. Soft x-ray backlighting of cryogenic implosions using a narrowband crystal imaging system (invited). Stoeckl, C; Bedzyk, M; Brent, G; Epstein, R; Fiksel, G; Guy, D; Goncharov, V N; Hu, S X; Ingraham, S; Jacobs-Perkins, D W; Jungquist, R K; Marshall, F J; Mileham, C; Nilson, P M; Sangster, T C; Shoup, M J; Theobald, W 2014-11-01 A high-performance cryogenic DT inertial confinement fusion implosion experiment is an especially challenging backlighting configuration because of the high self-emission of the core at stagnation and the low opacity of the DT shell. High-energy petawatt lasers such as OMEGA EP promise significantly improved backlighting capabilities by generating high x-ray intensities and short emission times. A narrowband x-ray imager with an astigmatism-corrected bent quartz crystal for the Si Heα line at ∼1.86 keV was developed to record backlit images of cryogenic direct-drive implosions. A time-gated recording system minimized the self-emission of the imploding target. A fast target-insertion system capable of moving the backlighter target ∼7 cm in ∼100 ms was developed to avoid interference with the cryogenic shroud system. With backlighter laser energies of ∼1.25 kJ at a 10-ps pulse duration, the radiographic images show a high signal-to-background ratio of >100:1 and a spatial resolution of the order of 10 μm. The backlit images can be used to assess the symmetry of the implosions close to stagnation and the mix of ablator material into the dense shell. 19. Gas gap heat switch for a cryogen-free magnet system Barreto, J.; Borges de Sousa, P.; Martins, D.; Kar, S.; Bonfait, G.; Catarino, I. 2015-12-01 Cryogen-free superconducting magnet systems (CFMS) have become popular over the last two decades for the simple reason that the use of liquid helium is rather cumbersome and that helium is a scarce resource. Some available CFMS use a mechanical cryocooler as the magnet's cold source. However, the variable temperature insert (VTI) for some existing CFMS are not strictly cryogen-free as they are still based on helium gas circulation through the sample space. We designed a prototype of a gas gap heat switch (GGHS) that allows a thermal management of a completely cryogen-free magnet system, with no helium losses. The idea relies on a parallel cooling path to a variable temperature insert (VTI) of a magnetic properties measurement system under development at Inter-University Accelerator Centre. A Gifford-McMahon cryocooler (1.5 W @ 4.2 K) would serve primarily as the cold source of the superconducting magnet, dedicating 1 W to this cooling, under quite conservative safety factors. The remaining cooling power (0.5 W) is to be diverted towards a VTI through a controlled GGHS that was designed and built with a 80 μm gap width. The built GGHS thermal performance was measured at 4 K, using helium as the exchange gas, and its conductance is compared both with a previously developed analytical model and a finite element method. Lessons learned lead to a new and more functional prototype yet to be reported. 20. Polarized proton beam for eRHIC Huang, H. [Brookhaven National Lab. (BNL), Upton, NY (United States); Meot, F. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ptitsyn, V. [Brookhaven National Lab. (BNL), Upton, NY (United States); Roser, T. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-05-03 RHIC has provided polarized proton collisions from 31 GeV to 255 GeV in the past decade. To preserve polarization through numerous depolarizing resonances through the whole accelerator chain, harmonic orbit correction, partial snakes, horizontal tune jump system and full snakes have been used. In addition, close attentions have been paid to betatron tune control, orbit control and beam line alignment. The polarization of 60% at 255 GeV has been delivered to experiments with 1.8×1011 bunch intensity. For the eRHIC era, the beam brightness has to be maintained to reach the desired luminosity. Since we only have one hadron ring in the eRHIC era, existing spin rotator and snakes can be converted to six snake configuration for one hadron ring. With properly arranged six snakes, the polarization can be maintained at 70% at 250 GeV. This paper summarizes the effort and plan to reach high polarization with small emittance for eRHIC. 1. Energy Efficiency of large Cryogenic Systems: the LHC Case and Beyond Claudet, S; Ferlin, G; Lebrun, P; Tavian, L; Wagner, U 2013-01-01 Research infrastructures for high-energy and nuclear physics, nuclear fusion and production of high magnetic fields are increasingly based on applied superconductivity and associated cryogenics in their quest for scientific breakthroughs at affordable capital and operation costs, a condition for their acceptance and sustained funding by society. The thermodynamic penalty for operating at low temperature makes energy efficiency a key requirement for their large cryogenic systems, from conceptual design to procurement, construction and operation. Meeting this requirement takes a combined approach on several fronts in parallel: management of heat loads and sizing of cooling duties, distribution of cooling power matching the needs of the superconducting devices, efficient production of refrigeration, optimal control resting on precise instrumentation and diagnostics, as well as a targeted industrial procurement policy. The case of the Large Hadron Collider (LHC) at CERN is presented. Potential improvements for fu... 2. Enhanced configuration of a water detritiation system; impact on ITER Isotope Separation System based cryogenic distillation Cristescu, Ion, E-mail: ion.cristescu@kit.edu 2016-11-01 Highlights: • An enhanced configuration of ITER WDS has been developed. • The proposed configuration allows minimization of hazards due to the reduction of tritium inventory. • The load on the tritium recovery system (ITER ISS) is minimized with benefits on mitigation of the explosion hazards. - Abstract: Tritiated water is generated in the ITER systems by various sources and may contain deuterium and tritium at various concentrations. The reference process for the ITER Water Detritiation System is based on Combined Electrolysis Catalytic Exchange (CECE) configuration. During long time operation of the CECE process, the accumulation of deuterium in the electrolysis unit and consequently along the Liquid Phase Catalytic Exchange (LPCE) column is unavoidable with consequences on the overall detritiation factor of the system. Beside the deuterium issue in the process, the large amount of the tritiated water with tritium activity up to 500 Ci/kg in the electrolysis cells is a concern from the safety aspect of the plant. The enhanced configuration of a system for processing tritiated water allows mitigation of the effects due to deuterium accumulation and also reduction of tritium inventory within the electrolysis system. In addition the benefits concerning to the interface between the water detritiation system and tritium recovery based cryogenic distillation are also presented. 3. Proceedings of the symposium on RHIC detector R&D Makdisi, Y.; Stevens, A.J. [eds. 1991-12-31 This report contains papers on the following topics: Development of Analog Memories for RHIC Detector Front-end Electronic Systems; Monolithic Circuit Development for RHIC at Oak Ridge National Laboratory; Highly Integrated Electronics for the STAR TPC; Monolithic Readout Circuits for RHIC; New Methods for Trigger Electronics Development; Neurocomputing methods for Pattern Recognition in Nuclear Physics; The Development of a Silicon Multiplicity Detector System; The Vertex Detector for the Lepton/Photon Collaboration; Simulations of Silicon Vertex Tracker for STAR Experiment at RHIC; Calorimeter/Absorber Optimization for a RHIC Dimuon Experiment (RD-10 Project); Applications of the LAHET simulation Code to Relativistic Heavy Ion Detectors; Highly Segmented, High Resolution Time-of-Flight System; Research and Development on a Sub 100 Picosecond Time-of-Flight System Based on Silicon Avalance Diodes; Behavior of TPCs in a High Particle Flux Environment; Generic R&D on Undoped Cesium Iodide and Lead Fluoride; and A Transition Radiation Detector for RHIC Featuring Accurate Tracking and dE/dx Particle Identification. Selected papers were processed separately for inclusion in the Energy Science and Technology Database. 4. Proceedings of the symposium on RHIC detector R&D Makdisi, Y.; Stevens, A.J. [eds. 1991-12-31 This report contains papers on the following topics: Development of Analog Memories for RHIC Detector Front-end Electronic Systems; Monolithic Circuit Development for RHIC at Oak Ridge National Laboratory; Highly Integrated Electronics for the STAR TPC; Monolithic Readout Circuits for RHIC; New Methods for Trigger Electronics Development; Neurocomputing methods for Pattern Recognition in Nuclear Physics; The Development of a Silicon Multiplicity Detector System; The Vertex Detector for the Lepton/Photon Collaboration; Simulations of Silicon Vertex Tracker for STAR Experiment at RHIC; Calorimeter/Absorber Optimization for a RHIC Dimuon Experiment (RD-10 Project); Applications of the LAHET simulation Code to Relativistic Heavy Ion Detectors; Highly Segmented, High Resolution Time-of-Flight System; Research and Development on a Sub 100 Picosecond Time-of-Flight System Based on Silicon Avalance Diodes; Behavior of TPCs in a High Particle Flux Environment; Generic R&D on Undoped Cesium Iodide and Lead Fluoride; and A Transition Radiation Detector for RHIC Featuring Accurate Tracking and dE/dx Particle Identification. Selected papers were processed separately for inclusion in the Energy Science and Technology Database. 5. Cryogenic System for the Test Facilities of the ATLAS Liquid Argon Calorimeter Modules Bremer, J; Chalifour, M; Haug, F; Passardi, Giorgio; Tischhauser, Johann 1998-01-01 To perform cold tests on the different modules of the ATLAS liquid argon calorimeter, a cryogenic system has been constructed and is now operated at the CERN North Experimental Area. Three different test cryostats will house the modules, which can also be exposed to particle beams for calibration purposes. The three cryostats share a common liquid argon and liquid nitrogen distribution system. The system is rather complex since it has to allow operations of the three cryostats at the same time. Liquid nitrogen is used as cold source for both the cool-down of the cryostats and for normal operation of the cryostats filled with liquid argon. 6. Control System and Operation of the Cryogenic Test Facilities for LHC Series Superconducting Magnets Axensalva, J; Lamboy, J P; Tovar-Gonzalez, A; Vullierme, B 2005-01-01 Prior to their final preparation before installation in the tunnel, the ~1800 series superconducting magnets of the LHC machine will be entirely tested at reception on modular test facilities using dedicated control systems. The test facilities are operated by teams of high-skilled and trained operators. This paper describes the architecture of the control & supervision system of the cryogenic test facilities as well as the tools and management systems developed to help in real time all involved operation teams in order to reach the required industrial production level. 7. Polarized proton collider at RHIC Alekseev, I.; Allgower, C.; Bai, M.; Batygin, Y.; Bozano, L.; Brown, K.; Bunce, G.; Cameron, P.; Courant, E.; Erin, S.; Escallier, J.; Fischer, W.; Gupta, R.; Hatanaka, K.; Huang, H.; Imai, K.; Ishihara, M.; Jain, A.; Lehrach, A.; Kanavets, V.; Katayama, T.; Kawaguchi, T.; Kelly, E.; Kurita, K.; Lee, S. Y.; Luccio, A.; MacKay, W. W.; Mahler, G.; Makdisi, Y.; Mariam, F.; McGahern, W.; Morgan, G.; Muratore, J.; Okamura, M.; Peggs, S.; Pilat, F.; Ptitsin, V.; Ratner, L.; Roser, T.; Saito, N.; Satoh, H.; Shatunov, Y.; Spinka, H.; Syphers, M.; Tepikian, S.; Tominaka, T.; Tsoupas, N.; Underwood, D.; Vasiliev, A.; Wanderer, P.; Willen, E.; Wu, H.; Yokosawa, A.; Zelenski, A. N. 2003-03-01 In addition to heavy ion collisions (RHIC Design Manual, Brookhaven National Laboratory), RHIC will also collide intense beams of polarized protons (I. Alekseev, et al., Design Manual Polarized Proton Collider at RHIC, Brookhaven National Laboratory, 1998 [2]), reaching transverse energies where the protons scatter as beams of polarized quarks and gluons. The study of high energy polarized protons beams has been a long term part of the program at BNL with the development of polarized beams in the Booster and AGS rings for fixed target experiments. We have extended this capability to the RHIC machine. In this paper we describe the design and methods for achieving collisions of both longitudinal and transverse polarized protons in RHIC at energies up to s=500 GeV. 8. Numerical Modeling of Fluid Transient in Cryogenic Fluid Network of Rocket Propulsion System Majumdar, Alok; Flachbart, Robin 2003-01-01 Fluid transients, also known as water hammer, can have a significant impact on the design and operation of both spacecraft and launch vehicles propulsion systems. These transients often occur at system activation and shut down. For ground safety reasons, many spacecrafts are launched with the propellant lines dry. These lines are often evacuated by the time the spacecraft reaches orbit. When the propellant isolation valve opens during propulsion system activation, propellant rushes into lines creating a pressure surge. During propellant system shutdown, a pressure surge is created due to sudden closure of a valve. During both activation and shutdown, pressure surges must be predicted accurately to ensure structural integrity of the propulsion system fluid network. The method of characteristics is the most widely used method of calculating fluid transients in pipeline [ 1,2]. The method of characteristics, however, has limited applications in calculating flow distribution in complex flow circuits with phase change, heat transfer and rotational effects. A robust cryogenic propulsion system analyzer must have the capability to handle phase change, heat transfer, chemical reaction, rotational effects and fluid transients in conjunction with subsystem flow model for pumps, valves and various pipe fittings. In recent years, such a task has been undertaken at Marshall Space Flight Center with the development of the Generalized Fluid System Simulation Program (GFSSP), which is based on finite volume method in fluid network [3]. GFSSP has been extensively verified and validated by comparing its predictions with test data and other numerical methods for various applications such as internal flow of turbo-pump [4], propellant tank pressurization [5,6], chilldown of cryogenic transfer line [7] and squeeze film damper rotordynamics [8]. The purpose of the present paper is to investigate the applicability of the finite volume method to predict fluid transient in cryogenic flow 9. Approach to modeling of the fast energy discharge in cryogenic systems in the form of an electric arc CERN. Geneva 2016-01-01 Superconducting magnets are supplied with a few kA of electric current and can store a large amount of energy. Therefore, cryogenic systems which are comprised of such magnets are subject to the risk of fast energy discharge from the magnets themselves in the form of an electric arc. The arcing can be a result of failure in the insulation of an electric circuit or in the connection between the magnet and its current lead. During the discharge, energy can be partially dissipated into the cryogen and partially into the cryogenic system metallic structure. The part of the energy that is transferred to the metallic structure will strongly heat up the metal surface, which can lead to material burning. In this case, the cryogen will flow through the perforation to the insulation vacuum space, which can trigger a rapid increase in pressure in the vacuum enclosure. However, the discharged energy that has been stored in the cryogen also causes a rapid increase in cryogenic pressure. Hence, the proper estimation of the... 10. Cryogenic cooling with cryocooler on a rotating system. Oguri, S; Choi, J; Kawai, M; Tajima, O 2013-05-01 We developed a system that continuously maintains a cryocooler for long periods on a rotating table. A cryostat that holds the cryocooler is set on the table. A compressor is located on the ground and supplies high-purity (>99.999%) and high-pressure (1.7 MPa) helium gas and electricity to the cryocooler. The operation of the cryocooler and other instruments requires the development of interface components between the ground and rotating table. A combination of access holes at the center of the table and two rotary joints allows simultaneous circulation of electricity and helium gas. The developed system provides two innovative functions under the rotating condition, cooling from room temperature and the maintenance of a cold condition for long periods. We have confirmed these abilities as well as temperature stability under a condition of continuous rotation at 20 rpm. The developed system can be applied in various fields, e.g., in tests of Lorentz invariance, searches for axion, radio astronomy, and cosmology, and application of radar systems. In particular, there is a plan to use this system for a radio telescope observing cosmic microwave background radiation. 11. Cryogenic cooling with cryocooler on a rotating system Oguri, S.; Choi, J.; Kawai, M.; Tajima, O. 2013-05-01 We developed a system that continuously maintains a cryocooler for long periods on a rotating table. A cryostat that holds the cryocooler is set on the table. A compressor is located on the ground and supplies high-purity (>99.999%) and high-pressure (1.7 MPa) helium gas and electricity to the cryocooler. The operation of the cryocooler and other instruments requires the development of interface components between the ground and rotating table. A combination of access holes at the center of the table and two rotary joints allows simultaneous circulation of electricity and helium gas. The developed system provides two innovative functions under the rotating condition, cooling from room temperature and the maintenance of a cold condition for long periods. We have confirmed these abilities as well as temperature stability under a condition of continuous rotation at 20 rpm. The developed system can be applied in various fields, e.g., in tests of Lorentz invariance, searches for axion, radio astronomy, and cosmology, and application of radar systems. In particular, there is a plan to use this system for a radio telescope observing cosmic microwave background radiation. Gu, X. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Altinbas, Z. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Bruno, D. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Binello, S. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Costanzo, M. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Drees, A. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Fischer, W. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Gassner, D. M. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Hock, J. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Hock, K. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Harvey, M. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Luo, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Marusic, A. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Mi, C. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Mernick, K. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Minty, M. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Michnoff, R. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Miller, T. A. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Pikin, A. I. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Robert-Demolaize, G. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Samms, T. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Shrey, T. C. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Schoefer, V. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Tan, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Than, R. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Thieberger, P. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; White, S. M. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept. 2015-05-03 In the Relativistic Heavy Ion Collider (RHIC) 100 GeV polarized proton run in 2015, two electron lenses were used to partially compensate for the head-on beam-beam effect for the first time. Here, we describe the design of the current electron lens, detailing the hardware modifications made after the 2014 commissioning run with heavy ions. A new electron gun with 15-mm diameter cathode is characterized. The electron beam transverse profile was measured using a YAG screen and fitted with a Gaussian distribution. During operation, the overlap of the electron and proton beams was achieved using the electron backscattering detector in conjunction with an automated orbit control program. 13. Two-Phase Cryogenic Heat Exchanger for the Thermodynamic Vent System Christie, Robert J. 2011-01-01 A two-phase cryogenic heat exchanger for a thermodynamic vent system was designed and analyzed, and the predicted performance was compared with test results. A method for determining the required size of the Joule-Thomson device was also developed. Numerous sensitivity studies were performed to show that the design was robust and possessed a comfortable capacity margin. The comparison with the test results showed very similar heat extraction performance for similar inlet conditions. It was also shown that estimates for Joule- Thomson device flow rates and exit quality can vary significantly and these need to be accommodated for with a robust system design. 14. Cryogenic thermal storage system for discontinuous industrial vacuum processes Scaringella M. 2012-10-01 Full Text Available Phase Change Materials are proposed for refrigerating systems in discontinuous industrial vacuum processes where temperatures as low as −140 ÷ −100°C are necessary within time-frames representing 10÷20% of total operating time. An application is proposed for cooling systems used in a Physical Vapour Deposition (PVD apparatus. A prototype has been manufactured which couples a cryopump with a reservoir filled with MethylCycloPentane (MCP-C6H12 and a distribution line where nitrogen in the gaseous state is flowing. Preliminary tests show that temperatures of about −120°C are actually achieved within time windows compatible with PVD applications. 15. Two stage sorption type cryogenic refrigerator including heat regeneration system Jones, Jack A.; Wen, Liang-Chi; Bard, Steven 1989-01-01 A lower stage chemisorption refrigeration system physically and functionally coupled to an upper stage physical adsorption refrigeration system is disclosed. Waste heat generated by the lower stage cycle is regenerated to fuel the upper stage cycle thereby greatly improving the energy efficiency of a two-stage sorption refrigerator. The two stages are joined by disposing a first pressurization chamber providing a high pressure flow of a first refrigerant for the lower stage refrigeration cycle within a second pressurization chamber providing a high pressure flow of a second refrigerant for the upper stage refrigeration cycle. The first pressurization chamber is separated from the second pressurization chamber by a gas-gap thermal switch which at times is filled with a thermoconductive fluid to allow conduction of heat from the first pressurization chamber to the second pressurization chamber. 16. Cryogenic cooling with cryocooler on a rotating system Oguri, Shugo; Kawai, Masanori; Tajima, Osamu 2013-01-01 We developed a system that continuously maintains a cryocooler for long periods on a rotating table. A cryostat that holds the cryocooler is set on the table. A compressor is located on the ground and supplies high-purity (> 99.999%) and high-pressure (1.7 MPa) helium gas and electricity to the cryocooler. The operation of the cryocooler and other instruments requires the development of interface components between the ground and rotating table. A combination of access holes at the center of the table and two rotary joints allows simultaneous circulation of electricity and helium gas. The developed system provides two innovative functions under the rotating condition; cooling from room temperature and the maintenance of a cold condition for long periods. We have confirmed these abilities as well as temperature stability under a condition of continuous rotation at 20 revolutions per minute. The developed system can be applied in various fields; e.g., in tests of Lorentz invariance, searches for axion, radio as... 17. Commissioning results of CERN HIE-ISOLDE and INFN ALPI cryogenic control systems Inglese, V.; Pezzetti, M.; Calore, A.; Modanese, P.; Pengo, R. 2017-02-01 The cryogenic systems of both accelerators, namely HIE ISOLDE (High Intensity and Energy Isotope Separator On Line DEvice) at CERN and ALPI (Acceleratore Lineare Per Ioni) at LNL, have been refurbished. HIE ISOLDE is a major upgrade of the existing ISOLDE facilities, which required the construction of a superconducting linear accelerator consisting of six cryomodules, each containing five superconductive RF cavities and superconducting solenoids. The ALPI linear accelerator, similar to HIE ISOLDE, is located at Legnaro National Laboratories (LNL) and became operational in the early 90’s. It is composed of 74 superconducting RF cavities, assembled inside 22 cryostats. The new control systems are equipped with PLC, developed on the CERN UNICOS framework, which include Schneider and Siemens PLCs and various fieldbuses (Profibus DP and PA, WorldFIP). The control systems were developed in synergy between CERN and LNL in order to build, effectively and with an optimized use of resources, control systems allowing to enhance ease of operation, maintainability, and long-term availability. This paper describes (i) the cryogenic systems, with special focus on the design of the control systems hardware and software, (ii) the strategy adopted in order to achieve a synergic approach, and (iii) the commissioning results after the cool-down to 4.5 K of the cryomodules. 18. Dielectric properties measurement system at cryogenic temperatures and microwave frequencies Molla, J.; Ibarra, A.; Margineda, J.; Zamarro, J. M.; Hernandez, A. 1994-07-01 A system based on the resonant cavity method has been developed to measure the permittivity and loss tangent at 12-18 GHz over the temperature range 80 K to 300 K. Changes of permittivity as low as 0.01 % in the range 1 to 30, and 3 x 10{sup 6} for loss tangent values below 10{sup 2}, can be obtained without requiring temperature stability. The thermal expansion coefficient and resistivity factor of copper have been measured between 80 K and 300 K. Permittivity of sapphire and loss tangent of alumina of 99.9 % purity in the same temperature range are presented. (Author) 23 refs. 19. Six movements measurement system employed for GAIA secondary mirror positioning system vacuum tests at cryogenic temperatures Ramos Zapata, Gonzalo; Sánchez Rodríguez, Antonio; Garranzo García-Ibarrola, Daniel; Belenguer Dávila, Tomás 2008-07-01 In this work, the optical measurement system employed to evaluate the performance of a 6 degrees of freedom (dof) positioning mechanism under cryogenic conditions is explored. The mechanism, the flight model of three translations and three rotations positioning mechanism, was developed by the Spanish company SENER (for ASTRIUM) to fulfil the high performance requirements from ESA technology preparatory program for the positioning of a secondary mirror within the GAIA Astrometric Mission. Its performance has been evaluated under vacuum and temperature controlled conditions (up to a 10-6mbar and 100K) at the facilities of the Space Instrumentation Laboratory (LINES) of the Aerospace Technical Nacional Institute of Spain (INTA). After the description of the 'alignment tool' developed to compare a fixed reference with the optical signal corresponding to the movement under evaluation, the optical system that allows measuring the displacements and the rotations in the three space directions is reported on. Two similar bread-boards were defined and mounted for the measurements purpose, one containing two distancemeters, in order to measure the displacements through the corresponding axis, and an autocollimator in order to obtain the rotations on the plane whose normal vector is the axis mentioned before, and other one containing one distancemeter and one autocollimator. Both distancemeter and autocollimator measurements have been combined in order to extract the information about the accuracy of the mechanism movements as well as their repeatability under adverse environmental conditions. 20. Numerical Modeling of Pressurization of Cryogenic Propellant Tank for Integrated Vehicle Fluid System Majumdar, Alok K.; LeClair, Andre C.; Hedayat, Ali 2016-01-01 This paper presents a numerical model of pressurization of a cryogenic propellant tank for the Integrated Vehicle Fluid (IVF) system using the Generalized Fluid System Simulation Program (GFSSP). The IVF propulsion system, being developed by United Launch Alliance, uses boiloff propellants to drive thrusters for the reaction control system as well as to run internal combustion engines to develop power and drive compressors to pressurize propellant tanks. NASA Marshall Space Flight Center (MSFC) has been running tests to verify the functioning of the IVF system using a flight tank. GFSSP, a finite volume based flow network analysis software developed at MSFC, has been used to develop an integrated model of the tank and the pressurization system. This paper presents an iterative algorithm for converging the interface boundary conditions between different component models of a large system model. The model results have been compared with test data. 1. Slosh wave excitation due to cryogenic liquid reorientation in space-based propulsion system Hung, R. J.; Shyu, K. L.; Lee, C. C. 1991-01-01 The objective of the cryogenic fluid management of the spacecraft propulsion system is to develop the technology necessary for acquistion or positioning of liquid and vapor within a tank in reduced gravity to enable liquid outflow or vapor venting. In this study slosh wave excitation induced by the resettling flow field activated by 1.0 Hz medium frequency impulsive reverse gravity acceleration during the course of liquid fluid reorientation with the initiation of geyser for liquid filled levels of 30, 50, and 80 percent have been studied. Characteristics of slosh waves with various frequencies excited are discussed. 2. A transient thermal model of a neutral buoyancy cryogenic fluid delivery system Bue, Grant C.; Conger, Bruce S. A thermal-performance model is presently used to evaluate a preliminary Neutral Buoyancy Cryogenic fluid-delivery system for underwater EVA training. Attention is given to the modeling of positional transients generated from the moving of internal components, including the control of cycling artifacts, as well as to the convection and boiling characteristics of the cryofluid, 250-psi N2/O2 gas, and water contained in the tank. Two piston designs are considered according to performance criteria; temperature and heat-transfer rate profiles are presented. 3. Physics Based Model for Online Fault Detection in Autonomous Cryogenic Loading System Kashani, Ali; Devine, Ekaterina Viktorovna P; Luchinsky, Dmitry Georgievich; Smelyanskiy, Vadim; Sass, Jared P.; Brown, Barbara L.; Patterson-Hine, Ann 2013-01-01 We report the progress in the development of the chilldown model for rapid cryogenic loading system developed at KSC. The nontrivial characteristic feature of the analyzed chilldown regime is its active control by dump valves. The two-phase flow model of the chilldown is approximated as one-dimensional homogeneous fluid flow with no slip condition for the interphase velocity. The model is built using commercial SINDAFLUINT software. The results of numerical predictions are in good agreement with the experimental time traces. The obtained results pave the way to the application of the SINDAFLUINT model as a verification tool for the design and algorithm development required for autonomous loading operation. 4. Design of a cryogenic system for a 20m direct current superconducting MgB2 and YBCO power cable Cheadle, Michael J.; Bromberg, Leslie; Jiang, Xiaohua; Glowacki, Bartek; Zeng, Rong; Minervini, Joseph; Brisson, John 2014-01-01 The Massachusetts Institute of Technology, the University of Cambridge in the United Kingdom, and Tsinghua University in Beijing, China, are collaborating to design, construct, and test a 20 m, direct current, superconducting MgB2 and YBCO power cable. The cable will be installed in the State Key Laboratory of Power Systems at Tsinghua University in Beijing beginning in 2013. In a previous paper [1], the cryogenic system was briefly discussed, focusing on the cryogenic issues for the superconducting cable. The current paper provides a detailed discussion of the design, construction, and assembly of the cryogenic system and its components. The two-stage system operates at nominally 80 K and 20 K with the primary cryogen being helium gas. The secondary cryogen, liquid nitrogen, is used to cool the warm stage of binary current leads. The helium gas provides cooling to both warm and cold stages of the rigid cryostat housing the MgB2 and YBCO conductors, as well as the terminations of the superconductors at the end of the current leads. A single cryofan drives the helium gas in both stages, which are thermally isolated with a high effectiveness recuperator. Refrigeration for the helium circuit is provided by a Sumitomo RDK415 cryocooler. This paper focuses on the design, construction, and assembly of the cryostat, the recuperator, and the current leads with associated superconducting cable terminations. 5. Intelligent control of a cryogenic cooling plant based on blackboard system architecture. 2000-01-01 Intelligent system techniques have been rapidly assimilating into process control engineering, with many applications reported in the last decade. Intelligent control is bringing a new perspective as well as new challenges to process control. In this paper, a software architecture for a Blackboard for Integrated Intelligent Control Systems (BIICS) is described. The system is designed to simultaneously support multiple heterogeneous intelligent methodologies, such as neural networks. expert systems, fuzzy logic, neural networks and genetic algorithms. It will be shown how such methodologies can be readily assimilated into the software architecture. The BIICS system represents a multi-purpose platform for design and simulation of intelligent control paradigms for different kinds of processes. Currently the system utilizes intelligent control techniques (neuro-fuzzy and genetic optimization) for controlling a cryogenic plant used for superconductor testing at temperatures below 100 K. 6. Center of mass energy and system-size dependence of photon production at forward rapidity at RHIC STAR Collaboration; Abelev, Betty 2010-07-05 We present the multiplicity and pseudorapidity distributions of photons produced in Au+Au and Cu+Cu collisions at {radical}s{sub NN} = 62.4 and 200 GeV. The photons are measured in the region -3.7 < {eta} < -2.3 using the photon multiplicity detector in the STAR experiment at RHIC. The number of photons produced per average number of participating nucleon pairs increases with the beam energy and is independent of the collision centrality. For collisions with similar average numbers of participating nucleons the photon multiplicities are observed to be similar for Au+Au and Cu+Cu collisions at a given beam energy. The ratios of the number of charged particles to photons in the measured pseudorapidity range are found to be 1.4 {+-} 0.1 and 1.2 {+-} 0.1 for {radical}s{sub NN} = 62.4 GeV and 200 GeV, respectively. The energy dependence of this ratio could reflect varying contributions from baryons to charged particles, while mesons are the dominant contributors to photon production in the given kinematic region. The photon pseudorapidity distributions normalized by average number of participating nucleon pairs, when plotted as a function of {eta} - ybeam, are found to follow a longitudinal scaling independent of centrality and colliding ion species at both beam energies. 7. Adjusting the IP $\\beta$ Functions in RHIC Wittmer, W; Pilat, F; Ptitsyn, V; Van Zeijts, J 2004-01-01 The beta functions at the IP can be adjusted without perturbation of other optics functions via several approaches. In this paper we describe a scheme based on a vector knob, which assigns fixed values to the different tuning quadrupoles and scales them by a common multiplier. The values for the knob vector were calculated for a lattice without any errors using MADX. Previous studies for the LHC [1] have shown that this approach can meet the design goals. A specific feature of the RHIC lattice is the nested power supply system. To cope with the resulting problems a detailed response matrix analysis has been carried out and different sets of knobs were calculated and compared. The knobs were tested at RHIC during the 2004 run and preliminary results are discussed. Simultaneously a new approach to measure the beam sizes of both colliding beams at the IP, based on the tunability provided by the knobs, was developed and tested. 8. SUCCESSFUL BUNCHED BEAM STOCHASTIC COOLING IN RHIC. BRENNAN, J.M.; BLASKIEWICZ, M.; SEVERINO, F. 2006-06-23 We report on a successful test of bunch-beam stochastic cooling in RHIC at 100 GeV. The cooling system is designed for heavy ions but was tested in the recent RHIC run which operated only with polarized protons. To make an analog of the ion beam a special bunch was prepared with very low intensity. This bunch had {approx}1.5 x 10{sup 9} protons, while the other 100 bunches contained {approx}1.2 x 10{sup 11} protons each. With this bunch a cooling time on the order 1 hour was observed through shortening of the bunch length and increase in the peak bunch current, together with a narrowing of the spectral line width of the Scottky power at 4 GHz. The low level signal processing electronics and the isolated-frequency kicker cavities are described. 9. Study and development of a cryogenic heat exchanger for life support systems Soliman, M. M. 1973-01-01 A prototype cryogenic heat exchanger for removal of waste heat from a spacecraft environmental control life support system was developed. The heat exchanger uses the heat sink capabilities of the cryogenic propellants and, hence, can operate over all mission phases from prelaunch to orbit, to post landing, with quiescent periods during orbit. A survey of candidate warm fluids resulted in the selection of E-2, a fluorocarbon compound, because of its low freezing point and high boiling point. The final design and testing of the heat exchanger was carried out, however, using Freon-21, which is similar to E-2 except for its low boiling point. This change was motivated by the desire for cost effectiveness of the experimental program. The transient performance of the heat exchanger was demonstrated by an analog simulation of the heat sink system. Under the realistic transient heat load conditions (20 sec ramp from minimum to maximum Freon-21 inlet temperature), the control system was able to maintain the warm fluid outlet temperature within + or - 3 F. For a 20-sec ramp from 0 F to -400 F in the hydrogen inlet temperature, at maximum heat load, the warm fluid outlet temperature was maintained within + or - 7 F. 10. Thermodynamic Control System for cryogenic propellant storage : experimental and analytical performance assessment Mer, Samuel; Thibault, Jean-Paul; Corre, Christophe 2016-11-01 Future operations in space exploration require to store cryogens for long duration. Residual heat loads induce cryogenic propellant vaporization and tank self-pressurization (SP), eventually leading to storage failure for large enough mission duration. The present study focuses on the Thermodynamic Venting System (TVS) control strategy : liquid propellant is pumped from the tank, cooled down by a heat exchanger and re-injected, as a jet, inside the tank. The injection is followed by vapor condensation and liquid bath destratification due to mixing. The system cold source is created thanks to a Vented Branch where a liquid fraction is withdrawn from the tank and expanded through a Joule-Thomson valve. The vented branch vaporization permits to cool down the injection loop. Quantitative analyses of SP and TVS control have been experimentally performed using a 110 L tank and a simulant fluid. A database of accurate temperature and pressure dynamics has been gathered and used to validate a homogeneous thermodynamic model which provides a fast prediction of the tank dynamics. The analytical model has been coupled with a multi-objective optimizer to identify system components and regulation strategies that maximize the tank storage duration for various mission types. The authors acknowledge the joint support of the Centre National d'Etudes Spatiales and Air Liquide Advanced Technologies. 11. Development of a Calibration System for Cryogenic Light Detectors in CUPID Luo, Meng; Kolomensky, Yury; O'Donnell, Thomas; Schmidt, Benjamin; Cupid Collaboration 2017-01-01 If neutrino is a Majorana particle, it is possible to observe neutrinoless double beta decay (0 νββ), whose signature is a monochromatic line at the Q-value of the decay in the energy spectrum of the two electrons. Cryogenic Underground Observatory for Rare Events (CUORE) is an experiment which aims to search for 0 νββ in 130Te with TeO2 bolometers, whose background is dominated by α particles from natural radioactivity in the detector material. CUPID (CUORE Upgrade with Particle IDentification) is the next generation experiment proposed to distinguish 0 νββ events from those of α particles with Cherenkov radiation. An important part of CUPID R&D is to design, build and characterize a calibration system that can generate a known amount of light and transport that light to the dilution refrigerator at mK temperatures. We describe the design, implementation and performance of a calibration system developed for bolometric light detectors. Preparation work includes researching and selecting a light source (LED). A transport system (optical fiber) was developed to direct the light to the coldest part of the dilution refrigerator. Additionally, the light yield attenuation of optical fiber at cryogenic temperatures was measured. This project is supported by National Science Foundation and UC-Berkeley. 12. The RHIC gold rush Schaefer, T. [Department of Physics, North Carolina State University (United States) 2003-06-01 Physicists are colliding gold nuclei to recreate the fireball that existed in the very early universe, and they may have found evidence for quark-gluon plasma. What happens to ordinary matter as you heat it to higher and higher temperatures, or compress it to greater and greater densities? This simple question underpins a major effort to create extreme conditions in the lab, which has recently taken the shape of the Relativistic Heavy Ion Collider (RHIC). This machine has been colliding gold nuclei since 2000, and has produced tantalizing hints that a new state of matter - the quark-gluon plasma - is created in the reactions. But it has also sparked surprises that are sending researchers back to the drawing board. (U.K.) 13. The RHIC gold rush Schäfer, T 2003-01-01 Physicists are colliding gold nuclei to recreate the fireball that existed in the very early universe, and they may have found evidence for quark-gluon plasma. What happens to ordinary matter as you heat it to higher and higher temperatures, or compress it to greater and greater densities? This simple question underpins a major effort to create extreme conditions in the lab, which has recently taken the shape of the Relativistic Heavy Ion Collider (RHIC). This machine has been colliding gold nuclei since 2000, and has produced tantalizing hints that a new state of matter - the quark-gluon plasma - is created in the reactions. But it has also sparked surprises that are sending researchers back to the drawing board. (U.K.) 14. The Local Helium Compound Transfer Lines for the Large Hadron Collider Cryogenic System Parente, C; Munday, A; Wiggins, P 2006-01-01 The cryogenic system for the Large Hadron Collider (LHC) under construction at CERN will include twelve new local helium transfer lines distributed among five LHC points in underground caverns. These lines, being manufactured and installed by industry, will connect the cold boxes of the 4.5-K refrigerators and the 1.8-K refrigeration units to the cryogenic interconnection boxes. The lines have a maximum of 30-m length and may possess either small or large re-distribution units to allow connection to the interface ports. Due to space restrictions the lines may have complex routings and require several elbowed sections. The lines consist of a vacuum jacket, a thermal shield and either three or four helium process pipes. Specific internal and external supporting and compensation systems were designed for each line to allow for thermal contraction of the process pipes (or vacuum jacket, in case of a break in the insulation vacuum) and to minimise the forces applied to the interface equipment. Whenever possible, f... 15. Optimal design of a Thermodynamic Vent System for cryogenic propellant storage Mer, Samuel; Fernandez, David; Thibault, Jean-Paul; Corre, Christophe 2016-12-01 Future operations in space exploration require to store cryogenic liquids for long duration. Residual heat loads, due to heat conduction in the launcher structure or solar radiation, induce cryogenic propellant vaporization and tank self-pressurization. The Thermodynamic Vent System (TVS) permits to control self-pressurization using the following procedure: a fraction of liquid propellant is removed from the tank by a pump, cooled down by a heat exchanger and re-injected inside the tank as a jet or a spray. As no natural heat sink is available in space, the cold source is created by removing another fraction of liquid propellant which is expanded in a Joule-Thomson valve and vented to space. The sub-cooled injection is followed by vapor condensation and liquid bath destratification due to mixing. In this work, an optimization method is applied to an extended homogeneous thermodynamic model to design a TVS system maximizing the storage duration under various heat load and tank size assumptions. 16. Cryogenic fluid management technologies for space transportation. Zero G thermodynamic vent system 1994-01-01 Long term storage of subcritical cryogens in space must address the problem of thermal stratification in the storage tanks, liquid acquisition devices, and associated feed systems. Due to the absence of gravity induced body forces, thermal stratification in zero-g is more severe than commonly experienced in a one-g environment. If left uncontrolled, the thermal gradients result in excessive tank pressure rise and the formation of undesirable liquid/vapor mixtures within the liquid bulk, liquid acquisition system, and propellant transfer lines. Since external heat leakage cannot be eliminated, a means of minimizing the thermal stratification in the ullage gas, liquid, and feed system is required. A subsystem which minimizes the thermal stratification and rejects the environmental heat leakage in an efficient manner is therefore needed for zero-g subcritical cryogenic systems. In ground based storage systems the ullage gas location is always known (top of the tank) and therefore direct venting of gases as a means of heat rejection is easily accomplished. In contrast, because the ullage location in a zero-g environment is not easily predictable, heat rejection through direct gaseous venting is difficult in space (requires liquid settling, or surface tension devices). A means of indirect venting through the use of a thermodynamic vent system (TVS) is therefore required. A thermodynamic vent system allows indirect venting of vapor through heat exchange between the vented fluid and the stored fluid. The objective is to ensure that only gas and not liquid is vented, in order to minimize the propellant losses. Consequently, the design of a TVS is a critical enabling technology for future applications such as solar thermal and electric propulsion, single-stage-to-orbit vertical landers and upper stages, and any space based operations involving subcritical cryogenics. To bridge this technology gap NASA MSFC initiated an effort to build and verify through ground tests a zero 17. Proposal for the award of a contract for the electronic instrumentation cards and crates for the LHC cryogenic system 2005-01-01 This document concerns the award of a contract for the manufacturing, assembly, testing and delivery of the electronic instrumentation cards and crates for the LHC cryogenic system. The Finance Committee is invited to agree to the negotiation of a contract with JUTRON (FI), the lowest bidder, for the manufacture, assembly, testing and delivery of the electronic instrumentation cards and crates for the LHC cryogenic system for a total amount of 1 217 950 euros (1 875 444 Swiss francs), not subject to revision. The rate of exchange used is that stipulated in the tender. 18. Highlights from BNL-RHIC Tannenbaum, M J 2012-01-01 Recent highlights from Brookhaven National Laboratory and the Relativistic Heavy Ion Collider (RHIC) are reviewed and discussed. Topics include: Discovery of the strongly interacting Quark Gluon Plasma (sQGP) in 2005; RHIC machine operation in 2011 as well as latest achievements from the superconducting Magnet Division and the National Synchrotron Light Source II project. Highlights from QGP physics at RHIC include: comparison of new measurements of charged multiplicity in A+A collisions by ALICE at the LHC to previous RHIC measurements; Observation of the anti-alpha particle by the STAR experiment; Collective Flow, including the Triangular Flow discovery and the latest results on v3; the RHIC beam energy scan in search of the QCD critical point. The pioneering use at RHIC of hard-scattering as a probe of the sQGP will also be reviewed and the latest results presented including: jet-quenching via suppression of high pT particles and two particle correlations; new results on fragmentation functions using gamma... 19. RHIC spin flipper AC dipole controller Oddo, P.; Bai, M.; Dawson, C.; Gassner, D.; Harvey, M.; Hayes, T.; Mernick, K.; Minty, M.; Roser, T.; Severino, F.; Smith, K. 2011-03-28 The RHIC Spin Flipper's five high-Q AC dipoles which are driven by a swept frequency waveform require precise control of phase and amplitude during the sweep. This control is achieved using FPGA based feedback controllers. Multiple feedback loops are used to and dynamically tune the magnets. The current implementation and results will be presented. Work on a new spin flipper for RHIC (Relativistic Heavy Ion Collider) incorporating multiple dynamically tuned high-Q AC-dipoles has been developed for RHIC spin-physics experiments. A spin flipper is needed to cancel systematic errors by reversing the spin direction of the two colliding beams multiple times during a store. The spin flipper system consists of four DC-dipole magnets (spin rotators) and five AC-dipole magnets. Multiple AC-dipoles are needed to localize the driven coherent betatron oscillation inside the spin flipper. Operationally the AC-dipoles form two swept frequency bumps that minimize the effect of the AC-dipole dipoles outside of the spin flipper. Both AC bumps operate at the same frequency, but are phase shifted from each other. The AC-dipoles therefore require precise control over amplitude and phase making the implementation of the AC-dipole controller the central challenge. 20. Highlights from BNL and RHIC 2014 Tannenbaum, M J 2015-01-01 Highlights of news from Brookhaven National Laboratory (BNL) and results from the Relativistic Heavy Ion Collider (RHIC) in the period July 2013-June 2014 are presented. It was a busy year for news, most notably a U. S. Government shutdown for 16 days beginning October 1, 2013 due to the lack of an approved budget for FY2014. Even with this unusual government activity, the $\\sqrt{s_{NN}}=200$ GeV Au+Au Run14 at RHIC was the best ever with integrated luminosity exceeding the sum of all previous runs. Additionally there was a brief He$^3$+Au run to continue the study of collective flow in small systems which was reinforced by new results presented on identified particle flow in d+Au. The other scientific highlights are also mostly concerned with soft (low $p_T$)'' physics complemented by the first preliminary results of reconstructed jets from hard-scattered partons in Au+Au collisions at RHIC . The measurements of transverse energy ($E_T$) spectra in p-p, d+Au and Au+Au collisions, which demonstrated last ye... 1. Thermal conductance modeling and characterization of the SuperCDMS-SNOLAB sub-Kelvin cryogenic system Dhuley, R. C. [Fermilab; Hollister, M. I. [Fermilab; Ruschman, M. K. [Fermilab; Martin, L. D. [Fermilab; Schmitt, R. L. [Fermilab; Tatkowski, Tatkowski,G.L. [Fermilab; Bauer, D. a. [Fermilab; Lukens, P. T. [Fermilab 2017-09-13 The detectors of the Super Cryogenic Dark Matter Search experiment at SNOLAB (SuperCDMS SNOLAB) will operate in a seven-layered cryostat with thermal stages between room temperature and the base temperature of 15 mK. The inner three layers of the cryostat, which are to be nominally maintained at 1 K, 250 mK, and 15 mK, will be cooled by a dilution refrigerator via conduction through long copper stems. Bolted and mechanically pressed contacts, at and cylindrical, as well as exible straps are the essential stem components that will facilitate assembly/dismantling of the cryostat. These will also allow for thermal contractions/movements during cooldown of the sub-Kelvin system. To ensure that these components and their contacts meet their design thermal conductance, prototypes were fabricated and cryogenically tested. The present paper gives an overview of the SuperCDMS SNOLAB sub-Kelvin architecture and its conductance requirements. Results from the conductance measurements tests and from sub-Kelvin thermal modeling are discussed. 2. A cryogen-free ultralow-field superconducting quantum interference device magnetic resonance imaging system Eom, Byeong Ho; Penanen, Konstantin; Hahn, Inseob, E-mail: ihahn@caltech.edu [Jet Propulsion Laboratory, California Institute of Technology, Pasadena, California 91109 (United States) 2014-09-15 Magnetic resonance imaging (MRI) at microtesla fields using superconducting quantum interference device (SQUID) detection has previously been demonstrated, and advantages have been noted. Although the ultralow-field SQUID MRI technique would not need the heavy superconducting magnet of conventional MRI systems, liquid helium required to cool the low-temperature detector still places a significant burden on its operation. We have built a prototype cryocooler-based SQUID MRI system that does not require a cryogen. The SQUID detector and the superconducting gradiometer were cooled down to 3.7 K and 4.3 K, respectively. We describe the prototype design, characterization, a phantom image, and areas of further improvements needed to bring the imaging performance to parity with conventional MRI systems. 3. A cryogen-free ultralow-field superconducting quantum interference device magnetic resonance imaging system Eom, Byeong Ho; Penanen, Konstantin; Hahn, Inseob 2014-09-01 Magnetic resonance imaging (MRI) at microtesla fields using superconducting quantum interference device (SQUID) detection has previously been demonstrated, and advantages have been noted. Although the ultralow-field SQUID MRI technique would not need the heavy superconducting magnet of conventional MRI systems, liquid helium required to cool the low-temperature detector still places a significant burden on its operation. We have built a prototype cryocooler-based SQUID MRI system that does not require a cryogen. The SQUID detector and the superconducting gradiometer were cooled down to 3.7 K and 4.3 K, respectively. We describe the prototype design, characterization, a phantom image, and areas of further improvements needed to bring the imaging performance to parity with conventional MRI systems. 4. Helium Recovery in the LHC Cryogenic System following Magnet Resistive Transitions Chorowski, M; Serio, L; Tavian, L; Wagner, U; Van Weelderen, R 1998-01-01 A resistive transition (quench) of the Large Hadron Collider magnets provokes the expulsion of helium from the magnet cryostats to the helium recovery system. A high-volume, vacuum-insulated recovery line connected to several uninsulated medium-pressure gas storage tanks, forms the main constituents of the system. Besides a dedicated hardware configuration, helium recovery also implies specific procedures that should follow a quench, in order to conserve the discharged helium and possibly make use of its refrigeration capability. The amount of energy transferred after a quench from the magnets to the helium leaving the cold mass has been estimated on the basis of experimental data. Based on these data, the helium thermodynamic state in the recovery system is calculated using a lumped parameter approach. The LHC magnet quenches are classified ina parametric way from their cryogenic consequences and procedures that should follow the quench are proposed. 5. Model Predictions and Observed Performance of JWST's Cryogenic Position Metrology System Lunt, Sharon R.; Rhodes, David; DiAntonio, Andrew; Boland, John; Wells, Conrad; Gigliotti, Trevis; Johanning, Gary 2016-01-01 The James Webb Space Telescope cryogenic testing requires measurement systems that both obtain a very high degree of accuracy and can function in that environment. Close-range photogrammetry was identified as meeting those criteria. Testing the capability of a close-range photogrammetric system prior to its existence is a challenging problem. Computer simulation was chosen over building a scaled mock-up to allow for increased flexibility in testing various configurations. Extensive validation work was done to ensure that the actual as-built system meet accuracy and repeatability requirements. The simulated image data predicted the uncertainty in measurement to be within specification and this prediction was borne out experimentally. Uncertainty at all levels was verified experimentally to be less than 0.1 millimeters. 6. Cryogen-free lkA-class Ic measurement system featuring an 8 T HTS magnet Strickland, N. M.; Hoffmann, C.; Wimbush, S. C.; Pooke, D. M.; Huang, T.; Lazic, Z.; Chamritski, V.; Talantsev, E. F.; Long, N. J.; Tallon, J. L. 2014-05-01 We have developed a cryogen-free critical-current (Ic) measuring system comprising a conduction-cooled 8 T HTS magnet and convection-cooled sample, both cooled by commercial cryocoolers. The sample can be rotated and transport currents of up to 800 A delivered with less than 0.5 K temperature rise during the Ic measurement. The system is automated with respect to variations in temperature (30-90 K), field (0-8 T), and field angle (0-360°). We have used this system to measure HTS wire samples, concentrating on metal-organic deposited YBCO on RABiTS substrates. Particular emphasis is given to the evolution of Ic anisotropy with temperature, and the dangers of extrapolating from 77 K to 30 K. 7. Model predictions and observed performance of JWST's cryogenic position metrology system Lunt, Sharon R.; Rhodes, David; DiAntonio, Andrew; Boland, John; Wells, Conrad; Gigliotti, Trevis; Johanning, Gary 2016-07-01 The James Webb Space Telescope (JWST) cryogenic testing requires measurement systems that both obtain a very high degree of accuracy and can function in that environment. Close-range photogrammetry was identified as meeting those criteria. Testing the capability of a close-range photogrammetric system prior to its existence is a challenging problem. Computer simulation was chosen over building a scaled mock-up to allow for increased flexibility in testing various configurations. Extensive validation work was done to ensure that the actual as-built system meets accuracy and repeatability requirements. The simulated image data predicted the uncertainty in measurement to be within specification and this prediction was borne out experimentally. Uncertainty at all levels was verified experimentally to be <0.1 mm. 8. Helium cryogenics Van Sciver, Steven W 2012-01-01 Twenty five years have elapsed since the original publication of Helium Cryogenics. During this time, a considerable amount of research and development involving helium fluids has been carried out culminating in several large-scale projects. Furthermore, the field has matured through these efforts so that there is now a broad engineering base to assist the development of future projects. Helium Cryogenics, 2nd edition brings these advances in helium cryogenics together in an updated form. As in the original edition, the author's approach is to survey the field of cryogenics with emphasis on helium fluids. This approach is more specialized and fundamental than that contained in other cryogenics books, which treat the associated range of cryogenic fluids. As a result, the level of treatment is more advanced and assumes a certain knowledge of fundamental engineering and physics principles, including some quantum mechanics. The goal throughout the work is to bridge the gap between the physics and engineering aspe... 9. Design, construction and cooling system performance of a prototype cryogenic stopping cell for the Super-FRS at FAIR Ranjan, M. [KVI-Center for Advanced Radiation Technology, University of Groningen - Zernikelaan 25, 9747 AA Groningen (Netherlands); Dendooven, P., E-mail: p.g.dendooven@rug.nl [KVI-Center for Advanced Radiation Technology, University of Groningen - Zernikelaan 25, 9747 AA Groningen (Netherlands); Purushothaman, S. [GSI Helmholtz Centre for Heavy Ion Research - Planckstraße 1, 64291 Darmstadt (Germany); Dickel, T. [GSI Helmholtz Centre for Heavy Ion Research - Planckstraße 1, 64291 Darmstadt (Germany); II. Physikalisches Institut, Justus-Liebig-Universität Gießen - Heinrich-Buff-Ring 16, 35392 Gießen (Germany); Reiter, M.P. [II. Physikalisches Institut, Justus-Liebig-Universität Gießen - Heinrich-Buff-Ring 16, 35392 Gießen (Germany); Ayet, S. [GSI Helmholtz Centre for Heavy Ion Research - Planckstraße 1, 64291 Darmstadt (Germany); Haettner, E. [GSI Helmholtz Centre for Heavy Ion Research - Planckstraße 1, 64291 Darmstadt (Germany); II. Physikalisches Institut, Justus-Liebig-Universität Gießen - Heinrich-Buff-Ring 16, 35392 Gießen (Germany); Moore, I.D. [University of Jyväskylä - FI-40014, Jyväskylä (Finland); Kalantar-Nayestanaki, N. [KVI-Center for Advanced Radiation Technology, University of Groningen - Zernikelaan 25, 9747 AA Groningen (Netherlands); and others 2015-01-11 A cryogenic stopping cell for stopping energetic radioactive ions and extracting them as a low energy beam was developed. This first ever cryogenically operated stopping cell serves as prototype device for the Low-Energy Branch of the Super-FRS at FAIR. The cell has a stopping volume that is 1 m long and 25 cm in diameter. Ions are guided by a DC field along the length of the stopping cell and by a combined RF and DC fields provided by an RF carpet at the exit-hole side. The ultra-high purity of the stopping gas required for optimum ion survival is reached by cryogenic operation. The design considerations and construction of the cryogenic stopping cell, as well as some performance characteristics, are described in detail. Special attention is given to the cryogenic aspects in the design and construction of the stopping cell and the cryocooler-based cooling system. The cooling system allows the operation of the stopping cell at any desired temperature between about 70 K and room temperature. The cooling system performance in realistic on-line conditions at the FRS Ion Catcher Facility at GSI is discussed. A temperature of 110 K at which efficient ion survival was observed is obtained after 10 h of cooling. A minimum temperature of the stopping gas of 72 K was reached. The expertise gained from the design, construction and performance of the prototype cryogenic stopping cell has allowed the development of a final version for the Low-Energy Branch of the Super-FRS to proceed. 10. Cryogenic exciter Bray, James William [Niskayuna, NY; Garces, Luis Jose [Niskayuna, NY 2012-03-13 The disclosed technology is a cryogenic static exciter. The cryogenic static exciter is connected to a synchronous electric machine that has a field winding. The synchronous electric machine is cooled via a refrigerator or cryogen like liquid nitrogen. The static exciter is in communication with the field winding and is operating at ambient temperature. The static exciter receives cooling from a refrigerator or cryogen source, which may also service the synchronous machine, to selected areas of the static exciter and the cooling selectively reduces the operating temperature of the selected areas of the static exciter. 11. The cryogenic system for the superconducting e{sup +}e{sup -} linear collider TESLA Horlitz, G. [Deutsches Elektronen-Synchrotron (DESY), Hamburg (Germany) 1996-09-01 The superconducting electron/positron collider TESLA (center of mass energy 500 GeV) requires a cryogenic supply system of total estimated cooling capacities of 33 kW at 2.0 K, 36 kW at 4.5 K, 243 kW at 40/80 K and current lead cooling flow rate (liquefaction power) of 0.2 kg/s. The system is spread over a linear range of about 30 km. A new layout is presented in this paper (reduction of HF - pulse rate from 10/s to 5/s results in reduced heat loads, lower numbers of cryo halls with refrigerators and increased subunit lengths). (author) 12. Commissioning of water detritiation and cryogenic distillation systems at TLK in view of ITER design Cristescu, I. [Forschungszentrum Karlsruhe, 76021 Karlsruhe (Germany)], E-mail: ion.cristescu@itp.fzk.de; Cristescu, Ioana R.; Doerr, L.; Glugla, M.; Hellriegel, G.; Michling, R. [Forschungszentrum Karlsruhe, 76021 Karlsruhe (Germany); Murdoch, D. [EFDA Close Support Unit, Boltzmannstrasse 2, D-85748 Garching (Germany); Schaefer, P.; Welte, S.; Wurster, W. [Forschungszentrum Karlsruhe, 76021 Karlsruhe (Germany) 2007-10-15 The ITER Isotope Separation System (ISS) and Water Detritiation System (WDS) will be integrated in order to reduce potential chronic tritium releases from the ISS by routing the top (protium) product from the ISS into the Liquid Phase Catalytic Exchange (LPCE) column of WDS. This provides an additional barrier against ISS tritium releases and should mitigate the memory effects due to process parameter fluctuations in the ISS. To support the research activities needed to characterize the performances of various components for WDS and ISS processes in various working conditions and configurations as needed for ITER design, an experimental facility called TRENTA and representative of the ITER WDS and ISS protium separation column has been commissioned at Tritium Laboratory Karlsruhe (TLK). The TRENTA facility consists of Combined Electrolysis Catalytic Exchange (CECE) process, with an LPCE column of 8 m, in combination with a cryogenic distillation (CD) process. The processes description and the status of commissioning of TRENTA facility is presented. 13. FRIB Cryogenic Plant Status Dixon, Kelly D. [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Ganni, Venkatarao [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Knudsen, Peter N. [Thomas Jefferson National Accelerator Facility (TJNAF), Newport News, VA (United States); Casagranda, Fabio [Michigan State Univ., East Lansing, MI (United States) 2015-12-01 After practical changes were approved to the initial conceptual design of the cryogenic system for MSU FRIB and an agreement was made with JLab in 2012 to lead the design effort of the cryogenic plant, many activities are in place leading toward a cool-down of the linacs prior to 2018. This is mostly due to using similar equipment used at CHLII for the 12 GeV upgrade at JLab and an aggressive schedule maintained by the MSU Conventional Facilities department. Reported here is an updated status of the cryogenic plant, including the equipment procurement status, plant layout, facility equipment and project schedule. Barron, R. F. During the past 50 years, the use of digital computers has significantly influenced the design and analysis of cryogenic systems. At the time when the first Cryogenic Engineering Conference was held, thermodynamic data were presented in graphical or tabular form (the "steam table" format), whereas thermodynamic data for cryogenic system design is computer generated today. The thermal analysis of cryogenic systems in the 1950s involved analytical solutions, graphical solutions, and relatively simple finite-difference approaches. These approaches have been supplanted by finite-element numerical programs which readily solve complicated thermal problems that could not be solved easily using the methods of the 1950s. In distillation column design, the use of the McCabe-Thiele graphical method for determination of the number of theoretical plates has been replaced by numerical methods that allow consideration of several different components in the feed and product streams. 15. Cryogenic Cooling System for 5 kA, 200 μH Class HTS DC Reactor Park, Heecheol; Kim, Seokho; Kim, Kwangmin; Park, Minwon; Park, Taejun; Kim, A.-rong; Lee, Sangjin DC reactors, made by aluminum busbar, are used to stabilize the arc of an electric furnace. In the conventional arc furnace, the transport current is several tens of kilo-amperes and enormous resistive loss is generated. To reduce the resistive loss at the DC reactor, a HTS DC reactor can be considered. It can dramatically improve the electric efficiency as well as reduce the installation space. Similar with other superconducting devices, the HTS DC reactor requires current leads from a power source in room temperature to the HTS coil in cryogenic environment. The heat loss at the metal current leads can be minimized through optimization process considering the geometry and the transport current. However, the transport current of the HTS DC reactor for the arc furnace is much larger than most of HTS magnets and the enormous heat penetration through the current lead should be effectively removed to keep the temperature around 70∼77 K. Current leads are cooled down by circulation of liquid nitrogen from the cooling system with a stirling cryocooler. The operating temperature of HTS coil is 30∼40 K and circulation of gaseous helium is used to remove the heat generation at the HTS coil. Gaseous helium is transported through the cryogenic helium blower and a single stage GM cryocooler. This paper describes design and experimental results on the cooling system for current leads and the HTS coil of 5 kA, 200 μH class DC reactor as a prototype. The results are used to verify the design values of the cooling systems and it will be applied to the design of scale-up cooling system for 50 kA, 200 μH class DC reactor. 16. A LOW NOISE RF SOURCE FOR RHIC. HAYES,T. 2004-07-05 The Relativistic Heavy Ion Collider (RHIC) requires a low noise rf source to ensure that beam lifetime during a store is not limited by the rf system. The beam is particularly sensitive to noise from power line harmonics. Additionally, the rf source must be flexible enough to handle the frequency jump required for rebucketing (transferring bunches from the acceleration to the storage rf systems). This paper will describe the design of a Direct Digital Synthesizer (DDS) based system that provides both the noise performance and the flexibility required. 17. Cryogenics for LHC experiments 2001-01-01 Cryogenic systems will be used by LHC experiments to maximize their performance. Institutes around the world are collaborating with CERN in the construction of these very low temperature systems. The cryogenic test facility in hall 180 for ATLAS magnets. High Energy Physics experiments have frequently adopted cryogenic versions of their apparatus to achieve optimal performance, and those for the LHC will be no exception. The two largest experiments for CERN's new flagship accelerator, ATLAS and CMS, will both use large superconducting magnets operated at 4.5 Kelvin - almost 270 degrees below the freezing point of water. ATLAS also includes calorimeters filled with liquid argon at 87 Kelvin. For the magnets, the choice of a cryogenic version was dictated by a combination economy and transparency to emerging particles. For the calorimeters, liquid argon was selected as the fluid best suited to the experiment's physics requirements. High Energy Physics experiments are the result of worldwide collaborations and... 18. Improvement plans for the RHIC/AGS on-line model environments Brown,K.A.; Ahrens, L.; Beebe-Wang, J.; Morris, J.; Nemesure, S.; Robert-Demolaize, G.; Satogata, T.; Schoefer, V.; Tepikian, S. 2009-08-31 The on-line models for Relativistic Ion Collider (RHIC) and the RHIC pre-injectors (the AGS and the AGS Booster) can be thought of as containing our best collective knowledge of these accelerators. As we improve these on-line models we are building the framework to have a sophisticated model-based controls system. Currently the RHIC on-line model is an integral part of the controls system, providing the interface for tune control, chromaticity control, and non-linear chromaticity control. What we discuss in this paper is our vision of the future of the on-line model environment for RHIC and the RHIC preinjectors. Although these on-line models are primarily used as Courant-Snyder parameter calculators using live machine settings, we envision expanding these environments to encompass many other problem domains. 19. RESEARCH PLAN FOR SPIN PHYSICS AT RHIC. AIDALA, C.; BUNCE, G.; ET AL. 2005-02-01 In this report we present the research plan for the RHIC spin program. The report covers (1) the science of the RHIC spin program in a world-wide context; (2) the collider performance requirements for the RHIC spin program; (3) the detector upgrades required, including timelines; (4) time evolution of the spin program. 20. Electron Cooling of RHIC Ben-Zvi, Ilan; Barton, Donald; Beavis, Dana; Blaskiewicz, Michael; Bluem, Hans; Brennan, Joseph M; Bruhwiler, David L; Burger, Al; Burov, Alexey; Burrill, Andrew; Calaga, Rama; Cameron, Peter; Chang, Xiangyun; Cole, Michael; Connolly, Roger; Delayen, Jean R; Derbenev, Yaroslav S; Eidelman, Yury I; Favale, Anthony; Fedotov, Alexei V; Fischer, Wolfram; Funk, L W; Gassner, David M; Hahn, Harald; Harrison, Michael; Hershcovitch, Ady; Holmes, Douglas; Hseuh Hsiao Chaun; Johnson, Peter; Kayran, Dmitry; Kewisch, Jorg; Kneisel, Peter; Koop, Ivan; Lambiase, Robert; Litvinenko, Vladimir N; MacKay, William W; Mahler, George; Malitsky, Nikolay; McIntyre, Gary; Meng, Wuzheng; Merminga, Lia; Meshkov, Igor; Mirabella, Kerry; Montag, Christoph; Nagaitsev, Sergei; Nehring, Thomas; Nicoletti, Tony; Oerter, Brian; Parkhomchuk, Vasily; Parzen, George; Pate, David; Phillips, Larry; Preble, Joseph P; Rank, Jim; Rao, Triveni; Rathke, John; Roser, Thomas; Russo, Thomas; Scaduto, Joseph; Schultheiss, Tom; Sekutowicz, Jacek; Shatunov, Yuri; Sidorin, Anatoly O; Skrinsky, Aleksander Nikolayevich; Smirnov, Alexander V; Smith, Kevin T; Todd, Alan M M; Trbojevic, Dejan; Troubnikov, Grigory; Wang, Gang; Wei, Jie; Williams, Neville; Wu, Kuo-Chen; Yakimenko, Vitaly; Zaltsman, Alex; Zhao, Yongxiang; ain, Animesh K 2005-01-01 We report progress on the R&D program for electron-cooling of the Relativistic Heavy Ion Collider (RHIC). This electron cooler is designed to cool 100 GeV/nucleon at storage energy using 54 MeV electrons. The electron source will be a superconducting RF photocathode gun. The accelerator will be a superconducting energy recovery linac. The frequency of the accelerator is set at 703.75 MHz. The maximum electron bunch frequency is 9.38 MHz, with bunch charge of 20 nC. The R&D program has the following components: The photoinjector and its photocathode, the superconducting linac cavity, start-to-end beam dynamics with magnetized electrons, electron cooling calculations including benchmarking experiments and development of a large superconducting solenoid. The photoinjector and linac cavity are being incorporated into an energy recovery linac aimed at demonstrating ampere class current at about 20 MeV. A Zeroth Order Design Report is in an advanced draft state, and can be found on the web at http://www.ags... 1. 2nd Biennial Conference on Refrigeration for Cryogenic Sensors and Electronic Systems 1983-01-01 This proceedings documents the output of the Second Biennial Conference on Refrigeration for Cryogenic Sensors and Electronic Systems held at the National Aeronautics and Space Administration's Goddard Space Flight Center, Greenbelt, Maryland, on December 7-8, 1982. Building on the first open meeting hosted by the National Bureau of Standards in 1980, the focus of this second meeting was again on low-temperature, closed-cycle cooler technology. However, higher temperature coolers (77 K), with technology applicable to the low temperature coolers, were considered to be within the scope of this meeting. This second conference consisted of 30 papers presented by representatives of industry, government, and academia. The conference proceedings reproduced here was published by the NASA Goddard Space Flight Center in Greenbelt Maryland as NASA Conference Publication 2287. 2. Cryogenic system configuration for the International Linear Collider (ILC) at mountainous site Nakai, H.; Okamura, T.; Delikaris, D.; Peterson, T.; Yamamoto, A. 2017-02-01 The International Linear Collider (ILC) plans to make use of ten cryoplants for its main linacs, each providing 19 kW at 4.5 K equivalent and among of it 3.6 kW at 2 K. Each cryoplant will consist of various cryogenic components such as a 4.5 K refrigerator cold box, a 2 K refrigerator cold box, and helium compressors and so on. In the technical design report (TDR) of the ILC, due to the mountainous topology, almost all cryogenic components would be installed in underground cryogenic caverns next to the main linac tunnels and only cooling towers on surface area. However, we would like to find a more effective and sophisticated configuration of the cryoplant components (cryogenic configuration). Under several constraints of technical, geographical, and environmental points of view, the cryogenic configuration should be considered carefully to satisfy such various conditions. After discussions on this topic conducted at various workshops and conferences, an updated cryogenic configuration is suggested. The proposed updated configuration may affect the total construction cost of the ILC and the entire structure of the ILC conventional facilities. The updated cryogenic configuration is presented and the on-going discussions with the conventional facilities and siting (CFS) colleagues for further improvement of the cryogenic configuration is introduced. 3. Strangeness in STAR at RHIC , 2016-01-01 We present the recent results of strangeness production at the mid-rapidity in Au + Au collisions at RHIC, from $\\sqrt{s_{\\rm NN}}$ = 7.7 to 200 GeV. The $v_2$ of multi-strange baryon $\\Omega$ and $\\phi$ mesons are similar to that of pions and protons in the intermediate $p_T$ range (2 - 5 GeV/$c$) in $\\sqrt{s_{\\rm NN}}$ = 200 GeV Au + Au collisions, indicating that the major part of collective flow has been built up at partonic stage. The breaking of mass ordering between $\\phi$ mesons and protons in the low $p_T$ range ($<$ 1 GeV/$c$) is consistent with a picture that $\\phi$ mesons are less sensitive to later hadronic interaction. The nuclear modification factor $R_{\\rm CP}$ and baryon to meson ratio change dramatically when the collision energy is lower than 19.6 GeV. It suggests a possible change of medium property of the system compared to those from high energies. 4. Simulations of silicon vertex tracker for star experiment at RHIC Odyniec, G.; Cebra, D.; Christie, W.; Naudet, C.; Schroeder, L.; Wilson, W. [Lawrence Berkeley Lab., CA (United States); Liko, D. [Institut fur Hochenenergiephysik, Vienna, (Austria); Cramer, J.; Prindle, D.; Trainor, T. [Univ. of Washington, Seattle (United States); Braithwaite, W. [Univ. of Arkansas, Little Rock (United States) 1991-12-31 The first computer simulations to optimize the Silicon Vertex Tracker (SVT) designed for the STAR experiment at RHIC are presented. The physics goals and the expected complexity of the events at RHIC dictate the design of a tracking system for the STAR experiment. The proposed tracking system will consist of a silicon vertex tracker (SVT) to locate the primary interaction and secondary decay vertices and to improve the momentum resolution, and a time projection chamber (TPC), positioned inside a solenoidal magnet, for continuous tracking. 5. Simulations of silicon vertex tracker for star experiment at RHIC Odyniec, G.; Cebra, D.; Christie, W.; Naudet, C.; Schroeder, L.; Wilson, W. [Lawrence Berkeley Lab., CA (United States); Liko, D. [Institut fur Hochenenergiephysik, Vienna, (Austria); Cramer, J.; Prindle, D.; Trainor, T. [Univ. of Washington, Seattle (United States); Braithwaite, W. [Univ. of Arkansas, Little Rock (United States) 1991-12-31 The first computer simulations to optimize the Silicon Vertex Tracker (SVT) designed for the STAR experiment at RHIC are presented. The physics goals and the expected complexity of the events at RHIC dictate the design of a tracking system for the STAR experiment. The proposed tracking system will consist of a silicon vertex tracker (SVT) to locate the primary interaction and secondary decay vertices and to improve the momentum resolution, and a time projection chamber (TPC), positioned inside a solenoidal magnet, for continuous tracking. 6. Comparison of reusable insulation systems for cryogenically-tanked earth-based space vehicles Sumner, I. E.; Barber, J. R. 1978-01-01 Three reusable insulation systems concepts have been developed for use with cryogenic tanks of earth-based space vehicles. Two concepts utilized double-goldized Kapton (DGK) or double-aluminized Mylar (DAM) multilayer insulation (MLI), while the third utilized a hollow-glass-microsphere, load-bearing insulation (LBI). All three insulation systems have recently undergone experimental testing and evaluation under NASA-sponsored programs. Thermal performance measurements were made under space-hold (vacuum) conditions for insulation warm boundary temperatures of approximately 291 K. The resulting effective thermal conductivity was approximately .00008 W/m-K for the MLI systems (liquid hydrogen test results) and .00054 W/m-K for the LBI system (liquid nitrogen test results corrected to liquid hydrogen temperature). The DGK MLI system experienced a maximum thermal degradation of 38 percent, the DAM MLI system 14 percent, and the LBI system 6.7 percent due to repeated thermal cycling representing typical space flight conditions. Repeated exposure of the DAM MLI system to a high humidity environment for periods as long as 8 weeks provided a maximum degradation of only 24 percent. 7. Polarization response of RHIC electron lens lattices V. H. Ranjbar 2016-10-01 Full Text Available Depolarization response for a system of two orthogonal snakes at irrational tunes is studied in depth using lattice independent spin integration. In particular we consider the effect of overlapping spin resonances in this system, to understand the impact of phase, tune, relative location and threshold strengths of the spin resonances. These results are benchmarked and compared to two dimensional direct tracking results for the RHIC e-lens lattice and the standard lattice. Finally we consider the effect of longitudinal motion via chromatic scans using direct six dimensional lattice tracking. 8. Cryogenic system for the ArTeMiS large sub millimeter camera Ercolani, E.; Relland, J.; Clerc, L.; Duband, L.; Jourdan, T.; Talvard, M.; Le Pennec, J.; Martignac, J.; Visticot, F. 2014-07-01 A new photonic camera has been developed in the framework of the ArTéMis project (Bolometers architecture for large field of view ground based telescopes in the sub-millimeter). This camera scans the sky in the sub-millimeter range at simultaneously three different wavelengths, namely 200 μm, 350 μm, 450 μm, and is installed inside the APEX telescope located at 5100m above sea level in Chile. Bolometric detectors cooled to 300 mK are used in the camera, which is integrated in an original cryostat developed at the low temperature laboratory (SBT) of the INAC institut. This cryostat contains filters, optics, mirrors and detectors which have to be implemented according to mass, size and stiffness requirements. As a result the cryostat exhibits an unusual geometry. The inner structure of the cryostat is a 40 K plate which acts as an optical bench and is bound to the external vessel through two hexapods, one fixed and the other one mobile thanks to a ball bearing. Once the cryostat is cold, this characteristic enabled all the different elements to be aligned with the optical axis. The cryogenic chain is built around a pulse tube cooler (40 K and 4 K) coupled to a double stage helium sorption cooler (300 mK). The cryogenic and vacuum processes are managed by a Siemens PLC and all the data are showed and stored on a CEA SCADA system. This paper describes the mechanical and thermal design of the cryostat, its command control, and the first thermal laboratory tests. This work was carried out in collaboration with the Astrophysics laboratory SAp of the IRFU institut. SAp and SBT have installed the camera in July 2013 inside the Cassegrain cabin of APEX. 9. The 5.8 T Cryogen-Free Gyrotron Superconducting Magnet System on HL-2A Xia, Donghui; Huang, Mei; Zhou, Jun; Bai, Xingyu; Zheng, Tieliu; Rao, Jun; Zhuang, Ge 2014-04-01 A 5.8 T cryogen-free superconducting magnet (SCM) system with a warm bore hole of 160 mm in diameter, used for gyrotrons operating in the frequency range from 68 GHz to 140 GHz, is installed on the site of the HL-2A tokamak. The SCM consists of two separate solenoidal magnetic coils connected in series, a 4.2 K Gifford-McMahon (GM) refrigerator, a compressor, a coil power supply and two temperature monitors. The performance, test and preliminary experimental results of this SCM system are described in this paper. The magnetic field distribution was measured along the axis, and a dummy tube was used for adjusting the magnet system. Finally, the magnet was used for the operation of a 68 GHz/500 kW gyrotron, which is part of an electron cyclotron resonance heating (ECRH) system. With an additional auxiliary coil and after adjusting the magnet system, a maximum output power for the ECRH system of up to 400 kW was achieved. 10. Development of NEG Coating for RHIC Experimental Beamtubes Weiss, Daniel; Hseuh Hsiao Chaun; Todd, Robert J 2005-01-01 As RHIC beam intensity increases beyond original scope, pressure rises in some regions have been observed. The luminosity limiting pressure rises are associated with electron multi-pacting, electron stimulated desorption and beam induced desorption. Non-Evaporable Getter (NEG) coated beampipes have been proven effective to suppress pressure rise in synchrotron radiation facilities. Standard beampipes have been NEG coated by a vendor and added to many RHIC UHV regions. BNL is developing a cylindrical magnetron sputtering system to NEG coat special beryllium beampipes installed in RHIC experimental regions. It features a hollow, liquid cooled cathode producing power density of 500W/m and deposition rate of 5000 Angstrom/hr on 7.5cm OD beampipe. The cathode, a titanium tube partially covered with zirconium and vanadium ribbons, is oriented for horizontal coating of 4m long chambers. Ribbons and magnets are arranged to provide uniform sputtering distribution and deposited NEG composition. Vacuum performance of NE... 11. POLARIZED NEUTRONS IN RHIC COURANT,E.D. 1998-04-27 There does not appear to be any obvious way to accelerate neutrons, polarized or otherwise, to high energies by themselves. To investigate the behavior of polarized neutrons the authors therefore have to obtain them by accelerating them as components of heavier nuclei, and then sorting out the contribution of the neutrons in the analysis of the reactions produced by the heavy ion beams. The best neutron carriers for this purpose are probably {sup 3}He nuclei and deuterons. A polarized deuteron is primarily a combination of a proton and a neutron with their spins pointing in the same direction; in the {sup 3}He nucleus the spins of the two protons are opposite and the net spin (and magnetic moment) is almost the same as that of a free neutron. Polarized ions other than protons may be accelerated, stored and collided in a ring such as RHIC provided the techniques proposed for polarized proton operation can be adapted (or replaced by other strategies) for these ions. To accelerate polarized particles in a ring, one must make provisions for overcoming the depolarizing resonances that occur at certain energies. These resonances arise when the spin tune (ratio of spin precession frequency to orbit frequency) resonates with a component present in the horizontal field. The horizontal field oscillates with the vertical motion of the particles (due to vertical focusing); its frequency spectrum is dominated by the vertical oscillation frequency and its modulation by the periodic structure of the accelerator ring. In addition, the magnet imperfections that distort the closed orbit vertically contain all integral Fourier harmonics of the orbit frequency. 12. Measurement of HOMs in the RHIC RF Cavities Abreu,N.P.; Choi, E. M. 2009-01-07 The authors present results of Higher Order Modes (HOMs) measurements in the RHIC accelerating (28 MHz system) and storage (197 MHz system) cavities. The power of the excited HOMs deposited into the HOM damper is measured and compared with an analytical calculation of the HOMs power. The quality factors (Q) are also measured and compared to previous measurements. 13. Waste heat recovery from the European Spallation Source cryogenic helium plants - implications for system design Jurns, John M.; Bäck, Harald; Gierow, Martin 2014-01-01 The European Spallation Source (ESS) neutron spallation project currently being designed will be built outside of Lund, Sweden. The ESS design includes three helium cryoplants, providing cryogenic cooling for the proton accelerator superconducting cavities, the target neutron source, and for the ESS instrument suite. In total, the cryoplants consume approximately 7 MW of electrical power, and will produce approximately 36 kW of refrigeration at temperatures ranging from 2-16 K. Most of the power consumed by the cryoplants ends up as waste heat, which must be rejected. One hallmark of the ESS design is the goal to recycle waste heat from ESS to the city of Lund district heating system. The design of the cooling system must optimize the delivery of waste heat from ESS to the district heating system and also assure the efficient operation of ESS systems. This report outlines the cooling scheme for the ESS cryoplants, and examines the effect of the cooling system design on cryoplant design, availability and operation. 14. Thermal Integration of a Liquid Acquisition Device into a Cryogenic Feed System Hastings, L. J.; Bolshinskiy, L. G.; Schunk, R. G.; Martin, A. K.; Eskridge, R. H.; Frenkel, A.; Grayson, G.; Pendleton, M. L. 2011-01-01 Primary objectives of this effort were to define the following: (1) Approaches for quantification of the accumulation of thermal energy within a capillary screen liquid acquisition device (LAD) for a lunar lander upper stage during periods of up to 210 days on the lunar surface, (2) techniques for mitigating heat entrapment, and (3) perform initial testing, data evaluation. The technical effort was divided into the following categories: (1) Detailed thermal modeling of the LAD/feed system interactions using both COMSOL computational fluid device and standard codes, (2) FLOW-3D modeling of bulk liquid to provide interfacing conditions for the LAD thermal modeling, (3) condensation conditioning of capillary screens to stabilize surface tension retention capability, and (4) subscale testing of an integrated LAD/feed system. Substantial progress was achieved in the following technical areas: (1) Thermal modeling and experimental approaches for evaluating integrated cryogen LAD/feed systems, at both the system and component levels, (2) reduced gravity pressure control analyses, (3) analytical modeling and testing for capillary screen conditioning using condensation and wicking, and (4) development of rapid turnaround testing techniques for evaluating LAD/feed system thermal and fluid integration. A comprehensive effort, participants included a diverse cross section of representatives from academia, contractors, and multiple Marshall Space Flight Center organizations. 15. Sub-cooled liquid nitrogen cryogenic system with neon turbo-refrigerator for HTS power equipment Yoshida, S.; Hirai, H.; Nara, N.; Ozaki, S.; Hirokawa, M.; Eguchi, T.; Hayashi, H.; Iwakuma, M.; Shiohara, Y. 2014-01-01 We developed a prototype sub-cooled liquid nitrogen (LN) circulation system for HTS power equipment. The system consists of a neon turbo-Brayton refrigerator with a LN sub-cooler and LN circulation pump unit. The neon refrigerator has more than 2 kW cooling power at 65 K. The LN sub-cooler is a plate-fin type heat exchanger and is installed in a refrigerator cold box. In order to carry out the system performance tests, a dummy cryostat having an electric heater was set instead of a HTS power equipment. Sub-cooled LN is delivered into the sub-cooler by the LN circulation pump and cooled within it. After the sub-cooler, sub-cooled LN goes out from the cold box to the dummy cryostat, and comes back to the pump unit. The system can control an outlet sub-cooled LN temperature by adjusting refrigerator cooling power. The refrigerator cooling power is automatically controlled by the turbo-compressor rotational speed. In the performance tests, we increased an electric heater power from 200 W to 1300 W abruptly. We confirmed the temperature fluctuation was about ±1 K. We show the cryogenic system details and performance test results in this paper. 16. Solid cryogen: a cooling system for future MgB2 MRI magnet Patel, Dipak; Hossain, Md Shahriar Al; Qiu, Wenbin; Jie, Hyunseock; Yamauchi, Yusuke; Maeda, Minoru; Tomsic, Mike; Choi, Seyong; Kim, Jung Ho 2017-03-01 An efficient cooling system and the superconducting magnet are essential components of magnetic resonance imaging (MRI) technology. Herein, we report a solid nitrogen (SN2) cooling system as a valuable cryogenic feature, which is targeted for easy usability and stable operation under unreliable power source conditions, in conjunction with a magnesium diboride (MgB2) superconducting magnet. The rationally designed MgB2/SN2 cooling system was first considered by conducting a finite element analysis simulation, and then a demonstrator coil was empirically tested under the same conditions. In the SN2 cooling system design, a wide temperature distribution on the SN2 chamber was observed due to the low thermal conductivity of the stainless steel components. To overcome this temperature distribution, a copper flange was introduced to enhance the temperature uniformity of the SN2 chamber. In the coil testing, an operating current as high as 200 A was applied at 28 K (below the critical current) without any operating or thermal issues. This work was performed to further the development of SN2 cooled MgB2 superconducting coils for MRI applications. 17. Performance test of the cryogenic cooling system for the superconducting fault current limiter Hong, Yong-Ju; In, Sehwan; Yeom, Han-Kil; Kim, Heesun; Kim, Hye-Rim 2015-12-01 A Superconducting Fault Current Limiter is an electric power device which limits the fault current immediately in a power grid. The SFCL must be cooled to below the critical temperature of high temperature superconductor modules. In general, they are submerged in sub-cooled liquid nitrogen for their stable thermal characteristics. To cool and maintain the target temperature and pressure of the sub-cooled liquid nitrogen, the cryogenic cooling system should be designed well with a cryocooler and coolant circulation devices. The pressure of the cryostat for the SFCL should be pressurized to suppress the generation of nitrogen bubbles in quench mode of the SFCL. In this study, we tested the performance of the cooling system for the prototype 154 kV SFCL, which consist of a Stirling cryocooler, a subcooling cryostat, a pressure builder and a main cryostat for the SFCL module, to verify the design of the cooling system and the electric performance of the SFCL. The normal operation condition of the main cryostat is 71 K and 500 kPa. This paper presents tests results of the overall cooling system. 18. Solid cryogen: a cooling system for future MgB2 MRI magnet Patel, Dipak; Hossain, Md Shahriar Al; Qiu, Wenbin; Jie, Hyunseock; Yamauchi, Yusuke; Maeda, Minoru; Tomsic, Mike; Choi, Seyong; Kim, Jung Ho 2017-01-01 An efficient cooling system and the superconducting magnet are essential components of magnetic resonance imaging (MRI) technology. Herein, we report a solid nitrogen (SN2) cooling system as a valuable cryogenic feature, which is targeted for easy usability and stable operation under unreliable power source conditions, in conjunction with a magnesium diboride (MgB2) superconducting magnet. The rationally designed MgB2/SN2 cooling system was first considered by conducting a finite element analysis simulation, and then a demonstrator coil was empirically tested under the same conditions. In the SN2 cooling system design, a wide temperature distribution on the SN2 chamber was observed due to the low thermal conductivity of the stainless steel components. To overcome this temperature distribution, a copper flange was introduced to enhance the temperature uniformity of the SN2 chamber. In the coil testing, an operating current as high as 200 A was applied at 28 K (below the critical current) without any operating or thermal issues. This work was performed to further the development of SN2 cooled MgB2 superconducting coils for MRI applications. PMID:28251984 19. A hardware overview of the RHIC LLRF platform Hayes, T.; Smith, K.S. 2011-03-28 The RHIC Low Level RF (LLRF) platform is a flexible, modular system designed around a carrier board with six XMC daughter sites. The carrier board features a Xilinx FPGA with an embedded, hard core Power PC that is remotely reconfigurable. It serves as a front end computer (FEC) that interfaces with the RHIC control system. The carrier provides high speed serial data paths to each daughter site and between daughter sites as well as four generic external fiber optic links. It also distributes low noise clocks and serial data links to all daughter sites and monitors temperature, voltage and current. To date, two XMC cards have been designed: a four channel high speed ADC and a four channel high speed DAC. The new LLRF hardware was used to replace the old RHIC LLRF system for the 2009 run. For the 2010 run, the RHIC RF system operation was dramatically changed with the introduction of accelerating both beams in a new, common cavity instead of each ring having independent cavities. The flexibility of the new system was beneficial in allowing the low level system to be adapted to support this new configuration. This hardware was also used in 2009 to provide LLRF for the newly commissioned Electron Beam Ion Source. 20. Hybrid Aerogel-MLI Insulation System for Cryogenic Storage in Space Applications Project National Aeronautics and Space Administration — The future of the NASA space program includes longer and more invasive missions into space. Long duration storage of large quantities of cryogenic fluids for... 1. A High Fidelity Computational Tool for Modeling Thermal Vent Systems in Cryogenic Tanks Project National Aeronautics and Space Administration — Control and management of cryogenic propellant tank pressures in low gravity is an important technical challenge to overcome for future long duration space missions.... 2. Cryogenic Cooling System for Zero-Venting Storage of Supercritical Air Packs Project National Aeronautics and Space Administration — Supercritical air at cryogenic temperature is an attractive source of breathing air because of its very high density and low pressure. However, heat leak into the... 3. Cryogenic Cooling System for Zero-Venting Storage of Supercritical Air Packs Project National Aeronautics and Space Administration — Supercritical air at cryogenic temperature is an attractive source of breathing air because of its very high density and low pressure. However, heat leak into the... 4. The cryogenic system for the Panda-X dark matter search experiment Gong, H.; Giboni, K. L.; Ji, X.; Tan, A.; Zhao, L. 2013-01-01 Panda-X is a liquid xenon dual-phase detector for the Dark Matter Search. The first modestly-sized module will soon be installed in the China JinPing Deep Underground Laboratory in Sichuan province, P.R. China. The cryogenic system is designed to handle much larger detectors, even the final version in the ton scale. Special attention has been paid to the reliability, serviceability, and adaptability to the requirements of a growing experiment. The system is cooled by a single Iwatani PC150 Pulse Tube Refrigerator. After subtracting all thermal losses, the remaining cooling power is still 82 W. The fill speed was 0.75 g/s, but could be boosted by LN2 assisted cooling to 3.3 g/s. For the continuous recirculation and purification through a hot getter, a heat exchanger was employed to reduce the required cooling power. The recirculation speed is limited to 2.9 g/s by the gas pump. At this speed, recirculation only adds 18.5 W to the heat load of the system, corresponding to a 95.2 % efficiency of the heat exchanger. 5. Concept and architecture of the RHIC LLRF upgrade platform Smith, K.S.; Hayes, T.; Severino, F. 2011-03-28 The goal of the RHIC LLRF upgrade has been the development of a stand alone, generic, high performance, modular LLRF control platform, which can be configured to replace existing systems and serve as a common platform for all new RF systems. The platform is also designed to integrate seamlessly into a distributed network based controls infrastructure, be easy to deploy, and to be useful in a variety of digital signal processing and data acquisition roles. Reuse of hardware, software and firmware has been emphasized to minimize development effort and maximize commonality of system components. System interconnection, synchronization and scaling are facilitated by a deterministic, high speed serial timing and data link, while standard intra and inter chassis communications utilize high speed, non-deterministic protocol based serial links. System hardware configuration is modular and flexible, based on a combination of a main carrier board which can host up to six custom or commercial daughter modules as required to implement desired functionality. This paper will provide an overview of the platform concept, architecture, features and benefits. The RHIC LLRF Upgrade Platform has been developed with the goal of providing a flexible, modular and scalable architecture which will support our current applications and satisfy new ones for the foreseeable future. The platform has been recently commissioned at both RHIC and the RHIC EBIS injector. To date the platform has demonstrated its versatility and utility, meeting the design goals as originally defined. 6. Investigation of Vapor Cooling Enhancements for Applications on Large Cryogenic Systems Ameen, Lauren; Zoeckler, Joseph 2017-01-01 The need to demonstrate and evaluate the effectiveness of heat interception methods for use on a relevant cryogenic propulsion stage at a system level has been identified. Evolvable Cryogenics (eCryo) Structural Heat Intercept, Insulation and Vibration Evaluation Rig (SHIIVER) will be designed with vehicle specific geometries (SLS Exploration Upper Stage (EUS) as guidance) and will be subjected to simulated space environments. One method of reducing structure-born heat leak being investigated utilizes vapor-based heat interception. Vapor-based heat interception could potentially reduce heat leak into liquid hydrogen propulsion tanks, increasing potential mission length or payload capability. Due to the high number of unknowns associated with the heat transfer mechanism and integration of vapor-based heat interception on a realistic large-scale skirt design, a sub-scale investigation was developed. The sub-project effort is known as the Small-scale Laboratory Investigation of Cooling Enhancements (SLICE). The SLICE aims to study, design, and test sub-scale multiple attachments and flow configuration concepts for vapor-based heat interception of structural skirts. SLICE will focus on understanding the efficiency of the heat transfer mechanism to the boil-off hydrogen vapor by varying the fluid network designs and configurations. Various analyses were completed in MATLAB, Excel VBA, and COMSOL Multiphysics to understand the optimum flow pattern for heat transfer and fluid dynamics. Results from these analyses were used to design and fabricate test article subsections of a large forward skirt with vapor cooling applied. The SLICE testing is currently being performed to collect thermal mechanical performance data on multiple skirt heat removal designs while varying inlet vapor conditions necessary to intercept a specified amount of heat for a given system. Initial results suggest that applying vapor-cooling provides a 50 heat reduction in conductive heat transmission 7. Dynamic simulations for preparing the acceptance test of JT-60SA cryogenic system Cirillo, R.; Hoa, C.; Michel, F.; Poncet, J. M.; Rousset, B. 2016-12-01 Power generation in the future could be provided by thermo-nuclear fusion reactors like tokamaks. There inside, the fusion reaction takes place thanks to the generation of plasmas at hundreds of millions of degrees that must be confined magnetically with superconductive coils, cooled down to around 4.5 K. Within this frame, an experimental tokamak device, JT-60SA is currently under construction in Naka (Japan). The plasma works cyclically and the coil system is subject to pulsed heat loads. In order to size the refrigerator close to the average power and hence optimizing investment and operational costs, measures have to be taken to smooth the heat load. Here we present a dynamic model of the JT-60SA's Auxiliary Cold box (ACB) for preparing the acceptance tests of the refrigeration system planned in 2016 in Naka. The aim of this study is to simulate the pulsed load scenarios using different process controls. All the simulations have been performed with EcosimPro® and the associated cryogenic library: CRYOLIB. 8. Study on mitigation of pulsed heat load for ITER cryogenic system Peng, N.; Xiong, L. Y.; Jiang, Y. C.; Tang, J. C.; Liu, L. Q. 2015-03-01 One of the key requirements for ITER cryogenic system is the mitigation of the pulsed heat load deposited in the magnet system due to magnetic field variation and pulsed DT neutron production. As one of the control strategies, bypass valves of Toroidal Field (TF) case helium loop would be adjusted to mitigate the pulsed heat load to the LHe plant. A quasi-3D time-dependent thermal-hydraulic analysis of the TF winding packs and TF case has been performed to study the behaviors of TF magnets during the reference plasma scenario with the pulses of 400 s burn and repetition time of 1800 s. The model is based on a 1D helium flow and quasi-3D solid heat conduction model. The whole TF magnet is simulated taking into account thermal conduction between winding pack and case which are cooled separately. The heat loads are given as input information, which include AC losses in the conductor, eddy current losses in the structure, thermal radiation, thermal conduction and nuclear heating. The simulation results indicate that the temperature variation of TF magnet stays within the allowable range when the smooth control strategy is active. 9. Control and materials characterization System for 6T Superconducting Cryogen Free Magnet Facility at IUAC, New Delhi Dutt, R. N.; Meena, D. K.; Kar, S.; Soni, V.; Nadaf, A.; Das, A.; Singh, F.; Datta, T. S. 2017-02-01 A system for carrying out automatic experimental measurements of various electrical transport characteristics and their relation to magnetic fields for samples mounted on the sample holder on a Variable Temperature Insert (VTI) of the Cryogen Free Superconducting Magnet System (CFMS) has been developed. The control and characterization system is capable of monitoring, online plotting and history logging in real-time of cryogenic temperatures with the Silicon (Si) Diode and Zirconium Oxy-Nitride sensors installed inside the magnet facility. Electrical transport property measurements have been automated with implementation of current reversal resistance measurements and automatic temperature set-point ramping with the parameters of interest available in real-time as well as for later analysis. The Graphical User Interface (GUI) based system is user friendly to facilitate operations. An ingenious electronics for reading Zirconium Oxy-Nitride temperature sensors has been used. Price to performance ratio has been optimized by using in house developed measurement techniques mixed with specialized commercial cryogenic measurement / control equipment. 10. A model for manuscript submitted to the nth IIR conference on overview of the long-baseline neutrino facility cryogenic system Montanari, David [FERMILAB; Adamowski, Mark [Fermilab; Bremer, Johan [CERN; Delaney, Michael [Fermilab; Aurelien, Diaz [CERN; Doubnik, Roza [Fermilab; Haaf, Kevin [Fermilab; Hentschel, Steve [Fermilab; Norris, Barry [Fermilab; Voirin, Erik [Fermilab 2017-03-09 The Deep Underground Neutrino Experiment (DUNE) collaboration is developing a multi-kiloton Long-Baseline neutrino experiment that will be located one mile underground at the Sanford Underground Research Facility (SURF) in Lead, SD. In the present design, detectors will be located inside four cryostats filled with a total of 68,400 ton of ultrapure liquid argon, at the level of impurities lower than 100 parts per trillion of oxygen equivalent contamination. The Long-Baseline Neutrino Facility (LBNF) is developing the conventional facilities and cryogenics infrastructure supporting this experiment. The cryogenics system is composed of several sub-systems: External/Infrastructure, Proximity, and Internal cryogenics. It will be engineered, manufactured, commissioned, and qualified by an international engineering team. This contribution highlights the main features of the LBNF cryogenic system. It presents its performance, functional requirements and modes of operations. As a result, it also details the status of the design, present and future needs. 11. Cryogenic engineering fifty years of progress Reed, Richard 2007-01-01 Cryogenic Engineering: Fifty Years of Progress is a benchmark reference work which chronicles the major developments in the field. Starting with an historical background dating to the 1850s, this book reviews the development of data resources now available for cryogenic fields and properties of materials. The advances in cryogenic fundamentals are covered by reviews of cryogenic principles, cryogenic insulation, low-loss storage systems, modern liquefaction processes, helium cryogenics and low-temperature thermometry. Several well-established applications resulting from cryogenic advances include aerospace cryocoolers and refrigerators, use of LTS and HTS systems in electrical applications, and recent changes in cryopreservation. Extensive references are provided for the readers interested in the details of these cryogenic engineering advances. 12. Development of the Cryogenic System of AEgIS at CERN Derking, J H; Burghart, G; Doser, M; Dudarev, A; Haider, S 2014-01-01 The AEgIS (Antimatter Experiment: Gravity, Interferometry, Spectroscopy) experiment is located at the antiproton decelerator complex of CERN. The main goal of the experiment is to perform the first direct measurement of the Earth’s gravitational acceleration on antihydrogen atoms within 1% precision. The antihydrogen is produced in a cylindrical Penning trap by combining antiprotons with positrons. To reach the precision of 1%, the antihydrogen has to be cooled to 100 mK to reduce its random velocity. A dilution refrigerator is selected to deliver the necessary cooling capacity of 100 μW at 50 mK. The AEgIS cryogenic system basically consists of cryostats for a 1-T and for a 5-T superconducting magnet, a central region cryostat, a dilution refrigerator cryostat and a measurement cryostat with a Moiré deflectometer to measure the gravitational acceleration. In autumn 2012, the 1-T cryostat, 5-T cryostat and central region cryostat were assembled and commissioned. The apparatus is cooled down in eight days ... 13. 冷冻靶制备用低温氦气循环系统%Cycling helium system for cryogenic target handling system 丁先庚; 丁怀况; 施锦 2012-01-01 The cryogenic target which is treated by high - pressure permeation charge at room temperature and after cryogenic cooling, needs cryogenic cycling helium with temperature below 20K to cool the high - pressure permeation cell and cryostat. Adopting GM crybcooler as the cold source and dedicated helium compressor as the cycling pump, with the design of high - efficiency regenerative heat interchanger, the system can obtain cryogenic helium with temperature below 20K through which the end fittings are cooled down, and thereby to achieve cryogenic and homothermal environment and cooling of permeation cell%常温高压渗透充气、低温冷却的冷冻靶球,需要20K以下的低温循环氦气,用于冷却高压渗透室和低温恒温腔.本套系统采用GM制冷机为冷源,采用专用氦压缩机为循环泵,设计高效率的回热式换热器,实现末端的20K以下低温氦气,通过低温氦气冷却终端部件,实现了20K的低温恒温环境和渗透室的冷却. 14. Cryogenically cooled low-noise amplifier for radio-astronomical observations and centimeter-wave deep-space communications systems Vdovin, V. F.; Grachev, V. G.; Dryagin, S. Yu.; Eliseev, A. I.; Kamaletdinov, R. K.; Korotaev, D. V.; Lesnov, I. V.; Mansfeld, M. A.; Pevzner, E. L.; Perminov, V. G.; Pilipenko, A. M.; Sapozhnikov, B. D.; Saurin, V. P. 2016-01-01 We report a design solution for a highly reliable, low-noise and extremely efficient cryogenically cooled transmit/receive unit for a large antenna system meant for radio-astronomical observations and deep-space communications in the X band. We describe our design solution and the results of a series of laboratory and antenna tests carried out in order to investigate the properties of the cryogenically cooled low-noise amplifier developed. The transmit/receive unit designed for deep-space communications (Mars missions, radio observatories located at Lagrangian point L2, etc.) was used in practice for communication with live satellites including "Radioastron" observatory, which moves in a highly elliptical orbit. 15. First results from RHIC-PHENIX Ghosh, T K; Adler, S S; Ajitanand, N N; Akiba, Y; Alexander, J; Aphecetche, L; Arai, Y; Aronson, S H; Averbeck, R; Awes, T C; Barish, K N; Barnes, P D; Barrette, J; Bassalleck, B; Bathe, S; Baublis, V; Bazilevsky, A; Belikov, S V; Bellaiche, F G; Belyaev, S T; Bennett, M J; Berdnikov, Yu A; Botelho, S S; Brooks, M L; Brown, D S; Bruner, N L; Bucher, D; Büsching, H; Bunce, G M; Burward-Hoy, J M; Butsyk, S; Carey, T A; Chand, P; Chang, J; Chang, W C; Chavez, L L; Chernichenko, S K; Chi, C Y; Chiba, J; Chiu, M; Choudhury, R K; Christ, T; Chujo, T; Chung, M S; Chung, P; Cianciolo, V; Cole, B A; D'Enterria, D G; Dávid, G; Delagrange, H; Denisov, A; Deshpande, A A; Desmond, E J; Dietzsch, O; Dinesh, B V; Drees, A; Durum, A A; Dutta, D; Ebisu, K; Efremenko, Yu V; Chenawi, K E; En-Yo, H; Esumi, S C; Ewell, L A; Ferdousi, T; Fields, D E; Fokin, S L; Fraenkel, Zeev; Franz, A; Frawley, A D; Fung, S Y; Garpman, S; Ghosh, T K; Glenn, A; Godoi, A L; Goto, Y; Greene, S V; Grosse-Perdekamp, M; Gupta, S K; Guryn, W; Gustafsson, Hans Åke; Haggerty, J S; Hamagaki, H; Hansen, A G; Hara, H; Hartouni, E P; Havano, R; Hayashi, N; He, X; Hemmick, T K; Heuser, J M; Hill, J C; Ho, D S; Homma, K; Hong, B; Hoover, A; Ichihara, T; Imai, K; Ippolitov, M S; Ishihara, M; Jacak, B V; Jang, W Y; Jia, J; Johnson, B M; Johnson, S C; Joo, K S; Kametani, S; Kang, J H; Kann, M; Kapoor, S S; Kelly, S; Khachaturov, B A; Khanzadeev, A V; Kikuchi, J; Kim, D J; Kim, H J; Kim, S Y; Kim, Y G; Kinnison, W W; Kistenev, E P; Kiyomichi, A; Klein-Bösing, C; Klinksiek, S A; Kochenda, L M; Kochetkov, D; Kochetkov, V; Köhler, D; Kohama, T; Kozlov, A; Kroon, P J; Kurita, K; Kweon, M J; Kwon, Y; Kyle, G S; Lacey, R A; Lajoie, J G; Lauret, J; Lebedev, A; Lee, D M; Leitch, M J; Li, X H; Li, Z; Lim, D J; Liu, M X; Liu, X; Liu, Z; Maguire, C F; Mahon, J; Makdisi, Y I; Man'ko, V I; Mao, Y; Mark, S K; Markacs, S; Martínez, G; Marx, M D; Massaike, A; Matathias, F; Matsumoto, T; McGaughey, P L; Melnikov, E A; Merschmeyer, M; Messer, F; Messer, M; Miake, Y; Miller, T E; Milov, A; Mioduszewski, S; Mischke, R E; Mishra, G C; Mitchell, J T; Mohanty, A K; Morrison, D P; Moss, J M; Muhlbacher, F; Muniruzzaman, M; Murata, J; Nagamiya, S; Nagasaka, Y; Nagle, J L; Nakada, Y; Nandi, B K; Newby, J; Nikkinen, L; Nilsson, P O; Nishimura, S; Nyanin, A S; Nystrand, J; O'Brien, E; Ogilvie, C A; Ohnishi, H; Ojha, I D; Ono, M; Onuchin, V A; Oskarsson, A; Österman, L; Otterlund, I; Oyama, K; Paffrath, L; Palounek, A P T; Pantuev, V S; Papavassiliou, V; Pate, S F; Peitzmann, Thomas; Petridis, A N; Pinkenburg, C H; Pisani, R P; Pitukhin, P; Plasil, F; Pollack, M E; Pope, K; Purschke, M L; Ravinovich, I; Read, K F; Reygers, K; Riabov, V; Riabov, Y; Rosati, M; Rose, A A; Ryu, S S; Saitô, N; Sakaguchi, A; Sakaguchi, T; Sako, H; Sakuma, T; Samsonov, V; Sangster, T C; Santo, R; Sato, H D; Sato, S; Sawada, S; Schlei, B R; Schutz, Y; Semenov, V; Seto, R; Shea, T K; Shein, I; Shibata, T A; Shigaki, K; Shiina, T; Shin, Y H; Sibiryak, Yu; Silvermyr, D; Sim, K S; Simon-Gillo, J; Singh, C P; Singh, V; Sivertz, M; Soldatov, A; Soltz, R A; Sørensen, S; Stankus, P W; Starinsky, N; Steinberg, P; Stenlund, E; Ster, A; Stoll, S P; Sugioka, M; Sugitate, T; Sullivan, J P; Sumi, Y; Sun, Z; Suzuki, M; Takagui, E M; Taketani, A; Tamai, M; Tanaka, Y; Taniguchi, E; Tannenbaum, M J; Thomas, J; Thomas, J H; Thomas, T L; Tian, W; Tojo, J; Torii, H A; Towell, R S; Tserruya, Itzhak; Tsuruoke, H; Tsvetkov, A A; Tuli, S K; Tydesjo, H; Tyurin, N; Ushiroda, T; van Hecke, H; Velissaris, C; Velkovska, J; Velkovsky, M; Vingradov, A A; Volkov, M A; Vorobyov, A A; Vznuzdaev, E A; Wang, H; Watanabe, Y; White, S N; Witzig, C; Wohn, F K; Woody, C L; Xie, W; Yagi, K; Yokkaichi, S; Young, G R; Yushmanov, I E; Zajc, W A; Zhang, Z; Zhou, S 2001-01-01 The PHENIX experiment consists of a large detector system located at the newly commissioned Relativistic Heavy Ion Collider (RHIC) at the Brookhaven National Laboratory. The primary goal of the PHENIX experiment is to look for signatures of the QCD prediction of a deconfined high-energy-density phase of nuclear matter and the quark gluon plasma. PHENIX started taking data for Au+Au collisions at square root (s/sub NN/)=130 GeV in June 2000. The signals from the beam-beam counter (BBC) and zero degree calorimeter (ZDC) are used to determine the centrality of the collision. A Glauber model reproduces the ZDC spectrum reasonably well to determine the participants in a collision. The charged particle multiplicity distribution from the first PHENIX paper is compared with the other RHIC experiment and the CERN and SPS results. Transverse momentum of photons are measured in the electro-magnetic calorimeter (EMCal) and preliminary results an presented. Particle identification is made by a time-of-flight (TOF) detecto... 16. First results from RHIC-PHENIX Tarun Kanti Ghosh; K Adcox; S S Adler; N Ajitanand; Y Akiba; J Alexander; L Aphecetche; Y Arai; S H Aronson; R Averbeck; T C Awes; K N Barish; P D Barnes; J Barrette; B Bassalleck; S Bathe; V Baublis; A Bazilevsky; S Belikov; F G Bellaiche; S T Belyaev; M J Bennett; Y Berdnikov; S Botelho; M L Brooks; D S Brown; N Bruner; D Bucher; H Buesching; V Bumazhnov; G Bunce; J Burward-Hoy; S Butsyk; T A Carey; P Chand; J Chang; W C Chang; L L Chavez; S Chernichenko; C Y Chi; J Chiba; M Chiu; R K Choudhury; T Christ; T Chujo; M S Chung; P Chung; V Cianciolo; B A Cole; D G D’Enterria; G David; H Delagrange; A Denisov; A Deshpande; E J Desmond; O Dietzsch; B V Dinesh; A Drees; A Durum; D Dutta; K Ebisu; Y V Efremenko; K El Chenawi; H En’yo; S Esumi; L Ewell; T Ferdousi; D E Fields; S L Fokin; Z Fraenkel; A Franz; A D Frawley; S-Y Fung; S Garpman; T K Ghosh; A Glenn; A L Godoi; Y Goto; S V Greene; M Grosse Perdekamp; S K Gupta; W Guryn; H-Å Gustafsson; J S Haggerty; H Hamagaki; A G Hansen; H Hara; E P Hartouni; R Hayano; N Hayashi; X He; T K Hemmick; J Heuser; J C Hill; D S Ho; K Homma; B Hong; A Hoover; T Ichihara; K Imai; M S Ippolitov; M Ishihara; B V Jacak; W Y Jang; J Jia; B M Johnson; S C Johnson; K S Joo; S Kametani; J H Kang; M Kann; S S Kapoor; S Kelly; B Khachaturov; A Khanzadeev; J Kikuchi; D J Kim; H J Kim; S Y Kim; Y G Kim; W W Kinnison; E Kistenev; A Kiyomichi; C Klein-Boesing; S Klinksiek; L Kochenda; D Kochetkov; V Kochetkov; D Koehler; T Kohama; A Kozlov; P J Kroon; K Kurita; M J Kweon; Y Kwon; G S Kyle; R Lacey; J G Lajoie; J Lauret; A Lebedev; D M Lee; M J Leitch; X H Li; Z Li; D J Lim; M X Liu; X Liu; Z Liu; C F Maguire; J Mahon; Y I Makdisi; V I Manko; Y Mao; S K Mark; S Markacs; G Martinez; M D Marx; A Masaike; F Matathias; T Matsumoto; P L McGaughey; E Melnikov; M Merschmeier; F Messer; M Messer; Y Miake; T E Miller; A Milov; S Mioduszewski; R E Mischke; G C Mishra; J T Mitchell; A K Mohanty; D P Morrison; J M Moss; F Mühlbacher; M Muniruzzaman; J Murata; S Nagamiya; Y Nagasaka; J L Nagle; Y Nakada; B K Nandi; J Newby; L Nikkinen; P Nilsson; S Nishimura; A S Nyanin; J Nystrand; E O’Brien; C A Ogilvie; H Ohnishi; I D Ojha; M Ono; V Onuchin; A Oskarsson; L Österman; I Otterlund; K Oyama; L Paffrath; A P T Palounek; V S Pantuev; V Papavassiliou; S F Pate; T Peitzmann; A N Petridis; C Pinkenburg; R P Pisani; P Pitukhin; F Plasil; M Pollack; K Pope; M L Purschke; I Ravinovich; K F Read; K Reygers; V Riabov; Y Riabov; M Rosati; A A Rose; S S Ryu; N Saito; A Sakaguchi; T Sakaguchi; H Sako; T Sakuma; V Samsonov; T C Sangster; R Santo; H D Sato; S Sato; S Sawada; B R Schlei; Y Schutz; V Semenov; R Seto; T K Shea; I Shein; T-A Shibata; K Shigaki; T Shiina; Y H Shin; I G Sibiriak; D Silvermyr; K S Sim; J Simon-Gillo; C P Singh; V Singh; M Sivertz; A Soldatov; R A Soltz; S Sorensen; P W Stankus; N Starinsky; P Steinberg; E Stenlund; A Ster; S P Stoll; M Sugioka; T Sugitate; J P Sullivan; Y Sumi; Z Sun; M Suzuki; E M Takagui; A Taketani; M Tamai; K H Tanaka; Y Tanaka; E Taniguchi; M J Tannenbaum; J Thomas; J H Thomas; T L Thomas; W Tian; J Tojo; H Torii; R S Towell; I Tserruya; H Tsuruoka; A A Tsvetkov; S K Tuli; H Tydesjö; N Tyurin; T Ushiroda; H W van Hecke; C Velissaris; J Velkovska; M Velkovsky; A A Vinogradov; M A Volkov; A Vorobyov; E Vznuzdaev; H Wang; Y Watanabe; S N White; C Witzig; F K Wohn; C L Woody; W Xie; K Yagi; S Yokkaichi; G R Young; I E Yushmanov; W A Zajc; Z Zhang; S Zhou 2001-08-01 The PHENIX experiment consists of a large detector system located at the newly commissioned relativistic heavy ion collider (RHIC) at the Brookhaven National Laboratory. The primary goal of the PHENIX experiment is to look for signatures of the QCD prediction of a deconfined high-energy-density phase of nuclear matter quark gluon plasma. PHENIX started data taking for Au+Au collisions at $\\sqrt{s_{NN}} = 130$ GeV in June 2000. The signals from the beam-beam counter (BBC) and zero degree calorimeter (ZDC) are used to determine the centrality of the collision. A Glauber model reproduces the ZDC spectrum reasonably well to determine the participants in a collision. Charged particle multiplicity distribution from the first PHENIX paper is compared with the other RHIC experiment and the CERN, SPS results. Transverse momentum of photons are measured in the electro-magnetic calorimeter (EMCal) and preliminary results are presented. Particle identification is made by a time of flight (TOF) detector and the results show clear separation of the charged hadrons from each other. 17. Joining the RHIC Online and Offline Models Malitsky, Nikolay; Fedotov, Alexei V; Kewisch, Jorg; Luccio, Alfredo U; Pilat, Fulvia Caterina; Ptitsyn, Vadim; Satogata, Todd; Talman, Richard M; Tepikian, Steven; Wei, Jie 2005-01-01 The paper presents an interface encompassing the RHIC online ramp model and the UAL offline simulation framework. The resulting consolidated facility aims to minimize the gap between design and operational data, and to facilitate analysis of RHIC performance and future upgrades in an operational context. The interface is based on the Accelerator Description Exchange Format (ADXF), and represents a snapshot of the RHIC online model which is in turn driven by machine setpoints. This approach is also considered as an intermediate step towards integrating the AGS and RHIC modeling environments to produce a unified online and offline AGS model for operations. 18. Workshop on the RHIC performance Khiari, F.; Milutinovic, J.; Ratti, A.; Rhoades-Brown, M.J. (eds.) 1988-07-01 The most recent conceptual design manual for the Relativistic Heavy Ion Collider (RHIC) at Brookhaven was published in May 1986 (BNL 51932). The purpose of this workshop was to review the design specifications in this RHIC reference manual, and to discuss in detail possible improvements in machine performance by addressing four main areas. These areas are beam-beam interactions, stochastic cooling, rf and bunch instabilities. The contents of this proceedings are as follows. Following an overview of the workshop, in which the motivation and goals are discussed in detail, transcripts of the first day talks are given. Many of these transcripts are copies of the original transparencies presented at the meeting. The following four sections contain contributed papers, that resulted from discussions at the workshop within each of the four working groups. In addition, there is a group summary for each of the four working groups at the beginning of each section. Finally, a list of participants is given. 19. Monolithic readout circuits for RHIC OConnor, P.; Harder, J. [Brookhaven National Laboratory, Upton, NY (United States) 1991-12-31 Several CMOS ASICs have been developed for a proposed RHIC experiment. This paper discusses why ASIC implementation was chosen for certain functions, circuit specifications and the design techniques used to meet them, and results of simulations and early prototypes. By working closely together from an early stage in the planning process, in-house ASIC designers and detector and data acquisition experimenters can achieve optimal use of this important technology. 20. Monolithic readout circuits for RHIC OConnor, P.; Harder, J. [Brookhaven National Laboratory, Upton, NY (United States) 1991-12-31 Several CMOS ASICs have been developed for a proposed RHIC experiment. This paper discusses why ASIC implementation was chosen for certain functions, circuit specifications and the design techniques used to meet them, and results of simulations and early prototypes. By working closely together from an early stage in the planning process, in-house ASIC designers and detector and data acquisition experimenters can achieve optimal use of this important technology. 1. High intensity protons in RHIC Montag, C.; Ahrens, L.; Blaskiewicz, M.; Brennan, J. M.; Drees, K. A.; Fischer, W.; Huang, H.; Minty, M.; Robert-Demolaize, G.; Thieberger, P.; Yip, K. 2012-01-05 During the 2012 summer shutdown a pair of electron lenses will be installed in RHIC, allowing the beam-beam parameter to be increased by roughly 50 percent. To realize the corresponding luminosity increase bunch intensities have to be increased by 50 percent, to 2.5 {center_dot} 10{sup 11} protons per bunch. We list the various RHIC subsystems that are most affected by this increase, and propose beam studies to ensure their readiness. The proton luminosity in RHIC is presently limited by the beam-beam effect. To overcome this limitation, electron lenses will be installed in IR10. With the help of these devices, the headon beam-beam kick experienced during proton-proton collisions will be partially compensated, allowing for a larger beam-beam tuneshift at these collision points, and therefore increasing the luminosity. This will be accomplished by increasing the proton bunch intensity from the presently achieved 1.65 {center_dot} 10{sup 11} protons per bunch in 109 bunches per beam to 2.5 {center_dot} 10{sup 11}, thus roughly doubling the luminosity. In a further upgrade we aim for bunch intensities up to 3 {center_dot} 10{sup 11} protons per bunch. With RHIC originally being designed for a bunch intensity of 1 {center_dot} 10{sup 11} protons per bunch in 56 bunches, this six-fold increase in the total beam intensity by far exceeds the design parameters of the machine, and therefore potentially of its subsystems. In this note, we present a list of major subsystems that are of potential concern regarding this intensity upgrade, show their demonstrated performance at present intensities, and propose measures and beam experiments to study their readiness for the projected future intensities. 2. HIGH-CURRENT ERL-BASED ELECTRON COOLING FOR RHIC. BEN-ZVI, I. 2005-09-18 The design of an electron cooler must take into account both electron beam dynamics issues as well as the electron cooling physics. Research towards high-energy electron cooling of RHIC is in its 3rd year at Brookhaven National Laboratory. The luminosity upgrade of RHIC calls for electron cooling of various stored ion beams, such as 100 GeV/A gold ions at collision energies. The necessary electron energy of 54 MeV is clearly out of reach for DC accelerator system of any kind. The high energy also necessitates a bunched beam, with a high electron bunch charge, low emittance and small energy spread. The Collider-Accelerator Department adopted the Energy Recovery Linac (ERL) for generating the high-current, high-energy and high-quality electron beam. The RHIC electron cooler ERL will use four Superconducting RF (SRF) 5-cell cavities, designed to operate at ampere-class average currents with high bunch charges. The electron source will be a superconducting, 705.75 MHz laser-photocathode RF gun, followed up by a superconducting Energy Recovery Linac (ERL). An R&D ERL is under construction to demonstrate the ERL at the unprecedented average current of 0.5 amperes. Beam dynamics performance and luminosity enhancement are described for the case of magnetized and non-magnetized electron cooling of RHIC. 3. The development of the control system for the cryogenics in the LHC tunnel; Le developpement du systeme de controle de la cryogenie dans le tunnel du LHC Fluder, C.; Blanco-Vinuela, E.; Casas-Cubillos, J.; Dubert, P.; Gomes, P.; Pezzetti, M.; Tovar-Gonzalez, A.; Zwalinski, L. [CERN-Conseil Europeen pour la recherche nucleaire, Geneve (Switzerland); Zwalinski, L. [Universite AGH des Sciences et Technologies (Poland) 2011-07-01 This paper reviews the control system architecture and the main hardware and software components; presents the hardware commissioning and software production methodologies; and illustrates some of the problems faced during development, commissioning and nominal cryogenics operation, together with the solutions applied. (authors) 4. Suitability of commercially available laboratory cryogenic refrigerators to support shipboard electro-optical systems in the 10 - 77 Kelvin region Hansen, R. G.; Byrd, E. A. 1983-12-01 The primary development of cryogenically cooled infrared systems was accomplished by FLIR systems designed for airborne, passive night vision. Essential to the development of these FLIR systems was a family of closed cycle refrigerators which had to meet a limited envelope requirement, utilize a nonlubricated compressor module, and be light in weight. Closed cycle refrigerators accomplished the same cooling function, they use modified oil lubricated reciprocating compressors which are limited in their axis of orientation to an angle of approximately 15-20 degrees maximum from horizon. 5. Performance of the primary mirror center-of-curvature optical metrology system during cryogenic testing of the JWST Pathfinder telescope Hadaway, James B.; Wells, Conrad; Olczak, Gene; Waldman, Mark; Whitman, Tony; Cosentino, Joseph; Connolly, Mark; Chaney, David; Telfer, Randal 2016-07-01 The James Webb Space Telescope (JWST) primary mirror (PM) is 6.6 m in diameter and consists of 18 hexagonal segments, each 1.5 m point-to-point. Each segment has a six degree-of-freedom hexapod actuation system and a radius of-curvature (RoC) actuation system. The full telescope will be tested at its cryogenic operating temperature at Johnson Space Center. This testing will include center-of-curvature measurements of the PM, using the Center-of-Curvature Optical Assembly (COCOA) and the Absolute Distance Meter Assembly (ADMA). The COCOA includes an interferometer, a reflective null, an interferometer-null calibration system, coarse and fine alignment systems, and two displacement measuring interferometer systems. A multiple-wavelength interferometer (MWIF) is used for alignment and phasing of the PM segments. The ADMA is used to measure, and set, the spacing between the PM and the focus of the COCOA null (i.e. the PM center-of-curvature) for determination of the ROC. The performance of these metrology systems was assessed during two cryogenic tests at JSC. This testing was performed using the JWST Pathfinder telescope, consisting mostly of engineering development and spare hardware. The Pathfinder PM consists of two spare segments. These tests provided the opportunity to assess how well the center-of-curvature optical metrology hardware, along with the software and procedures, performed using real JWST telescope hardware. This paper will describe the test setup, the testing performed, and the resulting metrology system performance. The knowledge gained and the lessons learned during this testing will be of great benefit to the accurate and efficient cryogenic testing of the JWST flight telescope. 6. Computer controlled vent and pressurization system. [of launch vehicle cryogenic propellant tanks Cieslewicz, E. J. 1975-01-01 The paper illustrates how the Centaur space launch vehicle airborne computer, which was primarily used to perform guidance, navigation, and sequencing tasks, was further used to monitor and control inflight pressurization and venting of the cryogenic propellant tanks. Computer software flexibility also provided a failure detection and correction capability necessary to adopt and operate redundant hardware techniques and enhance the overall vehicle reliability. 7. Evaluation of static mixer flow enhancements for cryogenic viscous compressor prototype for ITER vacuum system Duckworth, Robert C.; Baylor, Larry R.; Meitner, Steven J.; Combs, Stephen K.; Ha, Tam; Morrow, Michael; Biewer, T. [Fusion and Materials for Nuclear System Division, Oak Ridge National Laboratory, Oak Ridge (United States); Rasmussen, David A.; Hechler, Michael P. [U.S. ITER Project Office, Oak Ridge National Laboratory, Oak Ridge (United States); Pearce, Robert J. H.; Dremel, Mattias [ITER Organization, 13115 St. Paul-lez-Durance (France); Boissin, J.-C. [Consultant, Grenoble (France) 2014-01-29 As part of the U.S. ITER contribution to the vacuum systems for the ITER fusion project, a cryogenic viscous compressor (CVC) is being designed and fabricated to cryopump hydrogenic gases in the torus and neutral beam exhaust streams and to regenerate the collected gases to controlled pressures such that they can be mechanically pumped with controlled flows to the tritium reprocessing facility. One critical element of the CVC design that required additional investigation was the determination of flow rates of the low pressure (up to 1000 Pa) exhaust stream that would allow for complete pumping of hydrogenic gases while permitting trace levels of helium to pass through the CVC to be pumped by conventional vacuum pumps. A sub-scale prototype test facility was utilized to determine the effectiveness of a static mixer pump tube concept, which consisted of a series of rotated twisted elements brazed into a 2-mm thick, 5-cm diameter stainless steel tube. Cold helium gas flow provided by a dewar and helium transfer line was used to cool the exterior of the static mixer pump tube. Deuterium gas was mixed with helium gas through flow controllers at different concentrations while the composition of the exhaust gas was monitored with a Penning gauge and optical spectrometer to determine the effectiveness of the static mixer. It was found that with tube wall temperatures between 6 K and 9 K, the deuterium gas was completely cryopumped and only helium passed through the tube. These results have been used to design the cooling geometry and the static mixer pump tubes in the full-scale CVC prototype. 8. Evaluation of Static Mixer Flow Enhancements for Cryogenic Viscous Compressor Prototype for ITER Vacuum System Duckworth, Robert C [ORNL; Baylor, Larry R [ORNL; Meitner, Steven J [ORNL; Combs, Stephen Kirk [ORNL; Ha, Tam T [ORNL; Morrow, Michael [ORNL; Biewer, Theodore M [ORNL; Rasmussen, David A [ORNL; Hechler, Michael P [ORNL; Pearce, R.J.H. [ITER Organization, Cadarache, France; Dremel, M. [ITER Organization, Cadarache, France; Boissin, Jean Claude [Consultant 2014-01-01 As part of the U.S. ITER contribution to the vacuum systems for the ITER fusion project, a cryogenic viscous compressor (CVC) is being designed and fabricated to cryopump hydrogenic gases in the torus and neutral beam exhaust streams and to regenerate the collected gases to controlled pressures such that they can be mechanically pumped with controlled flows to the tritium reprocessing facility. One critical element of the CVC design that required additional investigation was the determination of flow rates of the low pressure (50 to 1000 Pa) exhaust stream that would allow for complete pumping of hydrogenic gases while permitting trace levels of helium to pass through the CVC to be pumped by conventional vacuum pumps. A sub-scale prototype test facility was utilized to determine the effectiveness of a static mixer pump tube concept, which consisted of a series of rotated twisted elements brazed into a 2-mm thick, 5-cm diameter stainless steel tube. Cold helium gas flow provided by a dewar and helium transfer line was used to cool the exterior of the static mixer pump tube. Deuterium gas was mixed with helium gas through flow controllers at different concentrations while the composition of the exhaust gas was monitored with a Penning gauge and optical spectrometer to determine the effectiveness of the static mixer. It was found that with tube wall temperatures between 6 K and 9 K, the deuterium gas was completely cryopumped and only helium passed through the tube. These results have been used to design the cooling geometry and the static mixer pump tubes in the full-scale CVC prototype 9. PHYSICS OF POLARITY AT RHIC-VOLUME 10. IMAI,K.; FIELDS,D. 1998-08-04 The RBRC Workshop on Physics of Polarimetry at RHIC was held from Aug 4 to 7, 1998 at BNL. The primary motive of the workshop is (1) to discuss the RHIC polarimeter using the elastic proton-carbon scattering at Coulomb-nuclear interference region (p-C CNI polarimeter) in detail and write a proposal for the test experiment a t the AGS, (2) to discuss the related physics, (3) and to discuss other options for the RHIC polarimetry. The idea of the p-C CNI polarimeter was proposed last year as a simple, inexpensive and efficient polarimeter for RHIC. In order to establish this polarimeter, we have decided to carry out a test experiment by using a polarized beam at the AGS. We have made a draft of the proposal during the workshop. For the p-C CNI polarimeter, a telescope detector using both the micro-channel plate (MCP) and the SSD was proposed to detect low energy recoil carbon ions, based on the test measurements at IUCF and Kyoto, where the carbon ions as low as 200 keV were successfully detected. The kinetic energy of carbon ion is measured with the SSD, and the velocity is measured by TOF between the two detectors and between the accelerator rf pulse and the two detectors. Counting rates for the background and true events were estimated. With the proposed polarimeter, one can expect to measure the beam polarization at the AGS and RHIC at an accuracy of 10% within a reasonable time period. We will test this detector system at Kyoto as soon as possible and install it in the AGS ring for the test measurement of A{sub N} during E880 which is scheduled early in the next year. Zizek, F. 1982-01-01 Theoretical, technical and design questions are examined of cryogenic current leads for SP of magnetic systems. Simplified mathematical models are presented for the current leads. To illustrate modeling, the calculation is made of the real current leads for 500 A and three variants of current leads for 1500 A for the enterprise ''Shkoda.'' 11. Elastic proton-proton scattering at RHIC Yip, K. 2011-09-03 Here we describe elastic proton+proton (p+p) scattering measurements at RHIC in p+p collisions with a special optics run of {beta}* {approx} 21 m at STAR, at the center-of-mass energy {radical}s = 200 GeV during the last week of the RHIC 2009 run. We present preliminary results of single and double spin asymmetries. 12. Beam dynamics and expected RHIC performance with 56MHz RF upgrade Fedotov,A.V.; Ben-Zvi, I. 2009-05-04 An upgrade of the RHIC storage RF system with a superconducting 56 MHz cavity was recently proposed. This upgrade will provide a significant increase in the acceptance of the RHIC 197 MHz storage RF bucket. This paper summarizes simulations of beam evolution due to intra-beam scattering (IBS) for beam parameters expected with the 56 MHz SRF cavity upgrade. Expected luminosity improvements are shown for Au ions at 100 GeV/nucleon and protons at 250 GeV. 13. Optics measurements and corrections at RHIC Bai M.; Aronson, J.; Blaskiewicz, M.; Luo, Y.; Robert-Demolaize, G.; White, S. 2012-05-20 The further improvement of RHIC luminosity performance requires more precise understanding of the RHIC modeling. Hence, it is necessary to minimize the beta-beat, deviation of measured beta function from the calculated beta functions based on an model. The correction of betabeat also opens up the possibility of exploring operating RHIC polarized protons at a working point near integer, a prefered choice for both luminosity as well as beam polarization. The segment-by-segment technique for reducing beta-beat demonstrated in the LHC operation for reducing the beta-beat was first tested in RHIC during its polarized proton operation in 2011. It was then fully implemented during the RHIC polarized proton operation in 2012. This paper reports the commissioning results. Future plan is also presented. 14. Benchmarking of collimation tracking using RHIC beam loss data. Robert-Demolaize,G.; Drees, A. 2008-06-23 State-of-the-art tracking tools were recently developed at CERN to study the cleaning efficiency of the Large Hadron Collider (LHC) collimation system. In order to estimate the prediction accuracy of these tools, benchmarking studies can be performed using actual beam loss measurements from a machine that already uses a similar multistage collimation system. This paper reviews the main results from benchmarking studies performed with specific data collected from operations at the Relativistic Heavy Ion Collider (RHIC). 15. ALL-FERRITE RHIC INJECTION KICKER HAHN,H.; FISCHER,W.; PTITSYN,V.I.; TUOZZOLO,J.E. 2001-06-18 Ion beams are transferred from the AGS into RHIC in boxcar fashion as single bunches. The nominal design assumes 60 bunches per ring but increasing the number of bunches to gain luminosity is possible, thereby requiring injection kickers with a shorter rise time. The original injection system consists of traveling-wave dielectric loaded kicker magnets and a Blumlein pulser with a rise time adequate for the present operation. Voltage breakdown in the dielectric kickers suggested the use of all-ferrite magnets. In order to minimize the conversion cost, the design of the all-ferrite kicker uses the same components as the dielectric loaded units. The all-ferrite kickers showed in bench measured good breakdown properties and a current rise time of < 50 ns. A prototype kicker has been installed in the blue ring and was tested with beam. Beam measurements indicate suitability of all-ferrite kicker magnets for upgraded operation. 16. RHIC AC DIPOLE DESIGN AND CONSTRUCTION. BAI,M.; METH,M.; PAI,C.; PARKER,B.; PEGGS,S.; ROSER,T.; SANDERS,R.; TRBOJEVIC,D.; ZALTSMAN,A. 2001-06-18 Two ac dipoles with vertical and horizontal magnetic field have been proposed at RHIC for applications in linear and non-linear beam dynamics and spin manipulations. A magnetic field amplitude of 380 Gm is required to produce a coherent oscillation of 5 times the rms beam size at the top energy. We take the ac dipole frequency to be 1.0% of the revolution frequency away from the betatron frequency. To achieve the strong magnetic field with minimum power loss, an air-core magnet with two seven turn winding of low loss Litz wire resonating at 64 kHz is designed. The system is also designed to allow one to connect the two magnet winding in series to resonate at 37 kHz for the spin manipulation. Measurements of a half length prototype magnet are also presented. 17. COUPLING MEASUREMENT AND CORRECTION AT RHIC. PILAT,F.; BEEBE-WANG,J.; FISCHER,W.; PTITSYN,V.; SATOGATA,T. 2002-06-02 Coupling correction at RHIC has been operationally achieved through a two-step process: using local triplet skew quadrupoles to compensate coupling corn rolled low-beta triplet quadrupoles, and minimizing the tune separation and residual coupling with orthogonal global skew quadrupole families. An application has been developed for global correction that allows skew quadrupole tuning and tune display with a choice of different tune measurement techniques, including tune-meter, Schottky and phase lock loop (PLL). Coupling effects have been analysed by using 1024-turn (TBT) information from the beam position monitor (BPM) system. These data allow the reconstruction of the off-diagonal terms of the transfer matrix, a measure of global coupling. At both injection and storage energies, coordination of tune meter kicks with TBT acquisition at 322 BPM's in each ring allows the measurement of local coupling at all BPM locations. 18. High-energy high-luminosity electron-ion collider eRHIC Litvinenko, V.N.; Ben-Zvi, I.; Hammons, L.; Hao, Y.; Webb, S.; et al 2011-08-09 19. Unruh gamma radiation at RHIC? Biro, T S; Schram, Z 2011-01-01 Varying the proposition that acceleration itself would simulate a thermal environment, we investigate the semiclassical photon radiation as a possible telemetric thermometer of accelerated charges. Based on the classical Jackson formula we obtain the equivalent photon intensity spectrum stemming from a constantly accelerated charge and demonstrate its resemblances to a thermal distribution for high transverse momenta. The inverse transverse slope differs from the famous Unruh temperature: it is larger by a factor of pi. We compare the resulting direct photon spectrum with experimental data for AuAu collisions at RHIC and speculate about further, analytically solvable acceleration histories. 20. CRYOGENIC MAGNETS Post, R.F.; Taylor, C.E. 1963-05-21 A cryogenic magnet coil is described for generating magnetic fields of the order of 100,000 gauss with a minimum expenditure of energy lost in resistive heating of the coil inductors and energy lost irreversibly in running the coil refrigeration plant. The cryogenic coil comprises a coil conductor for generating a magnetic field upon energization with electrical current, and refrigeration means disposed in heat conductive relation to the coil conductor for cooling to a low temperature. A substantial reduction in the power requirements for generating these magnetic fields is attained by scaling the field generating coil to large size and particular dimensions for a particular conductor, and operating the coil at a particular optimum temperature commensurate with minimum overall power requirements. (AEC) 1. Cryogenics; Criogenia Gutierrez R, C.; Jimenez D, J.; Cejudo A, J.; Hernandez M, V. [Instituto Nacional de Investigaciones Nucleares, A.P. 18-1027, 11801 Mexico D.F. (Mexico) 1997-07-01 Cryogenics is one of these technologies which contributes to scientific research that supports to the industry in the following benefits: 1. Storage ability and a great quantity of dense gases with cryogenic liquid which is found at high pressure. 2. Production ability at low cost with high purity gases through distillation or condensation. 3. Ability to use low temperatures in the refrigerating materials or alteration of the physical properties. This technology is used for reprocessing of those short and long half life radioactive wastes which always have been required that to be separated with classical methods. In this text we report the radioactive wastes separation by more sophisticated methods but more quickly and reliable. (Author) 2. The Control System for the Cryogenics in the LHC Tunnel [First Experience and Improvements Gomes, P; Casas, J; Fluder, C; Fortescue, E; Le Roux, P; Penacoba, G; Pezzetti, M; Soubiran, M; Tovar, A; Zwalinski, L 2010-01-01 The Large Hadron Collider (LHC) was commissioned at CERN and started operation with beams in 2008. Several months of operation in nominal cryogenic conditions have triggered an optimisation of the process functional analysis. This lead to a few revisions of the control logic, which were realised on-the-fly. During the 2008-09 shut-down, and in order to enhance the safety, availability and operability of the LHC cryogenics, a major rebuild of the logic and several hardware modifications were implemented. The databases, containing instruments and controls in-formation, are being rationalized; the automatic tool, that extracts data for the control software, is being simplified. This paper describes the main improvements and sug-gests perspectives of further developments. 3. Mechanical and Thermal Characteristics of Insulation Materials for the KSTAR Magnet System at Cryogenic Temperature Chung, Wooho; Lim, Bungsu; Kim, Myungkyu; Park, Hyunki; Kim, Keeman; Chu, Yong; Lee, Sangil 2004-06-01 The KSTAR(Korea Superconducting Tokamak Advanced Research) superconducting magnet is electrically insulated by the composite material of epoxy resin and glass fiber (2.5 kV/mm) and Kapton (8 kV/mm). The insulation composite material of epoxy resin and glass fiber is prepared using a VPI (Vacuum Pressure Impregnation) process. The superconducting magnet is under mechanical stress caused by the large temperature difference between the operation temperature of the magnet and room temperature. The large electro-magnetic force during the operation of the magnet is also exerted on the magnet. Therefore, the characteristics of the insulation material at cryogenic temperatures are very important and the tensile stress and thermal expansion coefficient for the insulation materials of the KSTAR superconducting magnet are measured. This paper presents results on mechanical properties of the insulation material for KSTAR magnets, such as density, ultimate tensile stress and thermal contraction between room temperature and cryogenic temperatures. 4. NASA Cryogenic Propellant Systems Technology Development and Potential Opportunities for Discussion Meyer, Michael L. 2015-01-01 Members of the eCryo Team are traveling to France to meet with CNES (Centre National d'Etudes Spatiales) on the benchmarking of CFM (Cryogenic Fluids Management) analytical models the week of January 26th, 2015. Mike Meyer is representing the Agency and eCryo Project and will conduct a conversation to explore future work. This slide package (28 charts and 3 movies) requires approval via a 1676. ISS data in this chart set has been copied from public websites. 5. Has the QCD critical point been observed at RHIC? Antoniou, N G; Diakonos, F K 2016-01-01 The experimental search for the location of the QCD critical point in the phase diagram is of primary importance. In a recent publication it is claimed that measurements at RHIC lead not only to the location of the critical point ($\\mu_{cep}=95$ MeV, $T_{cep}=165$ MeV) but also to the verification of its universality class ($3d$ Ising system) by extracting the values of the critical exponents ($\\gamma=1.2$, $\ 6. Modelling and Model-Based-Designed PID Control of the JT-60SA Cryogenic System Using the Simcryogenics Library Bonne, F.; Bonnay, P.; Hoa, C.; Mahoudeau, G.; Rousset, B. 2017-02-01 This papers deals with the Japan Torus-60 Super Advanced fusion experiment JT-60SA cryogenic system. A presentation of the JT-60SA cryogenic system model, from 300K to 4.4K -using the Matlab/Simulink/Simscape Simcryogenics library- will be given. As a first validation of our modelling strategy, the obtained operating point will be compared with the one obtained from HYSYS simulations. In the JT60-SA tokamak, pulsed heat loads are expected to be coming from the plasma and must be handled properly, using both appropriate refrigerator architecture and appropriate control model, to smooth the heat load. This paper presents model-based designed PID control schemes to control the helium mass inside the phase separator. The helium mass inside the phase separator as been chosen to be the variable of interest in the phase separator since it is independent of the pressure which can vary from 1 bar to 1.8 bar during load smoothing. Dynamics simulations will be shown to assess the legitimacy of the proposed strategy. This work is partially supported through the French National Research Agency (ANR), task agreement ANR-13-SEED-0005. 7. Synchrotron Radiation in eRHIC Interaction Region Beebe-Wang, Joanne; Montag, Christoph; Rondeau, Daniel J; Surrow, Bernd 2005-01-01 The eRHIC currently under study at BNL consists of an electron storage ring added to the existing RHIC complex. The interaction region of this facility has to provide the required low-beta focusing while accommodating the synchrotron radiation generated by beam separation close to the interaction point. In the current design, the synchrotron radiation caused by 10GeV electrons bent by low-beta triplet magnets will be guided through the interaction region and dumped 5m downstream. However, it is unavoidable to stop a fraction of the photons at the septum where the electron and ion vacuum system are separated. In order to protect the septum and minimize the backward scattering of the synchrotron radiation, an absorber and collimation system will be employed. In this paper, we first present the overview of the current design of the eRHIC interaction region with special emphasis on the synchrotron radiation. Then the initial design of the absorber and collimation system, including their geometrical and physical p... 8. Matter in extremis: Ultrarelativistic nuclear collisions at RHIC Jacobs, Peter; Wang, Xin-Nian 2004-08-20 We review the physics of nuclear matter at high energy density and the experimental search for the Quark-Gluon Plasma at the Relativistic Heavy Ion Collider (RHIC). The data obtained in the first three years of the RHIC physics program provide several lines of evidence that a novel state of matter has been created in the most violent, head-on collisions of Au nuclei at {radical}s = 200 GeV. Jet quenching and global measurements show that the initial energy density of the strongly interacting medium generated in the collision is about two orders of magnitude larger than that of cold nuclear matter, well above the critical density for the deconfinement phase transition predicted by lattice QCD. The observed collective flow patterns imply that the system thermalizes early in its evolution, with the dynamics of its expansion consistent with ideal hydrodynamic flow based on a Quark-Gluon Plasma equation of state. 9. Control mechanism for attenuation of thermal energy pulses using cold circulators in the cryogenic distribution system of fusion devices in tokamak configuration Bhattacharya, R.; Sarkar, B.; Vaghela, H.; Shah, N. 2014-01-01 Operation and control of superconducting (SC) magnets in the fusion devices having tokamak configuration opens up the domain of varying peak thermal energy environment as a function of time, commensurate with the plasma pulses. The varied thermal energy environment, thus propagated to upstream of the cooling system, is responsible for the system level instability of the overall cryogenic system. The cryogenic distribution system, the regime of first impact point, therefore, has to be tuned so as to stay at the nearly stable zone of operation. The configuration of the cryogenic distribution system, considered in the present study, involves a liquid helium (LHe) bath as a thermal buffer, LHe submerged heat exchangers and cold circulator apart from the valves for implementations of the precise controls. The cold circulator supplies the forced flow supercritical helium, used for the cooling of SC magnets. The transients of the thermal energy pulses can be attenuated in the cryogenic distribution system by various methodologies. One of the adopted methodologies in the present study is with the precise speed control of the cold circulators. The adopted methodology is applied to various configurations of arrangements of internal components in the distribution system for obtaining system responses with superior attenuation of energy pulses. The process simulation approach, assumptions, considered inputs and constraints, process modeling with different configuration as well as results to accomplish the control scheme for the attenuation of the thermal energy pulses are described. 10. A Piezoelectric Cryogenic Heat Switch Jahromi, Amir E.; Sullivan, Dan F. 2014-01-01 We have measured the thermal conductance of a mechanical heat switch actuated by a piezoelectric positioner, the PZHS (PieZo electric Heat Switch), at cryogenic temperatures. The thermal conductance of the PZHS was measured between 4 K and 10 K, and on/off conductance ratios greater than 100 were achieved when the positioner applied its maximum force of 8 N. We discuss the advantages of using this system in cryogenic applications, and estimate the ultimate performance of an optimized PZHS. 11. Spin physics at RHIC: Present and future Abhay Deshpande 2003-11-01 In 2001–2002 the relativistic heavy-ion collider (RHIC) at the Brookhaven National Laboratory (BNL) was first commissioned for polarized proton collisions. Polarized protons were injected into the RHIC, accelerated to 100 GeV, stored and the two beams were made to collide in four interaction regions. I will review the progress made by the RHIC spin program, followed by the physics goals for the next few years. After that I will present a brief overview of a proposal to build a high intensity polarized electron/positron beam facility at BNL which would enable deep inelastic scattering (DIS) experiments to be pursued at BNL by its collisions with the RHIC hadron beams. 12. Experts dismiss doomsday scenarios for RHIC Levi, B G 2000-01-01 A panel of particle physicists examining the possibility that operation of RHIC could generate blackholes or 'strangelets' which would consume ordinary matter, have declared that such scenarios are 'firmly excluded' (1 p). 13. Polarization transmission at RHIC, numerical simulations Meot F.; Bai, M.; Liu, C.; Minty, M.; Ranjbar, V. 2012-05-20 Typical tracking simulations regarding the transmission of the polarization in the proton-proton collider RHIC are discussed. They participate in general studies aimed at understanding and improving polarization performances during polarized proton-proton runs. 14. GLOBAL DECOUPLING ON THE RHIC RAMP. LUO, Y.; CAMERON, P.; DELLA PENNA, A.; FISCHER, W.; ET AL. 2005-05-16 The global betatron decoupling on the ramp is an important issue for the operation of the Relativistic Heavy Ion Collider (RHIC), especially in the RHIC polarized proton (pp) run. To avoid the major betatron and spin resonances on the ramp, the betatron tunes are constrained. And the rms value of the vertical closed orbit should be smaller than 0.5mm. Both require the global coupling on the ramp to be well corrected. Several ramp decoupling schemes were found and tested at RHIC, like N-turn map decoupling, three-ramp correction, coupling amplitude modulation, and coupling phase modulation. In this article, the principles of these methods are shortly reviewed and compared. Among them, coupling angle modulation is a robust and fast one. It has been applied to the global decoupling in the routine RHIC operation. 15. RHIC heavy ion operations performance Satogata, T; Ferrone, R; Pilat, F 2006-01-01 The Relativistic Heavy Ion Collider (RHIC) completed its fifth year of operation in 2005, colliding copper ion beams with ps=200 GeV/u and 62.4 GeV/u[1]. Previous heavy ion runs have collided gold ions at ps=130 GeV/u, 200 GeV/u, and 62.4 GeV/u[2], and deuterons and gold ions at ps=200 GeV/u[3]. This paper discusses operational performance statistics of this facility, including Cu- Cu delivered luminosity, availability, calendar time spent in physics stores, and time between physics stores. We summarize the major factors affecting operations efficiency, and characterize machine activities between physics stores. 16. ANALYSIS OF ELECTRON CLOUD AT RHIC. IRISO,U.; BLASKIEWICZ,M.; CAMERON,P.; DREES,A.; FISCHER,W.; ET AL. 2004-07-05 Pressure rises with high intense beams are among the main luminosity limitations at RHIC. Observations during the latest runs show beam induced electron multipacting as one of the causes for these pressure rises. Experimental studies are carried out at RHIC using devoted instrumentation to understand the mechanism leading to electron clouds. In the following, we report the experimental electron cloud data and the analyzed results using computer simulation codes. 17. Results from STAR experiment at RHIC Bedangadas Mohanty; STAR Collaboration 2006-11-01 We present some of the important experimental results from nucleus–nucleus collision studies carried out by the STAR experiment at Relativistic Heavy Ion Collider (RHIC). The results suggests that central Au+Au collisions at RHIC has produced a dense and rapidly thermalizing matter with initial energy densities above the critical values predicted by lattice QCD for establishment of a quark-gluon plasma (QGP). 18. Cryogenics and the Human Exploration of Mars Salerno, Louis J.; Kittel, Peter; Rasky, Daniel J. (Technical Monitor) 1997-01-01 Current plans within NASA involve extending the human exploration of space from low earth orbit into the solar system, with the first human exploration of Mars presently planned in 2011. Integral to all hum Mars mission phases is cryogenic fluid management. Cryogenic fluids will be required both as propellant and for In-Situ Resource Utilization (ISRU). Without safe and efficient cryogen storage human Mars missions will not be possible. Effective control and handling of cryogenic fluids is the key to affordable Mars missions, and advancing active thermal control technology is synergistic with all of NASA's exploration initiatives and with existing and future instrument cooling programs, including MTPE and Origins. Present mission scenarios for human exploration require cryogenic propellant storage for up to 1700 days and for up to 60 metric tons. These requirements represent increases of an order of magnitude over previous storage masses and lifetimes. The key cryogenic terminology areas to be addressed in human Mars missions are long-term propellant storage, cryogenic refrigeration, cryogenic liquefaction, and zero gravity fluid management. Long-term storage for the thermal control of cryogenic propellants is best accomplished with a mix of passive and active technologies. Passive technologies such as advanced multilayer insulation (MLI) concepts will be combined with the development of active coolers (cryogenic refrigerators). Candidates for long-life active cooling applications include Reverse Turbo-Brayton, Stirling, and Pulse-Tube coolers. The integration of passive and active technologies will form a hybrid system optimized to minimize the launch mass while preserving the cryogenic propellants. Since cryogenic propellants are the largest mass that Mars missions must launch from earth, even a modest reduction in the percentage of propellant carried results in a significant weight saving. This paper will present a brief overview of cryogenic fluid management 19. Continuous-Reading Cryogen Level Sensor Barone, F. E.; Fox, E.; Macumber, S. 1984-01-01 Two pressure transducers used in system for measuring amount of cryogenic liquid in tank. System provides continuous measurements accurate within 0.03 percent. Sensors determine pressure in liquid and vapor in tank. Microprocessor uses pressure difference to compute mass of cryogenic liquid in tank. New system allows continuous sensing; unaffected by localized variations in composition and density as are capacitance-sensing schemes. 20. A luminosity model of RHIC gold runs Zhang, S.Y. 2011-11-01 In this note, we present a luminosity model for RHIC gold runs. The model is applied to the physics fills in 2007 run without cooling, and with the longitudinal cooling applied to one beam only. Having good comparison, the model is used to project a fill with the longitudinal cooling applied to both beams. Further development and possible applications of the model are discussed. To maximize the integrated luminosity, usually the higher beam intensity, smaller longitudinal and transverse emittance, and smaller {beta} are the directions to work on. In past 10 years, the RHIC gold runs have demonstrated a path toward this goal. Most recently, a successful commissioning of the bunched beam stochastic cooling, both longitudinal and transverse, has offered a chance of further RHIC luminosity improvement. With so many factors involved, a luminosity model would be useful to identify and project gains in the machine development. In this article, a preliminary model is proposed. In Section 2, several secondary factors, which are not yet included in the model, are identified based on the RHIC operation condition and experience in current runs. In Section 3, the RHIC beam store parameters used in the model are listed, and validated. In Section 4, the factors included in the model are discussed, and the luminosity model is presented. In Section 5, typical RHIC gold fills without cooling, and with partial cooling are used for comparison with the model. Then a projection of fills with more coolings is shown. In Section 6, further development of the model is discussed. 1. A fast framing camera system for observation of acceleration and ablation of cryogenic hydrogen pellet in ASDEX Upgrade plasmas Kocsis, G.; Kálvin, S.; Veres, G.; Cierpka, P.; Lang, P. T.; Neuhauser, J.; Wittman, C.; ASDEX Upgrade Team 2004-11-01 An observation system using fast digital cameras was developed to measure a cryogenic hydrogen pellet's cloud structure, trajectory, and velocity changes during its ablation in ASDEX Upgrade plasmas. In this article the system, the applied numerical methods, and the results are presented. The three-dimensional pellet trajectory and velocity components were reconstructed from images of observations from two different directions. Pellet acceleration both in the radial and toroidal directions was detected. The pellet cloud distribution was measured with high spatio-temporal resolution. The cloud surrounding the pellet was found to be elongated along the magnetic field lines. Its typical size is 5-7 cm along the field lines and 2 cm in the perpendicular directions. A cloud extension in the poloidal direction was also observed which may be related to the drift of the detached part of the cloud. 2. Optimization and testing of the Beck Engineering free-piston cryogenic pump for LNG systems on heavy vehicles. Final technical report Beck, Douglas S. 2003-01-10 Task 7 was completed by reaching Milestone 7: Test free piston cryogenic pump (FPCP) in Integrated LNG System. Task 4: Alternative Pump Design was also completed. The type of performance of the prototype LNG system is consistent with requirements of fuel systems for heavy vehicles; however, the maximum flow capacity of the prototype LNG system is significantly less than the total flow requirement. The flow capacity of the prototype LNG system is determined by a cavitation limit for the FPCP. 3. The evolution of cryogenic safety at Fermilab Stanek, R.; Kilmer, J. 1992-12-01 Over the past twenty-five years, Fermilab has been involved in cryogenic technology as it relates to pursuing experimentation in high energy physics. The Laboratory has instituted a strong cryogenic safety program and has maintained a very positive safety record. The solid commitment of management and the cryogenic community to incorporating safety into the system life cycle has led to policies that set requirements and help establish consistency for the purchase and installation of equipment and the safety analysis and documentation. 4. Transverse and longitudinal dynamics at RHIC Staszel, P.; BRAHMS Collaboration 2008-04-01 We review results obtained by the BRAHMS experiment at the Relativistic Heavy Ion Collider (RHIC) for the systems of Au+Au and p+p colliding at \\rootsnn{200} and at \\rootsnn{62.4} . Rapidity-dependent \\barp/\\pi^- and K±/π± ratios within 0 < y < 3 for Au+Au at \\rootsnn{200} are found to be enhanced in nucleus-nucleus collisions as compared to p+p collisions. The particle ratios are discussed in terms of their system size and rapidity dependence. From comparison of RAA for different systems and energies it is found that RAA increases with decreasing collision energy, decreasing system size and when going toward more peripheral collisions. However, RAA shows only a very weak dependence on rapidity (for 0 < y < 3.2), both for pions and protons. The \\barp/p ratio for p+p collisions at \\roots{200} is below 0.1 at high pT (≈4 GeV/c) and y ≈ 3. Such a strong asymmetry in p and \\barp production cannot be described within next-to-leading order perturbative QCD utilizing any known sets of FFs. At the same rapidity but lower energy (\\roots{62.4}) the charge asymmetry is stronger by an order of magnitude for both protons and kaons. 5. Cryogenic treatment of gas Bravo, Jose Luis [Houston, TX; Harvey, III, Albert Destrehan; Vinegar, Harold J [Bellaire, TX 2012-04-03 Systems and methods of treating a gas stream are described. A method of treating a gas stream includes cryogenically separating a first gas stream to form a second gas stream and a third stream. The third stream is cryogenically contacted with a carbon dioxide stream to form a fourth and fifth stream. A majority of the second gas stream includes methane and/or molecular hydrogen. A majority of the third stream includes one or more carbon oxides, hydrocarbons having a carbon number of at least 2, one or more sulfur compounds, or mixtures thereof. A majority of the fourth stream includes one or more of the carbon oxides and hydrocarbons having a carbon number of at least 2. A majority of the fifth stream includes hydrocarbons having a carbon number of at least 3 and one or more of the sulfur compounds. 6. Cryogenic treatment of gas Bravo, Jose Luis [Houston, TX; Harvey, III, Albert Destrehan (Kingwood, TX); Vinegar, Harold J [Bellaire, TX 2012-04-03 Systems and methods of treating a gas stream are described. A method of treating a gas stream includes cryogenically separating a first gas stream to form a second gas stream and a third stream. The third stream is cryogenically contacted with a carbon dioxide stream to form a fourth and fifth stream. A majority of the second gas stream includes methane and/or molecular hydrogen. A majority of the third stream includes one or more carbon oxides, hydrocarbons having a carbon number of at least 2, one or more sulfur compounds, or mixtures thereof. A majority of the fourth stream includes one or more of the carbon oxides and hydrocarbons having a carbon number of at least 2. A majority of the fifth stream includes hydrocarbons having a carbon number of at least 3 and one or more of the sulfur compounds. 7. CRYOGENIC DEWAR Chamberlain, W.H.; Maseck, H.E. 1964-01-28 This patent relates to a dewar for storing cryogenic gase and is of the type having aii inner flask surrounded by a vacuum jacket and having a vent spout through which evaporating gas escapes. Heretofore substantial gas loss has resulted from the radiation of heat towards the flask from the warmer outer elements of the dewar. In this invention, the mask is surrounded by a thermally conducting shield which is disposed in the vacuum space between the flask and the outer elements of the dewar. The shield contacts only the vent spout, which is cooled by the evaporating gas, and thus is maintained at a temperature very close to that of the flask itself. Accordingly, heat radiated toward the flask is intercepted and conducted to the evaporating gas rather than being re-radiated towards the hask. In a liquid helium dewar of typical configniration the mention reduces the boil-off rate by approximately one-half.(AEC) 8. Advances in Helium Cryogenics Sciver, S. W. Van This review provides a survey of major advances that have occurred in recent years in the area of helium cryogenics. Helium-temperature cryogenics is the enabling technology for a substantial and growing number of low-temperature systems from superconducting magnets to space-based experimental facilities. In recent years there have been many advances in the technology of low-temperature helium, driven mostly by new applications. However, to keep the review from being too broad, this presentation focuses mainly on three of the most significant advances. These are: (1) the development of large-scale recuperative refrigeration systems mainly for superconducting magnet applications in accelerators and other research facilities; (2) the use of stored superfluid helium (He II) as a coolant for spacebased astrophysics experiments; and (3) the application of regenerative cryocoolers operating at liquid helium temperatures primarily for cooling superconducting devices. In each case, the reader should observe that critical technologies were developed to facilitate these applications. In addition to these three primary advances, other significant helium cryogenic technologies are briefly reviewed at the end of this chapter, along with some vision for future developments in these areas. 9. Cryogenic design of liquid-nitrogen circulation system for long-length HTS cables with altitude variation Chang, Ho-Myung; Ryu, Ki Nam; Yang, Hyung Suk 2017-04-01 Liquid-nitrogen (LN2) circulation systems with altitude variation are investigated and designed for application to practical long-length HTS cables. This study is motivated by the KEPCO's immediate plan to install new HTS cables in existing utility tunnels including inclined and vertical sections. The distribution of pressure and temperature along the LN2 circulation loop is examined for various geographic conditions, taking into account the gravitational effect of altitude variation. The cryogenic cooling requirements are defined in terms of the pressure and temperature of LN2, and a design process is graphically demonstrated on phase diagram. It is concluded that the LN2 flow rate along the cable with altitude variation should be carefully determined with the constraints on pressure and temperature, and the proposed graphical method is useful in the design. 10. Manufacture of Novel Cryogenic Thermal Protection Materials Project National Aeronautics and Space Administration — Advanced Materials Technology, Inc (AMTI) responds to the NASA SBIR solicitation X8 "Space Cryogenic Systems" under subtopic X8.01, "Cryogenic Fluid Transfer and... 11. 大型超流氦低温冷却系统的研究进展%DEVELOPMENT OF LARGE SCALE CRYOGENIC SUPERFLUID HELIUM REFRIGERATION SYSTEM 何超峰; 郁欢强; 孙兴中; 陈耀锋; 武义锋; 周家屹; 张华标 2016-01-01 低温超导技术在基础科学研究中的广泛应用,极大的带动了低温工程的技术发展。通过介绍超流氦在大型低温系统中的应用优势,国内外几个典型超流氦低温冷却系统的流程、性能指标以及运行情况。分析了大型超流氦低温冷却系统的主要设备构成、热力循环方案以及关键设备、关键技术突破方向。%While the applications of low temperature superconducting technology in basic scientific research spread widely,the development of cryogenic engineering was greatly accelerated,and lots of large helium cryogenic systems were established successively. The processes,performances and operating conditions of several typical superfluid helium cryogenic systems were described in detail. The advantages of the application of the super flow helium in the large-scale cryogenic system are introduced and the main equipment composition,the thermodynamic cycle scheme and the key equipment,the key technology breakthrough direction are analyzed in this paper. 12. Medium energy heavy ion operations at RHIC Drees, K.A.; Ahrens, L.; Bai, M.; Beebe-Wang, J.; Blackler, I.M.C.; Blaskiewicz, M.; Brown, K.A.; Brennan, M.; Bruno, D.; Butler, J.; Carlson, C.; Connolly, R.; D' Ottavio, T.; Fischer, W.; Fu, W.; Gassner, D.; Harvey, M.; Hayes, T.; Huang, H.; Hulsart, R.; Ingrassia, P.; Kling, N.; Lafky, M.; Laster, J.; Lee, R.C.; Litvinenko, V.; Luo, Y.; MacKay, W.W.; Marr, G.; Mapes. M.; Marusic, A.; Mernick, K.; Michnoff, R.; Minty, M.; Montag, C.; Morris, J.; Naylor, C.; Nemesure, S.; Pilat, F.; Ptitsyn, V.; Robert-Demolaize, G.; Roser, T.; Sampson, P.; Satogata, T.; Schoefer, V.; Schultheiss, C.; Severino, F.; Shrey, T.; Smith, K.S.; Tepikian, S.; Thieberger, P.; Trbojevic, D.; Tsoupas, N.; Tuozzolo, J.; van Kuik, B.; Wilinski, M.; Zaltsman, A.; Zeno, K.; Zhang, S.Y. 2011-03-28 As part of the search for a phase transition or critical point on the QCD phase diagram, an energy scan including 5 different energy settings was performed during the 2010 RHIC heavy ion run. While the top beam energy for heavy ions is at 100 GeV/n and the lowest achieved energy setpoint was significantly below RHICs injection energy of approximately 10 GeV/n, we also provided beams for data taking in a medium energy range above injection energy and below top beam energy. This paper reviews RHIC experience and challenges for RHIC medium energy operations that produced full experimental data sets at beam energies of 31.2 GeV/n and 19.5 GeV/n. The medium energy AuAu run covered two beam energies, both above the RHIC injection energy of 9.8 GeV but well below the standard store energy of 100 GeV (see table 1). The low energy and full energy runs with heavy ions in FY10 are summarized in [1] and [2]. Stochastic Cooling ([3]) was only used for 100 GeV beams and not used in the medium energy run. The efficiency of the transition from 100 GeV operation to 31.2 GeV and then to 19.5 GeV was remarkable. Setup took 32 h and 19 h respectively for the two energy settings. The time in store, defined to be the percentage of time RHIC provides beams in physics conditions versus calendar time, was approximately 52% for the entire FY10 heavy ion run. In both medium energy runs it was well above this average, 68% for 31.5 GeV and 82% for 19.5 GeV. For both energies RHIC was filled with 111 bunches with 1.2 10{sup 9} and 1.3 10{sup 9} ions per bunch respectively. 13. Design of a simple cryogenic system for ultraviolet-visible absorption spectroscopy with a back-reflectance fiber optic probe. Vinyard, Andrew; Hansen, Kaj A; Byrd, Ross; Stuart, Douglas A; Hansen, John E 2014-01-01 We report a convenient and inexpensive technique for the rapid acquisition of absorption spectra from small samples at cryogenic temperatures using a home built cryostat with novel collection optics. A cylindrical copper block was constructed with a coaxial bore to hold a 4.00 mm diameter electron paramagnetic resonance (EPR) tube and mounted on a copper feed in thermal contact with liquid nitrogen. A 6.35 mm diameter hole was bored into the side of the cylinder so a fiber optic cable bundle could be positioned orthogonally to the EPR tube. The light passing through the sample is reflected off of the opposing surfaces of the EPR tube and surrounding copper, back through the sample. The emergent light is then collected using the fiber optic bundle and analyzed using a dispersive spectrometer. Absorption spectra for KMnO4 were measured between 400 and 700 nm. Absorption intensity at 506, 525, 545, and 567 nm was found to be proportional to concentration, displaying Beer's law-like behavior. The EPR tube had an internal diameter of 3.2 mm; the double pass of the probe beam through the sample affords a central path length of about 6.4 mm. Comparing these measurements with those recorded on a conventional tabletop spectrometer using a cuvette with a 10.00 mm path length, we consistently found a ratio between intensities of 0.58 rather than the anticipated 0.64. These 6% smaller values we attribute to the curvature of the EPR tube and transmission/reflection losses. This system is particularly well-suited to studying the kinetics and dynamics of chemical reactions at cryogenic temperatures. The rapid response (100 ms) and multiplex advantage provided the opportunity of recording simultaneous time courses at several wavelengths following initiation of a chemical reaction with a pulsed laser source. 14. Highlights from BNL-RHIC-2012 Tannenbaum, M J 2013-01-01 Recent highlights from Brookhaven National Laboratory and the Relativistic Heavy Ion Collider (RHIC) are reviewed and discussed in the context of the discovery of the strongly interacting Quark Gluon Plasma (sQGP) at RHIC in 2005 as confirmed by results from the CERN-LHC Pb+Pb program. Outstanding RHIC machine operation in 2012 with 3-dimensional stochastic cooling and a new EBIS ion source enabled measurements with Cu+Au, U+U, for which multiplicity distributions are shown, as well as with polarized p-p collisions. Differences of the physics and goals of p-p versus A+A are discussed leading to a review of RHIC results on pi0 suppression in Au+Au collisions and comparison to LHC Pb+Pb results in the same range 5 30 GeV. Improved measurements of direct photon production and correlation with charged particles at RHIC are shown, including the absence of a low pT (thermal) photon enhancement in d+Au collisions. Attempts to understand the apparent equality of the energy loss of light and heavy quarks in the QGP by... 15. RHIC PERFORMANCE DURING THE FY10 200 GeV Au+Au HEAVY ION RUN Brown, K.A.; Ahrens, L.; Bai, M.; Beebe-Wang, J.; Blaskiewicz, M.; Brennan, J.; Bruno, D.; Carlson, C.; Connolly, R.; de Maria, R.; D’Ottavio, T.; Drees, A.; Fischer, W.; Fu, W.; Gardner, C.; Gassner, D.; Glenn, J.W.; Hao, Y.; Harvey, M.; Hayes, T.; Hoff, L.; Huang, H.; Laster, J.; Lee, R.; Litvinenko, V.; Luo, Y.; MacKay, W.; Marr, G.; Marusic, A.; Mernick, K.; Michnoff, R.; Minty, M.; Montag, C.; Morris, J.; Nemesure, S.; Oerter, B.; Pilat, F.; Ptitsyn, V.; Robert-Demolaize, G.; Roser, T.; Russo, T.; Sampson, P.; Sandberg, J.; Satogata, T.; Severino, F.; Schoefer, V.; Schultheiss, C.; Smith, K.; Steski, D.; Tepikian, S.; Theisen, C.; Thieberger, P.; Trbojevic, D.; Tsoupas, N.; Tuozzolo, J.; Wang, G.; Wilinski, M.; Zaltsman, A.; Zeno, K.; Zhang, S.Y. 2010-05-23 Since the last successful RHIC Au+Au run in 2007 (Run-7), the RHIC experiments have made numerous detector improvements and upgrades. In order to benefit from the enhanced detector capabilities and to increase the yield of rare events in the acquired heavy ion data a significant increase in luminosity is essential. In Run-7 RHIC achieved an average store luminosity of = 12 x 10{sup 26} cm{sup -2} s{sup -1} by operating with 103 bunches (out of 111 possible), and by squeezing to {beta}* = 0.85 m. This year, Run-10, we achieved = 20 x 10{sup 26} cm{sup -2} s{sup -1}, which put us an order of magnitude above the RHIC design luminosity. To reach these luminosity levels we decreased {beta}* to 0.75 m, operated with 111 bunches per ring, and reduced longitudinal and transverse emittances by means of bunched-beam stochastic cooling. In addition we introduced a lattice to suppress intra-beam scattering (IBS) in both RHIC rings, upgraded the RF control system, and separated transition crossing times in the two rings. We present an overview of the changes and the results of Run-10 performance. 16. Commissioning of the helium cryogenic system for the HIE- ISOLDE accelerator upgrade at CERN Delruelle, N; Leclercq, Y; Pirotte, O; Williams, L 2015-01-01 The High Intensity and Energy ISOLDE (HIE-ISOLDE) project is a major upgrade of the existing ISOLDE and REX-ISOLDE facilities at CERN. The most significant improvement will come from replacing the existing REX accelerating structure by a superconducting linear accelerator (SC linac) composed ultimately of six cryo-modules installed in series, each containing superconducting RF cavities and solenoids operated at 4.5 K. In order to provide the cooling capacity at all temperature levels between 300 K and 4.5 K for the six cryo-modules, an existing helium refrigerator, manufactured in 1986 and previously used to cool the ALEPH magnet during LEP operation from 1989 to 2000, has been refurbished, reinstalled and recommissioned in a dedicated building located next to the HIE-ISOLDE experimental hall. This helium refrigerator has been connected to a new cryogenic distribution line, consisting of a 30-meter long vacuum-insulated transfer line, a 2000-liter storage dewar and six interconnecting valve boxes, one for eac... 17. The RHIC polarized H- ion source Zelenski, A.; Atoian, G.; Raparia, D.; Ritter, J.; Steski, D. 2016-02-01 A novel polarization technique had been successfully implemented for the Relativistic Heavy Ion Collider (RHIC) polarized H- ion source upgrade to higher intensity and polarization. In this technique, a proton beam inside the high magnetic field solenoid is produced by ionization of the atomic hydrogen beam (from external source) in the He-gaseous ionizer cell. Further proton polarization is produced in the process of polarized electron capture from the optically pumped Rb vapor. The use of high-brightness primary beam and large cross sections of charge-exchange cross sections resulted in production of high intensity H- ion beam of 85% polarization. The source very reliably delivered polarized beam in the RHIC Run-2013 and Run-2015. High beam current, brightness, and polarization resulted in 75% polarization at 23 GeV out of Alternating Gradient Synchrotron (AGS) and 60%-65% beam polarization at 100-250 GeV colliding beams in RHIC. 18. A high performance DAC /DDS daughter module for the RHIC LLRF platform Hayes, T.; Harvey, M.; Narayan, G.; Severino, F.; Smith, K.S.; Yuan, S. 2011-03-28 The RHIC LLRF upgrade is a flexible, modular system. Output signals are generated by a custom designed XMC card with 4 high speed digital to analog (DAC) converters interfaced to a high performance field programmable gate array (FPGA). This paper discusses the hardware details of the XMC DAC board as well as the implementation of a low noise rf synthesizer with digital IQ modulation. This synthesizer also provides injection phase cogging and frequency hop rebucketing capabilities. A new modular RHIC LLRF system was recently designed and commissioned based on custom designed XMC cards. As part of that effort a high speed, four channel DAC board was designed. The board uses Maxim MAX5891 16 bit DACs with a maximum update rate of 600 Msps. Since this module is intended to be used for many different systems throughout the Collider Accelerator complex, it was designed to be as generic as possible. One major application of this DAC card is to implement digital synthesizers to provide drive signals to the various cavities at RHIC. Since RHIC is a storage ring with stores that typically last many hours, extremely low RF noise is a critical requirement. Synchrotron frequencies at RHIC range from a few hertz to several hundred hertz depending on the species and point in the acceleration cycle so close in phase noise is a major concern. The RHIC LLRF system uses the Update Link, a deterministic, high speed data link that broadcasts the revolution frequency and the synchronous phase angle. The digital synthesizers use this data to generate a properly phased analog drive signal. The synthesizers must also provide smooth phase shifts for cogging and support frequency shift rebucketing. One additional feature implemented in the FPGA is a digital waveform generator (WFG) that generates I and Q data pairs based on a user selected amplitude and phase profile as a function of time. 19. A Reference Guide for Cryogenic Properties of Materials Weisend, John G 2003-09-16 A thorough knowledge of the behavior of materials at cryogenic temperatures is critical for the design of successful cryogenic systems. Over the past 50 years, a tremendous amount of material properties at cryogenic temperatures have been measured and published. This guide lists resources for finding these properties. It covers online databases, computer codes, conference proceedings, journals, handbooks, overviews and monographs. It includes references for finding reports issued by government laboratories and agencies. Most common solids and fluids used in cryogenics are covered. 20. Cryogenic Propulsion Project National Aeronautics and Space Administration — The storage of cryogenic propellants is challenging because heat leaks into the cryogenic storage tanks no matter how good the insulation, resulting in a necessity... 1. Energy Efficient Cryogenics Meneghelli, Barry J.; Notardonato, William; Fesmire, James E. 2016-01-01 The Cryogenics Test Laboratory, NASA Kennedy Space Center, works to provide practical solutions to low-temperature problems while focusing on long-term technology targets for the energy-efficient use of cryogenics on Earth and in space. 2. Magnetic bearings for cryogenic turbomachines Iannello, Victor; Sixsmith, Herbert 1991-01-01 Magnetic bearings offer a number of advantages over gas bearings for the support of rotors in cryogenic turboexpanders and compressors. Their performance is relatively independent of the temperature or pressure of the process gas for a large range of conditions. Active magnetic bearing systems that use capacitive sensors have been developed for high speed compressors for use in cryogenic refrigerators. Here, the development of a magnetic bearing system for a miniature ultra high speed compressor is discussed. The magnetic bearing has demonstrated stability at rotational speeds exceeding 250,000 rpm. This paper describes the important features of the magnetic bearing and presents test results demonstrating its performance characteristics. 3. A prototype ionization profile monitor for RHIC Connolly, R.; Cameron, P.; Ryan, W. [and others 1997-07-01 Transverse beam profiles in the Relativistic Heavy-Ion Collider (RHIC) will be measured with ionization profile monitors (IPMs). Each IPM collects and measures the distribution of electrons in the beamline resulting from residual gas ionization during bunch passage. The electrons are swept transversely from the beamline and collected on strip anodes oriented parallel to the beam axis. At each bunch passage the charge pulses are amplified, integrated, and digitized for display as a profile histogram. A prototype detector was tested in the injection line during the RHIC Sextant Test. This paper describes the detector and gives results from the beam tests. 4. Summary of the RHIC Retreat 2008 Pilat,F.; Brennan, M.; Brown, K.; Fischer, W.; Montag, C. 2008-08-01 The main goal of the RHIC Retreat is to review last run's performance and prepare for the next. As always though we also discussed the longer term goals and plans for the facility to put the work in perspective and in the right priority. A straw-man plan for the facility was prepared for the DOE that assumes 30 cryoweek and running 2 species per year. The plan outlines RHIC operations for 2008-2012 and integrates well accelerator and detector upgrades to optimize the physics output with high luminosities. The plans includes guidance from the PAC and has been reviewed by DOE. 5. Global Decoupling on the RHIC Ramp Luo, Yun; Della Penna, Al; Fischer, Wolfram; Laster, Jonathan S; Marusic, Al; Pilat, Fulvia Caterina; Roser, Thomas; Trbojevic, Dejan 2005-01-01 The global betatron decoupling on the ramp is an important issue for the operation of the Relativistic Heavy Ion Collider (RHIC). In the polarized proton run, the betatron tunes are required to keep almost constant on the ramp to avoid spin resonance line crossing and the beam polarization loss. Some possible correction schemes on the ramp, like three-ramp correction, the coupling amplitude modulation and the coupling phase modulaxtion, have been found. The principles of these schemes are shortly reviewed and compared. Operational results of their applications on the RHIC ramps are given. 6. The Odderon at RHIC and LHC Nicolescu, Basarab 2007-01-01 The Odderon remains an elusive object, 33 years after its invention. The Odderon is now a fundamental object in QCD and CGC and it has to be found experimentally if QCD and CGC are right. In the present talk, we show how to find it at RHIC and LHC. The most spectacular signature of the Odderon is the predicted difference between the differential cross-sections for proton-proton and antiproton-proton at high s and moderate t. This experiment can be done by using the STAR detector at RHIC and b... 7. Cryogenic immersion microscope Le Gros, Mark; Larabell, Carolyn A. 2010-12-14 A cryogenic immersion microscope whose objective lens is at least partially in contact with a liquid reservoir of a cryogenic liquid, in which reservoir a sample of interest is immersed is disclosed. When the cryogenic liquid has an index of refraction that reduces refraction at interfaces between the lens and the sample, overall resolution and image quality are improved. A combination of an immersion microscope and x-ray microscope, suitable for imaging at cryogenic temperatures is also disclosed. 8. Design of the cryogenic systems for the Near and Far LAr-TPC detectors of the Short-Baseline Neutrino program (SBN) at Fermilab Geynisman, M. [Fermilab; Bremer, J. [CERN; Chalifour, M. [CERN; Delaney, M. [Fermilab; Dinnon, M. [Fermilab; Doubnik, R. [Fermilab; Hentschel, S. [Fermilab; Kim, M. J. [Fermilab; Montanari, C. [INFN, Pavia; Monatanari, D. [Fermilab; Nichols, T. [Fermilab; Norris, B. [Fermilab; Sarychev, M. [Fermilab; Schwartz, F. [Fermilab; Tillman, J. [Fermilab; Zuckerbrot, M. [Fermilab 2017-08-31 The Short-Baseline Neutrino (SBN) physics program at Fermilab and Neutrino Platform (NP) at CERN are part of the international Neutrino Program leading to the development of Long-Baseline Neutrino Facility/Deep Underground Neutrino Experiment (LBNF/DUNE) science project. The SBN program consisting of three Liquid Argon Time Projection Chamber (LAr-TPC) detectors positioned along the Booster Neutrino Beam (BNB) at Fermilab includes an existing detector known as MicroBooNE (170-ton LAr-TPC) plus two new experiments known as SBN’s Near Detector (SBND, ~260 tons) and SBN’s Far Detector (SBN-FD, ~760 tons). All three detectors have distinctly different design of their cryostats thus defining specific requirements for the cryogenic systems. Fermilab has already built two new facilities to house SBND and SBN-FD detectors. The cryogenic systems for these detectors are in various stages of design and construction with CERN and Fermilab being responsible for delivery of specific sub-systems. This contribution presents specific design requirements and typical implementation solutions for each sub-system of the SBND and SBN-FD cryogenic systems. 9. Design and test of the RHIC CMD10 abort kicker Hahn, H. [Brookhaven National Lab. (BNL), Upton, NY (United States); Blaskiewicz, M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Drees, A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Fischer, W. [Brookhaven National Lab. (BNL), Upton, NY (United States); Mi, J. [Brookhaven National Lab. (BNL), Upton, NY (United States); Meng, W. [Brookhaven National Lab. (BNL), Upton, NY (United States); Montag, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Pai, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Sandberg, J. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tsoupas, N. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tuozzolo, J. E. [Brookhaven National Lab. (BNL), Upton, NY (United States); Zhang, W. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-05-03 In recent RHIC operational runs, planned and unplanned pre-fire triggered beam aborts have been observed that resulted in quenches of SC main ring magnets, indicating a weakened magnet kick strength due to beam-induced ferrite heating. An improvement program was initiated to reduce the longitudinal coupling impedance with changes to the ferrite material and the eddy-current strip geometry. Results of the impedance measurements and of magnet heating tests with CMD10 ferrite up to 190°C are reported. All 10 abort kickers in the tunnel have been modified and were provided with a cooling system for the RUN 15. 10. Heat flux to the helium cryogenic system elements in the case of incidental vacuum vessel ventilation with atmospheric air CERN. Geneva 2016-01-01 The selection process for size in safety equipment for cold vessels or process pipes in cryogenic systems should take into consideration the incidental ventilation of the vacuum vessel with atmospheric air. In this case, a significant heat input toward the cold elements of the system can be expected. A number of experimental investigations have been done for the elements at liquid helium temperature which have been covered with 10 layers of MLI. The typical values of the heat flux were measured in a range of 3.7 to 5.0 kW/m2 of the element surface. The helium temperature parts are typically surrounded by thermal shields that are kept in a temperature range of 50-80K. On the external side, the thermal shields are covered with 30-40 layers of MLI while on the internal side, the shields are bare. The theoretical calculations of heat flux to the thermal shield, with respect to the possibility of air condensation and freezing on the bare side of the thermal shield, show that the heat flux to the thermal shield can... 11. Cryogenic system for the 43 T Hybrid Magnet at LNCMI Grenoble: from the needs to the commissioning Ronayette, L.; Crispel, S.; Berriaud, C.; Berthier, R.; Caplanne, G.; Gorski, M.; Graffin, P.; Hanoux, P.; Hergat, T.; Hervieu, B.; Juster, FP; Pfister, R.; Pissard, M.; Pugnat, P.; Vincent, B. 2017-02-01 LNCMI is one of the unique worldwide laboratories offering the scientific community access to various experimental conditions with continuous magnetic fields well above 20 T. LNCMI is currently developing a large field flexible experimental platform. One configuration will produce a continuous magnetic field of 43 T in a 34 mm warm bore aperture from the combination of homemade resistive electromagnet inserts and a large bore outer superconducting magnet (1.1 m internal cold dia.), the latter being built in close collaboration with CEA-IRFU Saclay. The superconducting magnet with its mechanical structure and its helium vessel will represent a mass of 22 tons to cool down to 1.8 K and maintain at this temperature 10 months per year. An overview of the project will be given focusing on the cryogenics and particularly on the helium liquefier designed and manufactured by Air Liquide Advanced Technologies. This system - the most powerful even produced in the Helial ML range - and its ancillaries has been integrated and commissioned as a turnkey system in the existing site of LNCMI. 12. Cryogenic heat transfer Barron, Randall F 2016-01-01 Cryogenic Heat Transfer, Second Edition continues to address specific heat transfer problems that occur in the cryogenic temperature range where there are distinct differences from conventional heat transfer problems. This updated version examines the use of computer-aided design in cryogenic engineering and emphasizes commonly used computer programs to address modern cryogenic heat transfer problems. It introduces additional topics in cryogenic heat transfer that include latent heat expressions; lumped-capacity transient heat transfer; thermal stresses; Laplace transform solutions; oscillating flow heat transfer, and computer-aided heat exchanger design. It also includes new examples and homework problems throughout the book, and provides ample references for further study. 13. Numerical modeling of self-pressurization and pressure control by a thermodynamic vent system in a cryogenic tank Majumdar, Alok; Valenzuela, Juan; LeClair, Andre; Moder, Jeff 2016-03-01 This paper presents a numerical model of a system-level test bed-the multipurpose hydrogen test bed (MHTB) using the Generalized Fluid System Simulation Program (GFSSP). MHTB is representative in size and shape of a space transportation vehicle liquid hydrogen propellant tank, and ground-based testing was performed at NASA Marshall Space Flight Center (MSFC) to generate data for cryogenic storage. GFSSP is a finite volume-based network flow analysis software developed at MSFC and used for thermofluid analysis of propulsion systems. GFSSP has been used to model the self-pressurization and ullage pressure control by the Thermodynamic Vent System (TVS). A TVS typically includes a Joule-Thompson (J-T) expansion device, a two-phase heat exchanger (HEX), and a mixing pump and liquid injector to extract thermal energy from the tank without significant loss of liquid propellant. For the MHTB tank, the HEX and liquid injector are combined into a vertical spray bar assembly. Two GFSSP models (Self-Pressurization and TVS) were separately developed and tested and then integrated to simulate the entire system. The Self-Pressurization model consists of multiple ullage nodes, a propellant node, and solid nodes; it computes the heat transfer through multilayer insulation blankets and calculates heat and mass transfer between the ullage and liquid propellant and the ullage and tank wall. A TVS model calculates the flow through a J-T valve, HEX, and spray and vent systems. Two models are integrated by exchanging data through User Subroutines of both models. Results of the integrated models have been compared with MHTB test data at a 50% fill level. Satisfactory comparison was observed between tests and numerical predictions. 14. Linear and chromatic optics measurements at RHIC Aiba, M.; Calaga, R.; Aiba, M.; Tomas, R.; Vanbavinkove, G. 2010-05-23 Measurements of chromatic beta-beating were carried out for the first time in the RHIC accelerator during Run 2009. The analysis package developed for the LHC was used to extract the off-momentum optics for injection and top energy. Results from the beam experiments and compassion to the optics model are presented. The primary goal of the RHIC experiments were execute an on-line measurement of the optics using the tools developed for the LHC. Turn-by-turn BPM trajectories (typically 1000 turns) acquired immediately after an external dipole kick are numerically analyzed to determine the optical parameters at the location of the beam position monitors (BPMs). For chromatic optics, a similar analysis, but on a beam with finite momentum offset(s). Each optical measurement typically is calculated from multiple data sets to capture statistical variations and ensure reproducibility. The procedure of measurement and analysis is detailed in ref [1, 2]. Two dedicated experiments were performed at RHIC with protons during Run 2009. The first at injection energy and optics and the other at 250 GeV and squeezed optics. The basic RHIC parameters relevant for the two experiments are listed in Table 1. 15. Hybrid helical snakes and rotators for RHIC Courant, E.D. 1995-06-13 The spin rotators and Siberian snakes presently envisaged for RHIC utilize helical dipole magnets. The snakes and the rotators each consist of four helices, each with a full twist (360{degrees}) of the field. Here we investigate an alternate layout, namely combinations of helical and pure bending magnet, and show that this may have advantages. 16. First Polarized Proton Collisions at RHIC Roser, T.; Ahrens, L.; Alessi, J.; Bai, M.; Beebe-Wang, J.; Brennan, J. M.; Brown, K. A.; Bunce, G.; Cameron, P.; Courant, E. D.; Drees, A.; Fischer, W.; Fliller, R.; Glenn, W.; Huang, H.; Luccio, A. U.; MacKay, W. W.; Makdisi, Y.; Montag, C.; Pilat, F.; Ptitsyn, V.; Satogata, T.; Tepikian, S.; Trbojevic, D.; Tsoupas, N.; van Zeijts, J.; Zelenski, A.; Zeno, K.; Deshpande, A.; Kurita, K.; Krueger, K.; Spinka, H.; Underwood, D.; Syphers, M.; Alekseev, I.; Svirida, D.; Ranjbar, V.; Tojo, J.; Jinnouchi, O.; Okamura, M.; Saito, N. 2003-05-01 We successfully injected polarized protons in both RHIC rings and maintained polarization during acceleration up to 100 GeV per ring using two Siberian snakes in each ring. Each snake consists of four helical superconducting dipoles which rotate the polarization by 180° about a horizontal axis. This is the first time that polarized protons have been accelerated to 100 GeV. 17. Heavy-flavour meson production at RHIC Mischke, A. 2010-01-01 Collisions of heavy atomic nuclei at very high beam energies allow to create and study hot QCD matter under laboratory-controlled conditions. Measurements at the SPS and RHIC facilities have yielded compelling evidence for the formation of this novel state of matter, the so-called Quark-Gluon Plasma 18. BEAM SCRUBBING FOR RHIC POLARIZED PROTON RUN. ZHANG,S.Y.FISCHER,W.HUANG,H.ROSER,T. 2004-07-05 One of the intensity limiting factor of RHIC polarized proton beam is the electron cloud induced pressure rise. A beam scrubbing study shows that with a reasonable period of time of running high intensity 112-bunch proton beam, the pressure rise can be reduced, allowing higher beam intensity. 19. Highlights from BNL and RHIC 2015 Tannenbaum, M J 2016-01-01 Highlights of news from Brookhaven National Laboratory (BNL) and results from the Relativistic Heavy Ion Collider (RHIC) in the period July 2014-June 2015 are presented. The news this year was mostly very positive. The major event at BNL was the startup and dedication of the new NSLS II, "the World's brightest Synchrotron Light Source". The operation of RHIC was outstanding with a polarized p+p run at$\\sqrt{s}=200$GeV with integrated luminosity that exceeded the sum of all previous p+p integrated luminosity at this$\\sqrt{s}$. For the first time at RHIC asymmetric p+Au and p+Al runs were made but the p+Al run caused damage in the PHENIX forward detectors from quenches that were inadequately shielded for this first p+A run. This was also the 10th anniversary of the 2005 announcement of the Perfect Liquid Quark Gluon Plasma at RHIC and a review is presented of the discoveries leading to this claim. A new result on net-charge fluctuations (with no particle identification) from PHENIX based on previous scans ov... 20. Copper coating specification for the RHIC arcs Blaskiewicz, M. 2010-12-01 Copper coating specifications for the RHIC arcs are given. Various upgrade scenarios are considered and calculations of resistive wall losses in the arcs are used to constrain the necessary quality and surface thickness of a copper coating. We find that 10 {mu}m of high purity copper will suffice. 1. Detectors for low energy electron cooling in RHIC Carlier, F. S. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2016-02-15 Low energy operation of RHIC is of particular interest to study the location of a possible critical point in the QCD phase diagram. The performance of RHIC at energies equal to or lower than 10 GV/nucleon is limited by nonlinearities, Intra-BeamScattering (IBS) processes and space-charge effects. To successfully address the luminosity and ion store lifetime limitations imposed by IBS the method of electron cooling has been envisaged. During electron cooling processes electrons are injected along with the ion beam at the nominal ion bunch velocities. The velocity spread of the ion beam is reduced in all planes through Coulomb interactions between the cold electron beam and the ion beam. The electron cooling system proposed for RHIC will be the first of its kind to use bunched beams for the delivery of the electron bunches, and will therefore be accompanied by the necessary challenges. The designed electron cooler will be located in IP2. The electron bunches will be accelerated by a linac before being injected along side the ion beams. Thirty consecutive electron bunches will be injected to overlap with a single ion bunch. They will first cool the yellow beam before being extracted turned by 180-degrees and reinjected into the blue beam for cooling. As such, both the yellow and blue beams will be cooled by the same ion bunches. This will pose considerable challenges to ensure proper electron beam quality to cool the second ion beam. Furthermore, no ondulator will be used in the electron cooler so radiative recombination between the ions and the electrons will occur. 2. Cryogenic technology for tracking detectors Granata, V; Watts, S; Borer, K; Janos, S; Pretzl, Klaus P; Dezillie, B; Li, Z; Casagrande, L; Collins, P; Grohmann, S; Heijne, Erik H M; Lourenço, C; Niinikoski, T O; Palmieri, V G; Sonderegger, P; Borchi, E; Bruzzi, Mara; Pirollo, S; Chapuy, S; Dimcovski, Zlatomir; Grigoriev, E; Bell, W; Devine, S R H; O'Shea, V; Ruggiero, G; Smith, K; Berglund, P; de Boer, Wim; Hauler, F; Heising, S; Jungermann, L; Abreu, M C; Rato-Mendes, P; Sousa, P; Cindro, V; Mikuz, M; Zavrtanik, M; Esposito, A P; Konorov, I; Paul, S; Buontempo, S; D'Ambrosio, D; Pagano, S; Eremin, V V; Verbitskaya, E 2001-01-01 A low-mass cryogenic cooling technique for silicon sensor modules has been developed in the framework of the RD39 Collaboration at CERN. A prototype low-mass beam tracker cryostat has been designed, constructed and tested for applications in fixed target experiments. We shall report here briefly the main features and results of the system. (2 refs). 3. USAF Space Sensing Cryogenic Considerations 2010-01-01 capacitance dilatometer for measuring thermal expansion and magnetostriction Rev. Sci. Instrum. 83, 095102 (2012) Compact radio-frequency resonator...enhancing when the refrigeration system is considered as part of an overall optimization problem. INTRODUCTION The use of cryogenics in space sensing 4. RELATIVISTIC HEAVY ION PHYSICS : RESULTS FROM AGS TO RHIC. STEINBERG,P. 2002-06-20 High-energy collisions of heavy ions provide a means to study QCD in a regime of high parton density, and may provide insight into its phme structure. Results from the four experiments at RHIC (BRAHMS, PHENIX, PHOBOS and STAR) are presented, and placed in context with the lower energy data from the AGS and SPS accelerators. The focus is on the insights these measurements provide into the time history of the collision process. Taken together, the data point to the creation of a deconfined state of matter that forms quickly, expands rapidly and freezes out suddenly. With the new RHIC data, systematic data now exists for heavy ion collisions as a function of {radical}s over several orders of magnitude and as a function of impact parameter. These data test the interplay between hard and soft processes in a large-volume system where nucleons are struck multiple times. The data is consistent with creating a deconfined state (jet quenching) that forms quickly (saturation models), expands rapidly (radial and elliptic flow) and freezes out suddenly (single freezeout and blast wave fits). There are also intriguing connections with particle production in elementary systems, which point to the role of the energy available for particle production on the features of the final state. Many in this field are optimistic that the careful understanding of this experimental data may lead t o the theoretical breakthroughs that will connect these complex systems to the fundamental lattice predict ions. 5. Implementation of Sub-Cooling of Cryogenic Propellants by Injection of Non-condensing Gas to the Generalized Fluid Systems Simulation Program (GFSSP) Huggett, Daniel J.; Majumdar, Alok 2013-01-01 Cryogenic propellants are readily heated when used. This poses a problem for rocket engine efficiency and effective boot-strapping of the engine, as seen in the "hot" LOX (Liquid Oxygen) problem on the S-1 stage of the Saturn vehicle. In order to remedy this issue, cryogenic fluids were found to be sub-cooled by injection of a warm non-condensing gas. Experimental results show that the mechanism behind the sub-cooling is evaporative cooling. It has been shown that a sub-cooled temperature difference of approximately 13 deg F below saturation temperature [1]. The phenomenon of sub-cooling of cryogenic propellants by a non-condensing gas is not readily available with the General Fluid System Simulation Program (GFSSP) [2]. GFSSP is a thermal-fluid program used to analyze a wide variety of systems that are directly impacted by thermodynamics and fluid mechanics. In order to model this phenomenon, additional capabilities had to be added to GFSSP in the form of a FORTRAN coded sub-routine to calculate the temperature of the sub-cooled fluid. Once this was accomplished, the sub-routine was implemented to a GFSSP model that was created to replicate an experiment that was conducted to validate the GFSSP results. 6. Numerical Modeling of an Integrated Vehicle Fluids System Loop for Pressurizing a Cryogenic Tank LeClair, A. C.; Hedayat, A.; Majumdar, A. K. 2017-01-01 This paper presents a numerical model of the pressurization loop of the Integrated Vehicle Fluids (IVF) system using the Generalized Fluid System Simulation Program (GFSSP). The IVF propulsion system, being developed by United Launch Alliance to reduce system weight and enhance reliability, uses boiloff propellants to drive thrusters for the reaction control system as well as to run internal combustion engines to develop power and drive compressors to pressurize propellant tanks. NASA Marshall Space Flight Center (MSFC) conducted tests to verify the functioning of the IVF system using a flight-like tank. GFSSP, a finite volume based flow network analysis software developed at MSFC, has been used to support the test program. This paper presents the simulation of three different test series, comparison of numerical prediction and test data and a novel method of presenting data in a dimensionless form. The paper also presents a methodology of implementing a compressor map in a system level code. 7. Survey of cryogenic semiconductor devices Talarico, L.J.; McKeever, J.W. 1996-04-01 Improved reliability and electronic performance can be achieved in a system operated at cryogenic temperatures because of the reduction in mechanical insult and in disruptive effects of thermal energy on electronic devices. Continuing discoveries of new superconductors with ever increasing values of T{sub c} above that of liquid nitrogen temperature (LNT) have provided incentive for developing semiconductor electronic systems that may also operate in the superconductors liquid nitrogen bath. Because of the interest in high-temperature superconductor (HTS) devices, liquid nitrogen is the cryogen of choice and LNT is the temperature on which this review is focused. The purpose of this survey is to locate and assemble published information comparing the room temperature (298 K), performance of commercially available conventional and hybrid semiconductor device with their performance at LNT (77K), to help establish their candidacy as cryogenic electronic devices specifically for use at LNT. The approach to gathering information for this survey included the following activities. Periodicals and proceedings were searched for information on the behavior of semiconductor devices at LNT. Telephone calls were made to representatives of semiconductor industries, to semiconductor subcontractors, to university faculty members prominent for their research in the area of cryogenic semiconductors, and to representatives of the National Aeronautics and Space Administration (NASA) and NASA subcontractors. The sources and contacts are listed with their responses in the introduction, and a list of references appears at the end of the survey. 8. Sprayable Thermal Insulation for Cryogenic Tanks Project National Aeronautics and Space Administration — The Sprayable Thermal Insulation for Cryogenic Tanks (STICT) is a thermal management system applied by either an automated or manual spraying process with less... 9. Sprayable Thermal Insulation for Cryogenic Tanks Project National Aeronautics and Space Administration — The innovation addressed in this proposal is Sprayable Thermal Insulation for Cryogenic Tanks, or STICT. This novel system could be applied in either an automated or... 10. A cryogenic test facility Veenendaal, Ian The next generation, space-borne instruments for far infrared spectroscopy will utilize large diameter, cryogenically cooled telescopes in order to achieve unprecedented sensitivities. Low background, ground-based cryogenic facilities are required for the cryogenic testing of materials, components and subsystems. The Test Facility Cryostat (TFC) at the University of Lethbridge is a large volume, closed cycle, 4K cryogenic facility, developed for this purpose. This thesis discusses the design and performance of the facility and associated external instrumentation. An apparatus for measuring the thermal properties of materials is presented, and measurements of the thermal expansion and conductivity of carbon fibre reinforced polymers (CFRPs) at cryogenic temperatures are reported. Finally, I discuss the progress towards the design and fabrication of a demonstrator cryogenic, far infrared Fourier transform spectrometer. 11. Central Exclusive Production in the STAR Experiment at RHIC Sikora, Rafal 2016-01-01 The STAR experiment at the Relativistic Heavy Ion Collider (RHIC) performs studies of diffractive processes with the focus on the exclusive production of particles in central range of rapidity. In 2015 STAR collected 18 pb$^{-1}$of data in polarized proton+proton collisions at$\\sqrt{s}$=200 GeV to measure Central Exclusive Production (CEP) process$pp\\to pXp$through Double Pomeron Exchange (DPE) mechanism. The intact protons moving inside the RHIC beampipe after the collision were measured in silicon strip detectors (SSD), which were placed in the Roman Pot vessels. This enables full control over interaction kinematics and verification of the exclusivity of the reaction by measuring the total (missing) transverse momenta of all final state particles: the central diffractive system in the Time Projection Chamber (TPC) and the forward protons in the Roman Pots. With the use of ionization energy loss in the TPC, d$E$/d$x$, it was possible to discriminate various production channels in$pp\\to pXp$reaction. Th... 12. Introduction to cryogenic engineering CERN. Geneva; Vandoni, Giovanna; Niinikoski, Tapio O 2005-01-01 Cryogenic engineering is one of the key technologies at CERN. It is widely used in research and has many applications in industry and last but not least in medicine. In research cryogenic engineering and its applications are omnipresent from the smallest laboratories to fusion reactors, hughe detectors and accelerators. With the termination of the LHC, CERN will in fact become the world's largest cryogenic installation. This series of talks intends to introduce the non-cryogenist to the basic principles and challenges of cryogenic engineering and its applications. The course will also provide a basis for practical application as well as for further learning. 13. ANALYSIS OF AVAILABILITY AND RELIABILITY IN RHIC OPERATIONS. PILAT, F.; INGRASSIA, P.; MICHNOFF, R. 2006-06-26 RHIC has been successfully operated for 5 years as a collider for different species, ranging from heavy ions including gold and copper, to polarized protons. We present a critical analysis of reliability data for RHIC that not only identifies the principal factors limiting availability but also evaluates critical choices at design times and assess their impact on present machine performance. RHIC availability data are typical when compared to similar high-energy colliders. The critical analysis of operations data is the basis for studies and plans to improve RHIC machine availability beyond the 50-60% typical of high-energy colliders. 14. Conceptual design of a quadrupole magnet for eRHIC Witte, H. [Brookhaven National Lab. (BNL), Upton, NY (United States); Berg, J. S. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-05-03 eRHIC is a proposed upgrade to the existing Relativistic Heavy Ion Collider (RHIC) hadron facility at Brookhaven National Laboratory, which would allow collisions of up to 21 GeV polarized electrons with a variety of species from the existing RHIC accelerator. eRHIC employs an Energy Recovery Linac (ERL) and an FFAG lattice for the arcs. The arcs require open-midplane quadrupole magnets of up to 30 T/m gradient of good field quality. In this paper we explore initial quadrupole magnet design concepts based on permanent magnetic material which allow to modify the gradient during operation. 15. Charm and beauty production at RHIC Kabana, Sonia [Laboratoire de Physique Subatomique et des Technologies Associees (SUBATECH), Ecole des Mines, 4 rue Alfred Kastler, 44307 Nantes (France) 2011-01-15 We review selected highlights on charm and beauty production at RHIC from p+p, d+Au and A+A collisions at {radical}(s{sub NN})=200GeV, and compare them to model calculations. We focus on two particular issues, jet quenching and quarkonia. Anomalous energy loss (jet quenching) of quarks passing through the dense and hot matter built in heavy ion collisions is one of the outstanding discoveries made at RHIC. This phenomenon allows for an estimate of the initial gluon density. Furthermore, color screening of hidden charm and beauty states is a key signature of the QCD phase transition, allowing an estimate of the initial temperature. We present results on the flavour dependence of jet quenching. Heavy flavour production in A+A as compared to p+p collisions will be discussed for open and hidden charm. 16. Observations of Snake Resonance in RHIC Bai, Mei; Lee, Shyh-Yuan; Lin, Fanglei; MacKay, William; Ptitsyn, Vadim; Roser, Thomas; Tepikian, Steven 2005-01-01 Siberian snakes now become essential in the polarized proton acceleration. With proper configuration of Siberian snakes, the spin precession tune of the beam becomes$\\frac{1}{2}$which avoids all the spin depolarizing resonance. However, the enhancement of the perturbations on the spin motion can still occur when the betatron tune is near some low order fractional numbers, called snake resonances, and the beam can be depolarized when passing through the resonance. The snake resonances have been confirmed in the spin tracking calculations, and observed in RHIC with polarized proton beam. Equipped with two full Siberian snakes in each ring, RHIC provides us a perfect facility for snake resonance studies. This paper presents latest experimental results. New insights are also discussed. 17. Helical spin rotators and snakes for RHIC Ptitsin, V.I.; Shatunov, Yu.M. [Budker Inst. of Nuclear Physics, Novosibirsk (Russian Federation); Peggs, S. [Brookhaven National Lab., Upton, NY (United States) 1995-05-01 The RHIC collider, now under construction at BNL, will have the possibility of polarized proton-proton collisions up to a beam energy of 250 Gev. Polarized proton beams of such high energy can be only obtained with the use of siberian snakes, a special kind of spin rotator that rotates the particle spin by 180{degree} around an axis lying in the horizontal plane. Siberian snakes help to preserve the beam polarization while numerous spin depolarizing resonances are crossed, during acceleration. In order to collide longitudinally polarized beams, it is also planned to install spin rotators around two interaction regions. This paper discusses snake and spin rotator designs based on sequences of four helical magnets. The schemes that were chosen to be applied at RHIC are presented. 18. Ferrite HOM Absorber for the RHIC ERL Hahn,H.; Choi, E.M.; Hammons, L. 2008-10-01 A superconducting Energy Recovery Linac is under construction at Brookhaven National Laboratory to serve as test bed for RHIC upgrades. The damping of higher-order modes in the superconducting five-cell cavity for the Energy-Recovery linac at RHIC is performed exclusively by two ferrite absorbers. The ferrite properties have been measured in ferrite-loaded pill box cavities resulting in the permeability values given by a first-order Debye model for the tiled absorber structure and an equivalent permeability value for computer simulations with solid ring dampers. Measured and simulated results for the higher-order modes in the prototype copper cavity are discussed. First room-temperature measurements of the finished niobium cavity are presented which confirm the effective damping of higher-order modes in the ERL. by the ferrite absorbers. 19. The Odderon at RHIC and LHC Nicolescu, Basarab 2007-01-01 The Odderon remains an elusive object, 33 years after its invention. The Odderon is now a fundamental object in QCD and CGC and it has to be found experimentally if QCD and CGC are right. In the present talk, we show how to find it at RHIC and LHC. The most spectacular signature of the Odderon is the predicted difference between the differential cross-sections for proton-proton and antiproton-proton at high s and moderate t. This experiment can be done by using the STAR detector at RHIC and by combining these future data with the already present UA4/2 data. The Odderon could also be found by ATLAS experiment at LHC by performing a high-precision measurement of the real part of the hadron elastic scattering amplitude at small t. 20. A Testbed for Implementing Prognostic Methodologies on Cryogenic Propellant Loading Systems National Aeronautics and Space Administration — Prognostics technologies determine the health state of a system and predict its remaining useful life. With this information, operators are able to make... 1. Cryogenic safety aspect of the low -$\\betamagnest systems at the Large Hadron Collider (LHC) Darve, C.; /Fermilab 2010-07-01 The low-{beta} magnet systems are located in the LHC insertion regions around the four interaction points. They are the key elements in the beams focusing/defocusing process and will allow proton collisions at a luminosity of up to 10{sup 34}cm{sup -2}s{sup -1}. Large radiation dose deposited at the proximity of the beam collisions dictate stringent requirements for the design and operation of the systems. The hardware commissioning phase of the LHC was completed in the winter of 2010 and permitted to validate this system safe operation. This paper presents the analysis used to qualify and quantify the safe operation of the low-{beta} magnet systems in the Large Hadron Collider (LHC) for the first years of operation. 2. eRHIC ERL modeling in Zgoubi Meot, F. [Brookhaven National Lab. (BNL), Upton, NY (United States); Brooks, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Hao, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States); Jing, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ptitsyn, V. [Brookhaven National Lab. (BNL), Upton, NY (United States); Trbojevic, D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tsoupas, N. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2016-01-01 This Note discusses on-going work regarding the modeling of eRHIC ERL in the ray-tracing code Zgoubi. The various pieces of the recirculator puzzle, their optical properties and their assemblage into an operational input data file in are addressed. The Note reports in particular on preparatory stages toward extensive end-to-end 6D polarized electron bunch transport simulations, which yield methods, as well a series of preliminary qualitative outcomes, discussed as well. 3. New Results from Spin Physics at RHIC Fatemi, Renee 2009-05-01 The sign and magnitude of the gluon spin contribution (δG) to the spin of the proton has been a topic of intense interest and speculation since inclusive deep inelastic scattering experiments found the total quark spin contribution to be surprisingly small. Starting in 2002, the Relativistic Heavy Ion Collider (RHIC) at Brookhaven National Lab has provided access to longitudinally and transversely polarized proton collisions. Both PHENIX and STAR, the two largest collaborations at RHIC, have used this data to complete a series of inclusive hadron and jet double spin asymmetry (ALL) measurements. The mid-rapidity 0̂ and jet results, now included in a global analysis of existing world data, are shown to provide significant constraints on δG within their range of kinematic sensitivity. Recent inclusive pion and jet ALL measurements will be presented. Plans to measure ALL in correlation channels, for example di-jets and photon-jets, and parity violating asymmetries for identified W^+/- in future longitudinal proton runs will be discussed. In addition to a successful δG program, the RHIC-Spin community is actively contributing to the new and rapidly expanding frontier within nucleon structure studies of transverse spin measurements. Quantum Chromodynamics predicts an extremely small (mq√s) spin asymmetry for leading hadron production in the reaction p^p->h+X. Contrary to expectations, transverse single-spin asymmetries (SSA) of up to 30% were discovered in forward particle production more than three decades ago, and surprisingly, asymmetries of the same magnitude have been found to persist at current RHIC center-of-mass energies. The most recent forward 0̂ and η SSA from STAR and PHENIX, as well as charged hadron measurements from the BRAHMS collaboration, will be discussed and compared with theoretical predictions. 4. Source of second order chromaticity in RHIC Luo, Y.; Gu, X.; Fischer, W.; Trbojevic, D. 2011-01-01 In this note we will answer the following questions: (1) what is the source of second order chromaticities in RHIC? (2) what is the dependence of second order chromaticity on the on-momentum {beta}-beat? (3) what is the dependence of second order chromaticity on {beta}* at IP6 and IP8? To answer these questions, we use the perturbation theory to numerically calculate the contributions of each quadrupole and sextupole to the first, second, and third order chromaticities. 5. Evaluation of Power Electronic Components and Systems at Cryogenic Temperatures For Space Missions Elbuluk, Malik E.; Gerber, Scott; Hammoud, Ahmad; Patterson, Richard L. 2005-01-01 Power electronic circuits and systems designed for deep space applications and outer planetary exploration are required to operate reliably and efficiently under extreme temperature conditions. This requirement is dictated by the fact that the operational environments associated with some of the space missions would encompass temperatures as low as -183 C. The development and utilization of electronics capable of low temperature operation would not only fulfill the advanced technology requirements, but also would contribute to improving circuit performance, increasing system efficiency, and reducing development and launch costs. These benefits are generally achieved by the improved intrinsic properties of some of the electronic materials at low temperature, reduced device losses, and the elimination of heating elements used in conventional systems at low temperatures. Power electronic circuits are widely used in space power systems in the areas of power management, conditioning, and control. In this work, the performance of certain power electronic components and systems was investigated under low temperature. These include inductors, capacitors, pulse-width-modulation (PWM) controllers, and advanced commercial DC/DC converter modules. Different properties were determined as a function of temperature in the range of 20 C to -140 C, at various current and voltages levels. The experimental procedures along with the experimental data obtained are presented and discussed in this paper. 6. QCD and Heavy Ions RHIC Overview Granier de Cassagnac, Raphael 2010-01-01 Nowadays, the most violent heavy ion collisions available to experimental study occur at the Relativistic Heavy Ion Collider (RHIC) of the Brookhaven National Laboratory. There, gold ions collide at psNN = 200 GeV. The early and most striking RHIC results were summarised in 2005 by its four experiments, BRAHMS, PHENIX, PHOBOS and STAR, in their so-called white papers [1, 2, 3, 4] that will be largely referenced thereafter. Beyond and after this, a wealth of data has been collected and analysed, providing additional information about the properties of the matter created at RHIC. It is categorically impossible to give a comprehensive review of these results in a 20 minutes talk or a 7 pages report. Here, I have made a selection of some of the most striking or intriguing signatures: jet quenching in Section 2, quarkonia suppressions in Section 3 and thermal photons in Section 4. A slightly longer and older version of this review can be found in [5]. Some updates are given here, as well as emphasis on new probes ... 7. Strongly interacting matter at RHIC: experimental highlights Okorokov, V A 2014-01-01 Recent experimental results obtained at the Relativistic Heavy-Ion Collider (RHIC) will be discussed. Investigations of different nucleus-nucleus collisions in recent years focus on two main tasks, namely, the detailed study of sQGP properties and the exploration of the QCD phase diagram. Results at top RHIC energy provide important information about event shapes as well as transport and thermodynamic properties of the hot medium for various flavors. Heavy-ion collisions are a unique tool for the study of topological properties of theory. Experimental results obtained for discrete QCD symmetries at finite temperatures are discussed. These results confirm indirectly the topologically non-trivial structure of the QCD vacuum. Most results obtained during phase-I of the RHIC beam energy scan (BES) program show smooth behavior vs initial energy. However, certain results suggest the transition in the domain of dominance of hadronic degrees of freedom at center-of-mass energies between 10-20 GeV. Future developments... 8. Determination of the Optimal Operating Parameters for Jefferson Laboratory's Cryogenic Cold Compressor Systems Wilson, J D 2003-01-01 The technology of Jefferson Laboratory's (JLab) Continuous Electron Beam Accelerator Facility (CEBAF) and Free Electron Laser (FEL) requires cooling from one of the world's largest 2K helium refrigerators known as the Central Helium Liquefier (CHL). The key characteristic of CHL is the ability to maintain a constant low vapor pressure over the large liquid helium inventory using a series of five cold compressors. The cold compressor system operates with a constrained discharge pressure over a range of suction pressures and mass flows to meet the operational requirements of CEBAF and FEL. The research topic is the prediction of the most thermodynamically efficient conditions for the system over its operating range of mass flows and vapor pressures with minimum disruption to JLab operations. The research goal is to find the operating points for each cold compressor for optimizing the overall system at any given flow and vapor pressure. 9. Systems Design, Fabrication, and Testing of a High-Speed Miniature Motor for Cryogenic Cooler Dipjyoti Acharya 2009-01-01 Full Text Available The long-term storage of liquid hydrogen for space missions is of considerable interest to NASA. To this end, the Reverse Turbo-Brayton Cryocooler (RTBC is considerably lighter than conventional designs and a potentially viable and attractive solution for NASA's long-term Zero-Boil-off (ZBO hydrogen storage system for future space missions. We present the systems design, fabrication, and performance evaluation of the Permanent Magnet Synchronous Motor (PMSM powering a cryocooler capable of removing 20 W of heat at 18 K with a COP of 0.005 and driven by two 2-kW permanent magnet synchronous motors operating at 200 000 rpm and at room temperature and 77 K. Structural, thermal, and rotordynamic aspects of system design are considered. 10. Determination of the Optimal Operating Parameters for Jefferson Laboratory's Cryogenic Cold Compressor Systems Joe D. Wilson, Jr. 2003-04-01 The technology of Jefferson Laboratory's (JLab) Continuous Electron Beam Accelerator Facility (CEBAF) and Free Electron Laser (FEL) requires cooling from one of the world's largest 2K helium refrigerators known as the Central Helium Liquefier (CHL). The key characteristic of CHL is the ability to maintain a constant low vapor pressure over the large liquid helium inventory using a series of five cold compressors. The cold compressor system operates with a constrained discharge pressure over a range of suction pressures and mass flows to meet the operational requirements of CEBAF and FEL. The research topic is the prediction of the most thermodynamically efficient conditions for the system over its operating range of mass flows and vapor pressures with minimum disruption to JLab operations. The research goal is to find the operating points for each cold compressor for optimizing the overall system at any given flow and vapor pressure. 11. A Cryogenic Flow Sensor Project National Aeronautics and Space Administration — Advanced Technologies Group, Inc. proposes the development of a Cryogenic Flow Sensor (CFS) for determining mass flow of cryogens in spacecraft propellant... 12. Tri-Gas Pressurization System Testing and Modeling for Cryogenic Applications Taylor, B.; Polsgrove, R.; Stephens, J.; Hedayat, A. 2014-01-01 The use of Tri-gas in rocket propulsion systems is somewhat of a new technology. This paper defines Tri-gas as a mixture of gases composed largely of helium with a small percentage of a stoichiometric mixture of hydrogen and oxygen. When exposed to a catalyst the hydrogen and oxygen in the mixture combusts, significantly raising the temperature of the mixture. The increase in enthalpy resulting from the combustion process significantly decreases the required quantity of gas needed to pressurize the ullage of the vehicle propellant tanks. The objective of this effort was to better understand the operating characteristics of Tri-gas in a pressurization system with low temperature applications. In conjunction with ongoing programs at NASA Marshall Space Flight Center, an effort has been undertaken to evaluate the operating characteristics of Tri-gas through modeling and bench testing. Through improved understanding of the operating characteristics, the risk of using this new technology in a launch vehicle propulsion system was reduced. Bench testing of Tri-gas was a multistep process that targeted gas characteristics and performance aspects that pose a risk to application in a pressurization system. Pressurization systems are vital to propulsion system performance. Keeping a target ullage pressure in propulsions tanks is necessary to supply propellant at the conditions and flow rates required to maintain desired engine functionality. The first component of testing consisted of sampling Tri-gas sources that had been stagnant for various lengths of time in order to determine the rate at which stratification takes place. Second, a bench test was set up in which Tri-gas was sent through a catalyst bed. This test was designed to evaluate the performance characteristics of Tri-gas, under low temperature inlet temperatures, in a flight-like catalyst bed reactor. The third, most complex, test examined the performance characteristics of Tri-gas at low temperature temperatures 13. Passive Gas-Gap Heat Switches for Use in Low-Temperature Cryogenic Systems Kimball, M. O.; Shirron, P. J.; Canavan, E. R.; Tuttle, J. G.; Jahromi, A. E.; Dipirro, M. J.; James, B. L.; Sampson, M. A.; Letmate, R. V. 2017-01-01 We present the current state of development in passive gas-gap heat switches. This type of switch does not require a separate heater to activate heat transfer but, instead, relies upon the warming of one end due to an intrinsic step in a thermodynamic cycle to raise a getter above a threshold temperature. Above this temperature sequestered gas is released to couple both sides of the switch. This enhances the thermodynamic efficiency of the system and reduces the complexity of the control system. Various gas mixtures and getter configurations will be presented. 14. Heavy flavor in heavy-ion collisions at RHIC and RHIC II Frawley, A D; Ullrich, T; Vogt, R 2008-03-30 In the initial years of operation, experiments at the Relativistic Heavy Ion Collider (RHIC) have identified a new form of matter formed in nuclei-nuclei collisions at energy densities more than 100 times that of a cold atomic nucleus. Measurements and comparison with relativistic hydrodynamic models indicate that the matter thermalizes in an unexpectedly short time, has an energy density at least 15 times larger than needed for color deconfinement, has a temperature about twice the critical temperature predicted by lattice QCD, and appears to exhibit collective motion with ideal hydrodynamic properties--a 'perfect liquid' that appears to flow with a near-zero viscosity to entropy ratio--lower than any previously observed fluid and perhaps close to a universal lower bound. However, a fundamental understanding of the medium seen in heavy-ion collisions at RHIC does not yet exist. The most important scientific challenge for the field in the next decade is the quantitative exploration of the new state of nuclear matter. That will require new data that will, in turn, require enhanced capabilities of the RHIC detectors and accelerator. In this report we discuss the scientific opportunities for an upgraded RHIC facility --RHIC II--in conjunction with improved capabilities of the two large RHIC detectors, PHENIX and STAR. We focus solely on heavy flavor probes. Their production rates are calculable using the well-established techniques of perturbative QCD and their sizable interactions with the hot QCD medium provide unique and sensitive measurements of its crucial properties making them one of the key diagnostic tools available to us. 15. Parametric analysis of a novel cryogenic CO2 capture system based on Stirling coolers. Song, Chun Feng; Kitamura, Yutaka; Li, Shu Hong; Jiang, Wei Zhong 2012-11-20 CO(2) capture and storage (CCS) is an important alternative to control greenhouse gas (GHG) effects. In previous work, a novel desublimation CO(2) capture process has been exploited making use of three free piston Stirling coolers (namely, SC-1, SC-2, and SC-3, respectively). Based on the developed system, moisture and CO(2) in the flue gas can condense and desublimate in the prefreezing and main-freezing towers, respectively. Meanwhile, the storage column is chilled by SC-3 to preserve the frosted CO(2), and permanent gas (such as N(2)) passes through the system without phase change. The whole process can be implemented at atmospheric pressure and reduce the energy penalty (e.g., solvent regeneration and pressure drop) in other technologies. In this work, the influence of process parameters has been investigated in detail. The optimal conditions for the system are as follows: idle operating time is 240 min, flow rate is 5 L/min, vacuum degree of the interlayer is 2.2 × 10(3) Pa, and temperatures of SC-1, -2, and -3 are -30, -120, and -120 °C, respectively. Under these conditions, the energy consumption of the system is around 0.5 MJ(electrical)/kg CO(2) with above 90% CO(2) recovery. 16. Two-statge sorption type cryogenic refrigerator including heat regeneration system Jones, Jack A. (Inventor); Wen, Liang-Chi (Inventor); Bard, Steven (Inventor) 1989-01-01 A lower stage chemisorption refrigeration system physically and functionally coupled to an upper stage physical adsorption refrigeration system. Waste heat generated by the lower stage cycle is regenerated to fuel the upper stage cycle thereby greatly improving the energy efficiency of a two-stage sorption refrigerator. The two stages are joined by disposing a first pressurization chamber providing a high pressure flow of a first refrigerant for the lower stage refrigeration cycle within a second pressurization chamber providing a high pressure flow of a second refrigerant for the upper stage refrigeration cycle. The first pressurization chamber is separated from the second pressurization chamber by a gas-gap thermal switch which at times is filled with a thermoconductive fluid to allow conduction of heat from the first pressurization chamber to the second pressurization chamber. 17. Cryogenic microcalorimeter system for ultra-high resolution alpha-particle spectrometry Rabin, Michael W [Los Alamos National Laboratory; Hoover, Andrew S [Los Alamos National Laboratory; Bacrania, Mnesh K [Los Alamos National Laboratory; Croce, Mark P [Los Alamos National Laboratory; Hoteling, N J [Los Alamos National Laboratory; Lamont, S P [Los Alamos National Laboratory; Plionis, A A [Los Alamos National Laboratory; Dry, D E [Los Alamos National Laboratory; Ullom, J N [NIST; Bennett, D A [NIST; Horansky, R [NIST; Kotsubo, V [NIST; Cantor, R [STAR CRYOELECTRONICS 2009-01-01 Microcalorimeters have been shown to yield unsurpassed energy resolution for alpha spectrometry, up to 1.06 keV FWHM at 5.3 MeV. These detectors use a superconducting transition-edge sensor (TES) to measure the temperature change in an absorber from energy deposited by an interacting alpha particle. Our system has four independent detectors mounted inside a liquid nitrogen/liquid helium cryostat. An adiabatic demagnetization refrigerator (ADR) cools the detector stage to its operating temperature of 80 mK. Temperature regulation with {approx}15 uK peak-to-peak variation is achieved by PID control of the ADR. The detectors are voltage-biased, and the current signal is amplified by a commercial SQUID readout system and digitized for further analysis, This paper will discuss design and operation of our microcalorimeter alpha spectrometer, and will show recent results. 18. Thermal Analysis on Cryogenic Liquid Hydrogen Tank on an Unmanned Aerial Vehicle System Wang, Xiao-Yen; Harpster, George; Hunter, James 2007-01-01 Thermal analyses are performed on the liquid hydrogen (LH2) tank designed for an unmanned aerial vehicle (UAV) powered by solar arrays and a regenerative proton-exchange membrane (PEM) fuel cell. A 14-day cruise mission at a 65,000 ft altitude is considered. Thermal analysis provides the thermal loads on the tank system and the boiling-off rates of LH2. Different approaches are being considered to minimize the boiling-off rates of the LH2. It includes an evacuated multilayer insulation (MLI) versus aerogel insulation on the LH2 tank and aluminum versus stainless steel spacer rings between the inner and outer tank. The resulting boil-off rates of LH2 provided by the one-dimensional model and three-dimensional finite element analysis (FEA) on the tank system are presented and compared to validate the results of the three-dimensional FEA. It concludes that heat flux through penetrations by conduction is as significant as that through insulation around the tank. The tank system with MLI insulation and stainless steel spacer rings result in the lowest boiling-off rate of LH2. 19. Photon Stimulated Desorption and the Effect of Cracking of Condensed Molecules in a Cryogenic Vacuum System Anashin, V V; Gröbner, Oswald; Malyshev, O B 2000-01-01 The design of the Large Hadron Collider (LHC) vacuum system requires a complete understanding of all processes which may affect the residual gas density in the cold bore of the 1.9 K cryomagnets. A wealth of data has been obtained which may be used to predict the residual gas density inside a cold vacuum system exposed to synchrotron radiation. In this study the effect of cracking of cryosorbed molecules by synchrotron radiation photons has been included. Cracking of the molecular species CO2 and CH4 has been observed in recent studies and these findings have been incorporated in a more detailed dynamic gas density model for the LHC. In this paper, we describe the relevant physical processes and the parameters required for a full evaluation. It is shown that the dominant gas species in the LHC vacuum system with its beam screen are H2 and CO. The important result of this study is that while the surface coverage of cryosorbed CH4 and CO2 molecules is limited due to cracking, the coverage of H2 and CO molecules... 20. Nanodielectrics for Cryogenic Applications Tuncer, Enis [ORNL; Sauers, Isidor [ORNL; James, David Randy [ORNL; Ellis, Alvin R [ORNL; Pace, Marshall O [ORNL; More, Karren [Oak Ridge National Laboratory (ORNL); Sathyamurthy, Srivatsan [University of Tennessee, Knoxville (UTK); Woodward, Jonathan [ORNL; Rondinone, Adam Justin [ORNL 2009-01-01 In this paper we report the recent advances in nanodielectrics that were developed and tested for cryogenic dielectric applications. The systems studied are composed of nanometer size particles. Particles were produced using either an ex-situ or in-situ technique. It is observed that there are clear differences in the structural properties of materials produced using these two approaches. Either no significant degradation or improvement in the electrical insulation properties were observed for ex-situ nano-particle samples processed with an ultrasonic processor and in-situ nano-particle samples. Nanodielectrics have the potential to be tailored with better thermal and mechanical properties without losing their electrical insulation characteristics. 1. Cryogenic system for X-ray Compton scattering measurements of superfluid helium below 2 K Tanaka, Hiroyuki; Yamaguchi, Akira; Koizumi, Akihisa; Kawasaki, Ikuto; Sumiyama, Akihiko; Itou, Masayoshi; Sakurai, Yoshiharu 2017-07-01 A cryostat was constructed for high-resolution X-ray Compton scattering measurements at temperature down to 1.7 K, in order to investigate superfluid helium-4. Compton profiles of helium were measured using synchrotron X-rays for gas and liquid phases, respectively. In the measurement of the liquid phase, we succeeded in measuring the Compton profile of the superfluid helium at 1.7 K. Comparison of the results with theoretical calculation reveals importance of many-body effects beyond the mean-field treatment of electron systems. 2. Construction of a 10 kW class helium cryogenic system for the large helical device LHD Group; Satoh, Sadao; Mito, Toshiyuki; Yamada, Syuuichi; Yamamoto, Junya; Motojima, Osamu; LHD Group The system has an equivalent design capacity of 9100Watt refrigeration at 4.4Kand will refrigerate all set of superconducting coils and their supporting structure of LHD. The fabrication at the vender's shop has been already completed and the construction at the site is now underway. In the paper, emphasis was given to discuss about the shop test methods and their results on compressors and turboexpanders. This is because the test facilities at the vender's shop are not directly applicable due to such large machine size or extreme operating conditions and the informations from shop run are largely dependent to the "equivalent" test performed. Nevertheless it was indicated that such result can be correlated to the design performances within the limit of assumptions made. 3. Radiation Tests on the Complete System of the Instrumentation of the LHC Cryogenics at the CERN Neutrinos to Gran Sasso (CNGS) Test Facility Gousiou, E; Casas Cubillos, J; de la Gama Serrano, J 2009-01-01 There are more than 6000 electronic cards for the instrumentation of the LHC cryogenics, housed in crates and distributed around the 27 km tunnel. Cards and crates will be exposed to a complex radiation field during the 10 years of LHC operation. Rad-tol COTS and rad-hard ASIC have been selected and individually qualified during the design phase of the cards. The test setup and the acquired data presented in this paper target the qualitative assessment of the compliance with the LHC radiation environment of an assembled system. It is carried out at the CNGS test facility which provides exposure to LHC-like radiation field. 4. Ring imaging Cherenkov detector of PHENIX experiment at RHIC Akiba, Y; Burward-Hoy, J; Chappell, R; Crook, D; Ebisu, K; Emery, M S; Ferriera, J; Frawley, A D; Hamagaki, H; Hara, H; Hayano, R S; Hemmick, T K; Hibino, M; Hutter, R; Kennedy, M; Kikuchi, J; Matsumoto, T; Moscone, C G; Nagasaka, Y; Nishimura, S; Oyama, K; Sakaguchi, T; Salomone, S; Shigaki, K; Tanaka, Y; Walker, J W; Wintenberg, A L; Young, G R 1999-01-01 The RICH detector of the PHENIX experiment at RHIC is currently under construction. Its main function is to identity electron tracks in a very high particle density, about 1000 charged particles per unit rapidity, expected in the most violent collisions at RHIC. The design and construction status of the detector and its expected performance are described. 5. Nuclear Stopping:. Paving the way from Rhic to Lhc Dalsgaard, Hans Hjersing Nuclear stopping has been measured at a range of different energies in heavy ion experiments. In this contribution proton data from the BRAHMS experiment at RHIC running at √ {SNN} = 62.4\\ GeV are presented. Furthermore data from AGS, SPS and RHIC are used to estimate the stopping, energy loss and multiplicity at LHC. 6. Experimental effects of orbit on polarization loss in RHIC Ranjbar V.; Bai, M.; Huang, H.; Marusic, A.; Ptitsyn, V.; Minty, M. 2012-05-20 We are performing several experiments during the RHIC ramp to better understand the impact of orbit errors on the polarization at our current working point. These will be conducted by exciting specified orbit harmonics during the final two large intrinsic resonance crossing in RHIC during the 250 GeV polarized proton ramp. The resultant polarization response will then be measured. 7. Measurements of strangeness production in the STAR experiment at RHIC Wilson, W.K. [Wayne State Univ., Detroit, MI (United States) 1995-07-15 Simulations of the ability of the STAR (Solenoidal Tracker at RHIC) detector to measure strangeness production in central Au+Au collisions at RHIC are presented. Emphasis is placed on the reconstruction of short lived particles using a high resolution inner tracker. The prospects for performing neutral kaon interferometry are discussed. Simulation results for measurements of strange and multi-strange baryons are presented. 8. Study of the Beam Energy Dependence of Azimuthal Anisotropy Coefficients and Non-Flow Effects in Small System d +Au Collisions at RHIC Yin, Pengqi 2016-09-01 Recent measurements of azimuthal anisotropy, v_n, in collision systems such as p,d,3He +Au suggest that a quark gluon plasma (QGP) may be formed in these small systems, which would be an unexpected discovery. However, this QGP lives for a shorter time than in larger A +A systems and it is not clear how the azimuthal anisotropy signals develop. Varying the collision energy in d +Au collisions can help to answer this question. However, non-flow effects are more dominant in small systems and must be accounted for in order to draw conclusions. We will show theoretical calculations of v_2 and v_3 in d +Au using different models at several collision energies, and we will present a method based on reference fitting to estimate the non-flow component in actual measurements so that they might be better compared to the theory. (Based on work published in) Division of Nuclear Physics of the U.S. Department of Energy under Grant No. DE-FG02-00ER41152. 9. Advances in cryogenic engineering. Volume 29 - Proceedings of the Cryogenic Engineering Conference, Colorado Springs, CO, August 15-17, 1983 Fast, R. W. Applications of superconductivity are discussed, taking into account the thermal performance of the MFTF magnets, the design and testing of a large bore superconducting magnet test facility, the development of a 12-tesla multifilamentary Nb3Sn magnet, a superconducting magnet for solid NMR studies, advanced applications of superconductors, transition and recovery of a cryogenically stable superconductor, and finite-difference modeling of the cryostability of helium II cooled conductor packs. Other topics explored are related to resource availability, heat exchangers, heat transfer to He I, liquid nitrogen, heat transfer in He II, refrigeration for superconducting and cryopump systems, refrigeration of cryogenic systems, refrigeration and liquefaction, dilution and magnetic refrigeration, cryocoolers, refrigeration for space applications, cryogenic applications, cryogenic instrumentation and data acquisition, and properties of fluids. Attention is given to biomedical applications of cryogenics in China, long-term cryogen storage in space, and a passive orbital disconnect strut. 10. Proceedings of the third workshop on experiments and detectors for a relativistic heavy ion collider (RHIC) Shivakumar, B.; Vincent, P. 1988-01-01 This report contains papers on the following topics: the RHIC Project; summary of the working group on calorimetry; J//Psi/ measurements in heavy ion collisions at CERN; QCD jets at RHIC; tracking and particle identification; a 4..pi.. tracking spectrometer for RHIC; Bose-Einstein measurements at RHIC in light of new data; summary of working group on read-out electronics; data acquisition for RHIC; summary of the working group on detector simulation; B-physics at RHIC; and CP violation revisited at BNL, B-physics at RHIC. 11. Polymers at cryogenic temperatures Fu, Shao-Yun 2013-01-01 Kalia and Fu's novel monograph covers cryogenic treatment, properties and applications of cryo-treated polymer materials. Written by numerous international experts, the twelve chapters in this book offer the reader a comprehensive picture of the latest findings and developments, as well as an outlook on the field. Cryogenic technology has seen remarkable progress in the past few years and especially cryogenic properties of polymers are attracting attention through new breakthroughs in space, superconducting, magnetic and electronic techniques. This book is a valuable resource for researchers, educators, engineers and graduate students in the field and at technical institutions. 12. Fundamentals of cryogenic engineering Mukhopadhyay, Mamata 2014-01-01 The author, with her vast and varied experience in teaching and allied fields, clearly enunciates the behaviour and various properties of common cryogenic fluids, methods of liquefaction, and separation and applications of cryogens with thermodynamic analysis for process selection. This profusely illustrated study with clear-cut diagrams and process charts, should serve not only as a textbook for students but also as an excellent reference for researchers and practising engineers on design of cryogenic refrigeration, and liquefaction and separation process plants for various applications. 13. RHIC spin physics: Proceedings. Volume 7 NONE 1998-12-01 This proceedings compiles one-page summaries and five transparencies for each talk, with the intention that the speaker should include a web location for additional information in the summary. Also, email addresses are given with the participant list. The order follows the agenda: gluon, polarimetry, accelerator, W production and quark/antiquark polarization, parity violation searches, transversity, single transverse spin, small angle elastic scattering, and the final talk on ep collisions at RHIC. The authors begin the Proceedings with the full set of transparencies from Bob Jaffe`s colloquium on spin, by popular request. 14. RHIC operation with asymmetric collisions in 2015 Liu, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Aschenauer, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Atoian, G. [Brookhaven National Lab. (BNL), Upton, NY (United States); Blaskiewicz, M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Brown, K. A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Bruno, D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Connolly, R. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ottavio, T. D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Drees, K. A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Fischer, W. [Brookhaven National Lab. (BNL), Upton, NY (United States); Gardner, C. J. [Brookhaven National Lab. (BNL), Upton, NY (United States); Gu, X. [Brookhaven National Lab. (BNL), Upton, NY (United States); Hayes, T. [Brookhaven National Lab. (BNL), Upton, NY (United States); Huang, H. [Brookhaven National Lab. (BNL), Upton, NY (United States); Laster, J. S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Luo, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States); Makdisi, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States); Marr, G. [Brookhaven National Lab. (BNL), Upton, NY (United States); Marusic, A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Meot, F. [Brookhaven National Lab. (BNL), Upton, NY (United States); Mernick, K. [Brookhaven National Lab. (BNL), Upton, NY (United States); Michnoff, R. [Brookhaven National Lab. (BNL), Upton, NY (United States); Minty, M. [Brookhaven National Lab. (BNL), Upton, NY (United States); Montag, C. [Brookhaven National Lab. (BNL), Upton, NY (United States); Morris, J. [Brookhaven National Lab. (BNL), Upton, NY (United States); Narayan, G. [Brookhaven National Lab. (BNL), Upton, NY (United States); Nayak, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Nemesure, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Pile, P. [Brookhaven National Lab. (BNL), Upton, NY (United States); Poblaguev, A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ranjbar, V. [Brookhaven National Lab. (BNL), Upton, NY (United States); Robert-Demolaize, G. [Brookhaven National Lab. (BNL), Upton, NY (United States); Roser, T. [Brookhaven National Lab. (BNL), Upton, NY (United States); Schmidke, B. [Brookhaven National Lab. (BNL), Upton, NY (United States); Schoefer, V. [Brookhaven National Lab. (BNL), Upton, NY (United States); Severino, F. [Brookhaven National Lab. (BNL), Upton, NY (United States); Shrey, T. [Brookhaven National Lab. (BNL), Upton, NY (United States); Smith, K. [Brookhaven National Lab. (BNL), Upton, NY (United States); Steski, D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tepikian, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Trbojevic, D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tsoupas, N. [Brookhaven National Lab. (BNL), Upton, NY (United States); Wang, G. [Brookhaven National Lab. (BNL), Upton, NY (United States); White, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Yip, K. [Brookhaven National Lab. (BNL), Upton, NY (United States); Zaltsman, A. [Brookhaven National Lab. (BNL), Upton, NY (United States); Zeno, K. [Brookhaven National Lab. (BNL), Upton, NY (United States); Zhang, S. Y. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-08-07 To study low-x shadowing/saturation physics as well as other nuclear effects [1], [2], proton-gold (p-Au, for 5 weeks) and proton-Aluminum (p-Al, for 2 weeks) collisions were provided for experiments in 2015 at the Relativistic Heavy Ion Collider (RHIC), with polarized proton beam in the Blue ring and Au/Al beam in the Yellow ring. The special features of the asymmetric run in 2015 will be introduced. The operation experience will be reviewed as well in the report. 15. Construction progress of the RHIC electron lenses Fischer W.; Altinbas, Z.; Anerella, M.; Beebe, E.; et al 2012-05-20 In polarized proton operation the RHIC performance is limited by the head-on beam-beam effect. To overcome this limitation two electron lenses are under construction. We give an overview of the construction progress. Guns, collectors and the warm electron beam transport solenoids with their power supplies have been constructed. The superconducting solenoids that guide the electron beam during the interaction with the proton beam are near completion. A test stand has been set up to verify the performance of the gun, collector and some of the instrumentation. The infrastructure is being prepared for installation, and simulations continue to optimize the performance. 16. Construction progress of the RHIC electron lenses Fischer W.; Altinbas, Z.; Anerella, M.; Beebe, E.; et al 2012-05-20 In polarized proton operation the RHIC performance is limited by the head-on beam-beam effect. To overcome this limitation two electron lenses are under construction. We give an overview of the construction progress. Guns, collectors and the warm electron beam transport solenoids with their power supplies have been constructed. The superconducting solenoids that guide the electron beam during the interaction with the proton beam are near completion. A test stand has been set up to verify the performance of the gun, collector and some of the instrumentation. The infrastructure is being prepared for installation, and simulations continue to optimize the performance. 17. The Cryogenic Dark Matter Search Sander, Joel 2004-05-01 The Cryogenic Dark Matter Search (CDMS) is an experiment to search for Weakly Interacting Massive Particles (WIMPs). The experiment initially was deployed at a shallow underground site, and is currently deployed at a deep underground site at the Soudan Mine in Minnesota. The detectors operate at cryogenic temperature, and are capable of distinguishing nuclear recoils from WIMP interactions from various backgrounds. The detectors are shielded from background by both active and passive elements. We will describe the components of the overall experiment, and focus on the novel data acquisition system that has been develop to control and monitor the experiment via the World Wide Web. Preliminary signals from the operation at Soudan will be discussed. 18. The cryogenic storage ring project Hahn, Robert von; Blaum, Klaus; Becker, Arno; Fellenberger, Florian; George, Sebastian; Grieser, Manfred; Grussie, Florian; Herwig, Philipp; Krantz, Claude; Kreckel, Holger; Lange, Michael; Menk, Sebastian; Repnow, Roland; Vogel, Stephen; Wolf, Andreas [Max-Planck-Institut fuer Kernphysik, Heidelberg (Germany); Spruck, Kaija [Justus-Liebig-Universitaet, Giessen (Germany) 2014-07-01 At MPIK the electrostatic cryogenic storage ring CSR is nearing completion. At beam energies of 20 to 300 keV per charge unit and 35 m circumference the CSR will allow experiments in a cryogenic environment providing conditions of extremely low vacuum and heat radiation. By using liquid helium at 2 K for cryopumping, the projected vacuum (confirmed at a prototype) lies at 1E-13 mbar or below, ensuring long storage times for slow singly charged and highly charged ions, molecules and clusters. Moreover, phase space cooling by electrons will be implemented. The internal quantum states of molecular and cluster ions can be cooled to low temperature, yielding well defined vibrational and for smaller systems also rotational structures. In the CSR construction, the cryogenic ion beam vacuum system has been set up. Extensive tests confirming the criteria on heat flow, alignment and high-voltage stability were successfully completed on the first quadrant. In addition beam diagnostic units for electric pickup signals and spatial profiles, detectors for neutral and charged fragments, the injection beam line, and an electron cooling device are under construction. 19. Advances in cryogenic engineering. Volume 41, Part A & B Kittel, P. [ed. 1996-12-31 This proceedings is of the 1995 Cryogenic Engineering Conference. It consists of 252 published papers covering the latest developments in all aspects of cryogenic engineering research. Contributions touch on fields including: cryobiology; heat and mass transfer (including data on boiling and superfluid helium); magnet technology; large-scale cryogenic systems, such as the large hadron collider and the TeV Electron Superconducting Linear Accelerator; cryofuels; minesweeping applications; space cryocooler applications; research on miscellaneous cryogenic machinery, techniques, and safety concerns. Separate abstracts have been submitted for contributions from this proceedings. 20. NASA's Cryogenic Fluid Management Technology Project Tramel, Terri L.; Motil, Susan M. 2008-01-01 The Cryogenic Fluid Management (CFM) Project's primary objective is to develop storage, transfer, and handling technologies for cryogens that will support the enabling of high performance cryogenic propulsion systems, lunar surface systems and economical ground operations. Such technologies can significantly reduce propellant launch mass and required on-orbit margins, reduce or even eliminate propellant tank fluid boil-off losses for long term missions, and simplify vehicle operations. This paper will present the status of the specific technologies that the CFM Project is developing. The two main areas of concentration are analysis models development and CFM hardware development. The project develops analysis tools and models based on thermodynamics, hydrodynamics, and existing flight/test data. These tools assist in the development of pressure/thermal control devices (such as the Thermodynamic Vent System (TVS), and Multi-layer insulation); with the ultimate goal being to develop a mature set of tools and models that can characterize the performance of the pressure/thermal control devices incorporated in the design of an entire CFM system with minimal cryogen loss. The project does hardware development and testing to verify our understanding of the physical principles involved, and to validate the performance of CFM components, subsystems and systems. This database provides information to anchor our analytical models. This paper describes some of the current activities of the NASA's Cryogenic Fluid Management Project. 1. High luminosity electron-hadron collider eRHIC Ptitsyn, V.; Aschenauer, E.; Bai, M.; Beebe-Wang, J.; Belomestnykh, S.; Ben-Zvi, I.; Blaskiewicz, M..; Calaga, R.; Chang, X.; Fedotov, A.; Gassner, D.; Hammons, L.; Hahn, H.; Hammons, L.; He, P.; Hao, Y.; Jackson, W.; Jain, A.; Johnson, E.C.; Kayran, D.; Kewisch, J.; Litvinenko, V.N.; Luo, Y.; Mahler, G.; McIntyre, G.; Meng, W.; Minty, M.; Parker, B.; Pikin, A.; Rao, T.; Roser, T.; Skaritka, J.; Sheehy, B.; Skaritka, J.; Tepikian, S.; Than, Y.; Trbojevic, D.; Tsoupas, N.; Tuozzolo, J.; Wang, G.; Webb, S.; Wu, Q.; Xu, W.; Pozdeyev, E.; Tsentalovich, E. 2011-03-28 We present the design of a future high-energy high-luminosity electron-hadron collider at RHIC called eRHIC. We plan on adding 20 (potentially 30) GeV energy recovery linacs to accelerate and to collide polarized and unpolarized electrons with hadrons in RHIC. The center-of-mass energy of eRHIC will range from 30 to 200 GeV. The luminosity exceeding 10{sup 34} cm{sup -2} s{sup -1} can be achieved in eRHIC using the low-beta interaction region with a 10 mrad crab crossing. We report on the progress of important eRHIC R&D such as the high-current polarized electron source, the coherent electron cooling, ERL test facility and the compact magnets for recirculation passes. A natural staging scenario of step-by-step increases of the electron beam energy by building-up of eRHIC's SRF linacs is presented. 2. Dilepton production from RHIC to the LHC Dahms Torsten 2015-01-01 Full Text Available The goal of ultrarelativistic heavy-ion collisions at RHIC and the LHC is to study the properties of the quark-gluon plasma (QGP, a phase of matter with partonic degrees of freedom. Electromagnetic radiation, in form of photons or lepton pairs, is a penetrating probe that allows the investigation of the full time evolution and dynamics of the produced matter as it does not undergo strong interaction in the final state. The dilepton spectrum is extremely rich in physics sources: Thermal black-body radiation is of particular interest as it carries information about the QGP temperature. Modifications of the spectral functions of light vector mesons are linked to the potential restoration of chiral symmetry in the QGP phase. Correlated lepton pairs from semi-leptonic charm and beauty decays provide additional information about the heavy-quark energy loss. Finally, the suppression of quarkonia in the QGP give access to an independent temperature measurement. In this proceedings, dilepton results from RHIC are reviewed and the status as well as prospects of low-mass dilepton measurements at the LHC are given. 3. Novel deflecting cavity design for eRHIC Wu, Q.; Belomestnykh, S.; Ben-Zvi, I. 2011-07-25 To prevent significant loss of the luminosity due to large crossing angle in the future ERL based Electron Ion Collider at BNL (eRHIC), there is a demand for crab cavities. In this article, we will present a novel design of the deflecting/crabbing 181 MHz superconducting RF cavity that will fulfil the requirements of eRHIC. The quarter-wave resonator structure of the new cavity possesses many advantages, such as compact size, high R{sub t}/Q, the absence of the same order mode and lower order mode, and easy higher order mode damping. We will present the properties and characteristics of the new cavity in detail. As the accelerator systems grow in complexity, developing compact and efficient deflecting cavities is of great interest. Such cavities will benefit situations where the beam line space is limited. The future linac-ring type electron-ion collider requires implementation of a crab-crossing scheme for both beams at the interaction region. The ion beam has a long bunches and high rigidity. Therefore, it requires a low frequency, large kicking angle deflector. The frequency of the deflecting mode for the current collider design is 181 MHz, and the deflecting angle is {approx}5 mrad for each beam. At such low frequency, the previous designs of the crab cavities will have very large dimensions, and also will be confronted by typical problems of damping the Lower Order Mode (LOM), the Same Order Mode (SOM), and as usual, the Higher Order Modes (HOM). In this paper we describe how one can use the concept of a quarter-wave (QW) resonator for a deflecting/crabbing cavity, and use its fundamental mode to deflect the beam. The simplicity of the cavity geometry and the large separation between its fundamental mode and the first HOM make it very attractive. 4. Feasibility study of parallel conduction cooling of NbTi magnet and sample probe in a cryogen-free magnet system Catarino, I.; Soni, V.; Barreto, J.; Martins, D.; Kar, S. 2017-02-01 The conduction cooling of both a 6 T superconducting magnet along with a sample probe in a parallel configuration is addressed in this work. A Gifford-McMahon (GM) cryocooler is directly cooling the NbTi magnet, which aims to be kept at 4 K, while a gas-gap heat switch (GGHS) manages the cooling power to be diverted to the sample probe, which may be swept from 4 K up to 300 K. A first prototype of a GGHS was customized and validated for this purpose. A sample probe assembly has been designed and assembled with the existing cryogen-free magnet system. The whole test setup and components are described and the preliminary experimental results on the integration are presented and discussed. The magnet was charged up to 3 T with a 4 K sample space and up to 1 T with a sweeping sample space temperature up to 300 K while acting on the GGHS. Despite some identified thermal insulation problems that occurred during this first test, the overall results demonstrated the feasibility of the cryogen-free parallel conduction cooling on study. 5. PHENIX EXPERIMENT AT RHIC: DECADAL PLAN 2004-2013 ZAJC,W.ET. AL. 2003-11-30 program is achievable using the present capabilities of PHENIX experimental apparatus, but the physics reach is considerably extended and the program made even more compelling by a proposed set of upgrades which include: (1) An aerogel and time-of-flight system to provide complete {pi}/K/p separation for momenta up to 10 GeV/c. (2) A vertex detector to detect displaced vertices from the decay of mesons containing charm or bottom quarks. (3) A hadron-blind detector to detect and track electrons near the vertex. (4) A micro-TPC to extend the range of PHENIX tracking in azimuth and pseudo-rapidity. (5) A forward detector upgrade for an improved muon trigger to preserve sensitivity at the highest projected RHIC luminosities. (6) A forward calorimeter to provide photon+jet studies over a wide kinematic range. The success of the proposed program is contingent upon several factors external to PHENIX. Implementation of the upgrades is predicated on the availability of R&D funds to develop the required detector technologies on a timely, and in some cases urgent, basis. The necessity for such funding, and the physics merit of the proposed PHENIX program, has been endorsed in the first meeting of BNL's Detector Advisory Committee in December, 2002. Progress towards the physics goals depends in an essential way on the development of the design values for RHIC luminosity, polarization and availability. An analysis based on the guidance from the Collider Accelerator Department indicates that moderate increases in the yearly running time lead to very considerable increases in progress toward the enunciated goals. Efficient access to the rarest probes in the proposed program is achieved via the order-of-magnitude increase in luminosity provided by RHIC-II. 6. Search for quark compositeness with polarized beams at RHIC Virey, J M 1996-01-01 Around 1999, thanks to the RHIC Spin Collaboration (RSC), the Relativistic Heavy Ion Collider (RHIC) will be used as a polarized proton-proton collider. A new handed interaction between quark subconstituents, which could explain the excess of large E_T jet found by the CDF collaboration, could be at the origin of some small parity violating effects in one-jet inclusive production. Using spin asymmetries it is possible, at RHIC, to disentangle this new effect from the Standard Model prediction due to QCD-ElectroWeak interferences. 7. AC dipole based optics measurement and correction at RHIC Shen, X; Bai, M; White, S; Robert-Domolaize, G; Luo, Y; Marusic, A; Tomas, R 2013-01-01 Independent component analysis (ICA) was applied to the AC dipole based optics measurement at RHIC to extract beta functions as well as phase advances at each BPM. Existence of excessive beta-beat was observed in both rings of RHIC at polarized proton store energy. A unique global optics correction scheme was then developed and tested successfully during the RHIC polarized proton run in 2013. The feasibility of using horizontal closed orbit bump at sextupole for arc beta-beat correction was also demonstrated. 8. Status of Proton Polarization in Rhic and AGS Mackay, W. W.; Bai, M.; Huang, H.; Ahrens, L.; Alekseev, I. G.; Bravar, A.; Brown, K.; Bunce, G.; Calaga, R.; Courant, E. D.; Drees, A.; Fischer, W.; Gardner, C.; Glenn, J. W.; Gupta, R.; Igo, G.; Iriso, U.; Jinnouchi, O.; Kurita, K.; Luccio, A. U.; Luo, Y.; Makdisi, Y.; Marr, G.; Montag, C.; Nass, A.; Okada, H.; Okamura, M.; Pilat, F.; Ptitsyn, V.; Roser, T.; Saito, N.; Satogata, T.; Spinka, H.; Stephenson, E. J.; Svirida, D. N.; Takano, J.; Tepikian, S.; Tomas, R.; Tsoupas, N.; Underwood, D.; Whitten, C.; Wood, J.; Zeijts, J. Van; Zelenski, A.; Zeno, K.; Zhang, S. Y. 2005-08-01 The Relativistic Heavy Ion Collider (RHIC) has collided protons with both transverse and longitudinal polarization at a centre-of-mass energy of 200 GeV. Future running will extend this to 500 GeV. This paper describes the methods used to accelerate and manipulate polarized proton beams in RHIC and its injectors. Special techniques include the use of a partial Siberian snake and an AC dipole in the AGS. In RHIC we use superconducting helical Siberian snakes for acceleration, and eight superconducting helical rotators for independent control of polarization directions at two interaction regions. The present status and future plans for the polarized proton program will be reviewed. 9. Status of RHIC head-on beam-beam compensation project Fischer, W.; Anerella, M.; Beebe, E.; Bruno, D.; Gassner, D.M.; Gu, X.; Gupta, R.C.; Hock, J.; Jain, A.K.; Lambiase, R.; Liu, C.; Luo, Y.; Mapes, M.; Montag, C.; Oerter, B.; Okamura, M.; Pikin, A.I.; Raparia, D.; Tan, Y.; Than, R.; Thieberger, P.; Tuozzolo, J.; Zhang, W. 2011-03-28 Two electron lenses are under construction for RHIC to partially compensate the head-on beam-beam effect in order to increase both the peak and average luminosities. The final design of the overall system is reported as well as the status of the component design, acquisition, and manufacturing. An overview of the RHIC head-on beam-beam compensation project is given in [1], and more details in [2]. With 2 head-on beam-beam interactions in IP6 and IP8, a third interaction with a low-energy electron beam is added near IP10 to partially compensate the the head-on beam-beam effect. Two electron lenses are under construction, one for each ring. Both will be located in a region common to both beams, but each lens will act only on one beam. With head-on beam-beam compensation up to a factor of two improvement in luminosity is expected together with a polarized source upgrade. The current RHIC polarized proton performance is documented in Ref. [4]. An electron lens (Fig. 1) consists of an DC electron gun, warm solenoids to focus the electron beam during transport, a superconducting main solenoid in which the interaction with the proton beam occurs, steering magnets, a collector, and instrumentation. The main developments in the last year are given below. The experimental program for polarized program at 100 GeV was expected to be finished by the time the electron lenses are commissioned. However, decadal plans by the RHIC experiments STAR and PHENIX show a continuing interest at both 100 GeV and 250 GeV, and a larger proton beam size has been accommodated in the design (Tab. 1). Over the last year beam and lattice parameters were optimized, and RHIC proton lattices are under development for optimized electron lens performance. The effect of the electron lens magnetic structure on the proton beam was evaluated, and found to be correctable. Experiments were done in RHIC and the Tevatron. 10. Cryogenic Fluid Management Technology Development Roadmaps Stephens, J. R.; Johnson, W. L. 2017-01-01 Advancement in Cryogenic Fluid Management (CFM) Technologies is essential for achieving NASA's future long duration missions. Propulsion systems utilizing cryogens are necessary to achieve mission success. Current State Of the Art (SOA) CFM technologies enable cryogenic propellants to be stored for several hours. However, some envisioned mission architectures require cryogens to be stored for two years or longer. The fundamental roles of CFM technologies are long term storage of cryogens, propellant tank pressure control and propellant delivery. In the presence of heat, the cryogens will "boil-off" over time resulting in excessive pressure buildup, off-nominal propellant conditions, and propellant loss. To achieve long term storage and tank pressure control, the CFM elements will intercept and/or remove any heat from the propulsion system. All functions are required to perform both with and without the presence of a gravitational field. Which CFM technologies are required is a function of the cryogens used, mission architecture, vehicle design and propellant tank size. To enable NASA's crewed mission to the Martian surface, a total of seventeen CFM technologies have been identified to support an In-Space Stage and a Lander/Ascent Vehicle. Recognizing that FY2020 includes a Decision Point regarding the In-Space Stage Architecture, a set of CFM Technology Development Roadmaps have been created identifying the current Technology Readiness Level (TRL) of each element, current technology "gaps", and existing technology development efforts. The roadmaps include a methodical approach and schedule to achieve a flight demonstration in FY2023, hence maturing CFM technologies to TRL 7 for infusion into the In-Space Stage Preliminary Design. 11. Operations and Performance of RHIC as a Cu-Cu Collider Pilat, Fulvia Caterina; Bai, Mei; Barton, Donald; Beebe-Wang, Joanne; Blaskiewicz, Michael; Brennan, Joseph M; Bruno, Donald; Cameron, Peter; Connolly, Roger; De Long, Joseph; Drees, Angelika; Fischer, Wolfram; Ganetis, George; Gardner, Chris J; Glenn, Joseph; Harvey, Margaret; Hayes, Thomas; Hseuh Hsiao Chaun; Huang, Haixin; Ingrassia, Peter; Iriso, Ubaldo; Lee, Roger C; Litvinenko, Vladimir N; Luo, Yun; MacKay, William W; Marr, Gregory J; Marusic, Al; Michnoff, Robert; Montag, Christoph; Morris, John; Nicoletti, Tony; Oerter, Brian; Ptitsyn, Vadim; Roser, Thomas; Russo, Thomas; Sandberg, Jon; Satogata, Todd; Schultheiss, Carl; Tepikian, Steven; Tomas, Rogelio; Trbojevic, Dejan; Tsoupas, Nicholaos; Tuozzolo, Joseph; Vetter, Kurt; Zaltsman, Alex; Zeno, Keith; Zhang, S Y; Zhang, Wu 2005-01-01 The 5th year of RHIC operations, started in November 2004 and expected to last till June 2005, consists of a physics run with Cu-Cu collisions at 100 GeV/u followed by one with polarized protons at 100 GeV. We will address here overall performance of the RHIC complex used for the first time as a Cu-Cu collider, and compare it with previous operational experience with Au, PP and asymmetric d-Au collisions. We will also discuss operational improvements, such as a ?* squeeze to 85cm in the high luminosity interaction regions from the design value of 1m, system improvements and machine performance limitations, such as vacuum pressure rise, intra-beam scattering, and beam beam interaction. 12. Strangeness production in heavy ion collisions at SPS and RHIC within two-source statistical model Lu, Z D; Fuchs, C; Zabrodin, E E; Lu, Zhong-Dao; Faessler, Amand 2002-01-01 The experimental data on hadron yields and ratios in central Pb+Pb and Au+Au collisions at SPS and RHIC energies, respectively, are analysed within a two-source statistical model of an ideal hadron gas. These two sources represent the expanding system of colliding heavy ions, where the hot central fireball is embedded in a larger but cooler fireball. The volume of the central source increases with rising bombarding energy. Results of the two-source model fit to RHIC experimental data at midrapidity coincide with the results of the one-source thermal model fit, indicating the formation of an extended fireball, which is three times larger than the corresponding core at SPS. 13. The E-lens test bench for RHIC beam-beam compensation Gu X.; Altinbas, F.Z.; Aronson, J.; Beebe, E. et al 2012-05-20 To compensate for the beam-beam effects from the proton-proton interactions at IP6 and IP8 in the Relativistic Heavy Ion Collider (RHIC), we are fabricating two electron lenses that we plan to install at RHIC IR10. Before installing the e-lenses, we are setting-up the e-lens test bench to test the electron gun, collector, GS1 coil, modulator, partial control system, some instrumentation, and the application software. Some e-lens power supplies, the electronics for current measurement will also be qualified on test bench. The test bench also was designed for measuring the properties of the cathode and the profile of the beam. In this paper, we introduce the layout and elements of the e-lens test bench; and we discuss its present status towards the end of this paper. 14. Re-visit local coupling correction in the interaction regions of RHIC Luo, Y.; Fischer, W.; Liu, C.; Marusic, A.; Minty, M.; Ptitsyn, V.; Schoefer, V.; Tepikian, S.; Trbojevic, D.; Zimmer, C. 2011-11-01 In this article we will re-visit the local coupling correction in the interaction regions (IRs) of the Relativistic Heavy Ion Collider (RHIC). We will review the measurement data of triplet quadrupole rolls, the local coupling correction strengths in the RHIC control system, and the methods for the local coupling correction with local skew quadrupole correctors. Based on the in-turnnel measurement data of triplet roll errors in 2011, we will analytically calculate and simulate IR-bump method to find out the local skew correction strengths and compare them at store and at injection with the Blue and Yellow ring lattices in the 2011 polarized proton (p-p) and Au-Au runs. The vertical dispersion from the triplet roll errors, local and global coupling correction skew quadrupoles, and the vertical dipole correctors are calculated and discussed. 15. OPERATIONS AND PERFORMANCE OF RHIC AS A CU-CU COLLIDER. PILAT, R.; AHRENS, L.; BAI, M.; BARTON, D.S.; ET AL. 2005-05-16 The 5th year of RHIC operations, started in November 2004 and expected to last till June 2005, consists of a physics run with Cu-Cu collisions at 100 GeV/u followed by one with polarized protons (pp) at 100 GeV [l]. We will address here the overall performance of the RHIC complex used for the first time as a Cu-Cu collider, and compare it with previous operational experience with Au, PP and asymmetric d-Au collisions. We will also discuss operational improvements, such as a {beta}* squeeze to 85cm in the high luminosity interaction regions from the design value of 1m, system improvements, machine performance and limitations, and address reliability and uptime issues. 16. RHIC POWER SUPPLIES - LESSONS LEARNED FROM THE 1999 - 2001 RHIC RUNS. BRUNO,D.ENG,W.GANETIS,G.LAMBIASE,R.F.LOUIE,W.SANDBERG,J.SCHULTHEISS,C. 2003-05-12 The Relativistic Heavy Ion Collider (RHIC) was commissioned in 1999 and 2000. The two RHIC rings require a total of 933 power supplies (PSs) to supply currents to highly inductive superconducting magnets. These units function as 4 main PSs, 237 insertion region (02) PSs, 24 sextupole PSs, 24 Gamma-T PSs, 8 snake PSs, 16 spin rotator PSs, and 620 correction PSs. PS reliability in this type of machine is of utmost importance because the IR PSs are nested within other IR PSs, and these are all nested within the main PSs. This means if any main or IR PS trips off due to a PS fault or quench indication, then all the IR and main PSs in that ring must follow. When this happens, the Quench Protection Assemblies (QPA's) for each unit disconnects the PSs from the circuit and absorb the stored energy in the magnets. Commissioning these power supplies and QPA's was and still is a learning experience. A summary of the major problems encountered during these first three RHIC runs will be presented along with solutions. 17. Long-Term Cryogenic Propellant Storage for the TOPS Mission Mustafi, Shuvo; Francis, John; Li, Xiaoyi; Purves, Lloyd; DeLee, Hudson; Riall, Sara; McGuinness, Dan; Willis, Dewey; Nixon, Conor; Devine Matt; 2015-01-01 Cryogenic propellants such as liquid hydrogen (LH2) and liquid oxygen (LOX) can dramatically enhance NASAs ability to explore the solar system because of their superior specific impulse (Isp) capability. Although these cryogenic propellants can be challenging to manage and store, they allow significant mass advantages over traditional hypergolic propulsion systems and are therefore technically enabling for many planetary science missions. New cryogenic storage techniques such as subcooling and the use of advanced insulation and low thermal conductivity support structures will allow for the long term storage and use of cryogenic propellants for solar system exploration and hence allow NASA to deliver more payloads to targets of interest, launch on smaller and less expensive launch vehicles, or both. Employing cryogenic propellants will allow NASA to perform missions to planetary destinations that would not be possible with the use of traditional hypergolic propellants. These new cryogenic storage technologies were implemented in a design study for the Titan Orbiter Polar Surveyor (TOPS) mission, with LH2 and LOX as propellants, and the resulting spacecraft design was able to achieve a 43 launch mass reduction over a TOPS mission, that utilized a conventional hypergolic propulsion system with mono-methyl hydrazine (MMH) and nitrogen tetroxide (NTO) propellants. This paper describes the cryogenic propellant storage design for the TOPS mission and demonstrates how these cryogenic propellants are stored passively for a decade-long Titan mission. 18. Marr, G.; Ahrens, L.; Bai, M.; Beebe-Wang, J.; Blackler, I.; Blaskiewicz, M.; Brennan, J.M.; Brown, K.A.; Bruno, D.; Butler, J.; Carlson, C.; Connolly, R.; D' Ottavio, T.; Drees, K.A.; Fedotov, A.V.; Fischer, W.; Fu, W.; Gardner, C.J.; Gassner, D.M.; Glenn, J.W.; Gu, X.; Harvey, M.; Hayes, T.; Hoff, L.; Huang, H.; Ingrassia, P.F.; Jamilkowski, J.P.; Kling, N.; Lafky, M.; Laster, J.S.; Liu, C.; Luo, Y.; Mapes, M.; Marusic, A.; Mernick, K.; Michnoff, R.J.; Minty, M.G.; Montag, C.; Morris, J.; Naylor, C.; Nemesure, S.; Polizzo, S.; Ptitsyn, V.; Robert-Demolaize, G.; Roser, T.; Sampson, P.; Sandberg, J.; Schoefer, V.; Schultheiss, C.; Severino, F.; Shrey, T.; Smith, K.; Steski, D.; Tepikian, S.; Thieberger, P.; Trbojevic, D.; Tsoupas, N.; Tuozzolo, J.E.; VanKuik, B.; Wang, G.; Wilinski, M.; Zaltsman, A.; Zeno, K.; Zhang, S.Y. 2011-09-04 Following the Fiscal Year (FY) 2010 (Run-10) Relativistic Heavy Ion Collider (RHIC) Au+Au run, RHIC experiment upgrades sought to improve detector capabilities. In turn, accelerator improvements were made to improve the luminosity available to the experiments for this run (Run-11). These improvements included: a redesign of the stochastic cooling systems for improved reliability; a relocation of 'common' RF cavities to alleviate intensity limits due to beam loading; and an improved usage of feedback systems to control orbit, tune and coupling during energy ramps as well as while colliding at top energy. We present an overview of changes to the Collider and review the performance of the collider with respect to instantaneous and integrated luminosity goals. At the conclusion of the FY 2011 polarized proton run, preparations for heavy ion run proceeded on April 18, with Au+Au collisions continuing through June 28. Our standard operations at 100 GeV/nucleon beam energy was bracketed by two shorter periods of collisions at lower energies (9.8 and 13.5 GeV/nucleon), continuing a previously established program of low and medium energy runs. Table 1 summarizes our history of heavy ion operations at RHIC. 19. Lattice design for the ERL electron ion collider in RHIC Trbojevic, D.; Beebe-Wang, J.; Tsoupas, N.; Chang, X.; Kayran, D.; Ptitsyn, V.; Litvinenko, V.; Hao, Y.; Parker, B.; Pozdeyev, E. 2010-05-23 We present electron ion collider lattice design for the Relativistic Heavy Ion Collider (eRHIC) where the electrons have multi-passes through recirculating linacs (ERL) and arcs placed in the existing RHIC tunnel. The present RHIC interaction regions (IR's), where the electron ion collisions will occur, are modified to allow for the large luminosity. Staging of eRHIC will bring the electron energy from 4 up to 20 (30) GeV as the superconducting cavities are built and installed sequentially. The synchrotron radiation from electrons at the IR is reduced as they arrive straight to the collision while ions and protons come with 10 mrad crossing angle using the crab cavities. 20. ERL Based Electron-Ion Collider eRHIC Litvinenko, Vladimir N; Bai, Mei; Beebe-Wang, Joanne; Ben-Zvi, Ilan; Blaskiewicz, Michael; Brennan, Joseph M; Calaga, Rama; Chang, Xiangyun; Deshpande, Abhay A; Farkhondeh, Manouchehr; Fedotov, Alexei V; Fischer, Wolfram; Kayran, Dmitry; Kewisch, Jorg; MacKay, William W; Montag, Christoph; Parker, Brett; Peggs, Steve; Ptitsyn, Vadim; Roser, Thomas; Ruggiero, Alessandro; Satogata, Todd; Surrow, Bernd; Tepikian, Steven; Trbojevic, Dejan; Yakimenko, Vitaly; Zhang, S Y 2005-01-01 We present the designs of a future polarized electron-hadron collider, eRHIC* based on a high current super-conducting energy-recovery linac (ERL) with energy of electrons up to 20 GeV. We plan to operate eRHIC in both dedicated (electron-hadrons only) and parallel(with the main hadron-hadron collisions) modes. The eRHIC has very large tunability range of c.m. energies while maintaining very high luminosity up to 1034 cm-2 s-1 per nucleon. Two of the most attractive features of this scheme are full spin transparency of the ERL at all operational energies and the capability to support up to four interaction points. We present two main layouts of the eRHIC, the expected beam and luminosity parameter, and discuss the potential limitation of its performance. 1. PROGRESS OF HIGH-ENERGY ELECTRON COOLING FOR RHIC. FEDOTOV,A.V. 2007-09-10 The fundamental questions about QCD which can be directly answered at Relativistic Heavy Ion Collider (RHIC) call for large integrated luminosities. The major goal of RHIC-I1 upgrade is to achieve a 10 fold increase in luminosity of Au ions at the top energy of 100 GeV/nucleon. Such a boost in luminosity for RHIC-II is achievable with implementation of high-energy electron cooling. The design of the higher-energy cooler for RHIC-II recently adopted a non-magnetized approach which requires a low temperature electron beam. Such electron beams will be produced with a superconducting Energy Recovery Linac (ERL). Detailed simulations of the electron cooling process and numerical simulations of the electron beam transport including the cooling section were performed. An intensive R&D of various elements of the design is presently underway. Here, we summarize progress in these electron cooling efforts. 2. Advanced Devices for Cryogenic Thermal Management Bugby, D.; Stouffer, C.; Garzon, J.; Beres, M.; Gilchrist, A. 2006-04-01 This paper describes six advanced cryogenic thermal management devices/subsystems developed by Swales Aerospace for ground/space-based applications of interest to NASA, DoD, and the commercial sector. The devices/subsystems described herein include the following: (a) a differential thermal expansion cryogenic thermal switch (DTE-CTSW) constructed with high purity aluminum end-pieces and an Ultem support rod for the 6 K Mid-Infrared Instrument (MIRI) on the James Webb Space Telescope (JWST) (b) a quad-redundant DTE-CTSW assembly for the 35 K science instruments (NIRCam, NIRSpec, and FGS) mounted on the JWST Integrated Science Instrument Module (ISIM) (c) a cryogenic diode heat pipe (CDHP) thermal switching system using methane as the working fluid for the 100 K CRISM hyperspectral mapping instrument on the Mars Reconnaissance Orbiter (MRO) and (d) three additional devices/subsystems developed during the AFRL-sponsored CRYOTOOL program, which include a dual DTE-CTSW/dual cryocooler test bed, a miniaturized neon cryogenic loop heat pipe (mini-CLHP), and an across gimbal cryogenic thermal transport system (GCTTS). For the first three devices/subsystems mentioned above, this paper describes key aspects of the development efforts including concept definition, design, fabrication, and testing. For the latter three, this paper provides brief overview descriptions as key details are provided in a related paper. 3. In situ cryogenic Raman spectroscopic studies on the synthetic fluid inclusions in the systems H2O and NaCl-H2O 2006-01-01 Salt-hydrates have diagnostic cryogenic Raman spectra, which can reflect the composition of the parent solutions. As analogue to the natural fluid inclusions, the synthetic inclusions can be used to validate numerous assumptions related to fluid inclusion research. They can also be used to test the feasibility of application of laser Raman spectroscopy to individual fluid inclusion analysis. Using the technique proposed by Sterner and Bodnar(1984), synthetic inclusions of the systems H2O and NaCl-H2O (with NaCl as 5.12 wt%, 9.06 wt%, 16.6 wt% and 25 wt%) were formed under the pressures from 50Mpa to 100Mpa and at the temperatures from 500℃ to 600℃. In situ cryogenic Raman spectra were collected at about -180℃ by combined use of freezing- heating stage and Laser Raman Microscopy. It is shown that hydrohalite (NaCl·2H2O), the salt-hydrate of NaCl in the fluid inclusions has the specific Raman spectrum and can be used as the standard to verify the existence of NaCl in the aqueous inclusions. The crystalline ice other than amorphous ice (glasses) formed from the aqueous phase whthin the synthetic inclusions during the initial freezing, but hydrohalite did not form. Subsequent warming of these inclusions induced a phase change, typically between approximately -40 and -22℃, that represents the hydrohalite crystallization event but not a eutectic melting event. So, for fluid inclusions in the system NaCl-H2O, interpretation of phase behavior below the eutectic temperature (-20.8℃) should be made with caution. The ratios of the relative intensity and the area of Raman spectra between 3423 cm-1 peak of hydrohalite and 3098 cm-1 peak of ice show positive correlations to the salinities in aqueous inclusions, which can be used to determine the salinity of NaCl- H2O system inclusions. 4. Material and structural mechanical modelling and reliability of thin-walled bellows at cryogenic temperatures. Application to LHC compensation system Garion, Cédric; Skoczen, Blazej The present thesis is dedicated to the behaviour of austenitic stainless steels at cryogenic temperatures. The plastic strain induced martensitic transformation and ductile damage are taken into account in an elastic-plastic material modelling. The kinetic law of →’ transformation and the evolution laws of kinematic/isotropic mixed hardening are established. Damage issue is analysed by different ways: mesoscopic isotropic or orthotropic model and a microscopic approach. The material parameters are measured from 316L fine gauge sheet at three levels of temperature: 293 K, 77 K and 4.2 K. The model is applied to thin-walled corrugated shell, used in the LHC interconnections. The influence of the material properties on the stability is studied by a modal analysis. The reliability of the components, defined by the Weibull distribution law, is analysed from fatigue tests. The impact on reliability of geometrical imperfections and thermo-mechanical loads is also analysed. 5. Physics at Relativistic Heavy Ion Collider (RHIC) Shuryak, E.V. 1990-08-01 This introductory talk contains a brief discussion of future experiments at RHIC related to physics of superdense matter. In particular, we consider the relation between space-time picture of the collision and spectra of the observed secondaries. We discuss where one should look for QGP signals and for possible manifestation of the phase transition. We pay more attention to a rather new topic: hadron modification in the gas phase, which is interesting by itself as a collective phenomenon, and also as a precursor indicating what happens with hadrons near the phase transition. We briefly review current understanding of the photon physics, dilepton production, charm and strangeness and J/{psi} suppression. At the end we try to classify all possible experiments. 47 refs., 3 figs. 6. Heavy Ion results from RHIC-BNL Esumi Shinlchi 2013-05-01 Full Text Available Recent results from heavy ion collision experiments from RHIC at BNL are presented and discussed in terms of Quark Gluon Plasm properties, such as partonic collectivity and partonic energy loss. The experimental results with direct photons and heavy quarks have given important additional insights of the plasma on top of what has been known with light hadrons. Higher order event anisotropies and the related results have provided the geometrical, temporal and dynamical information of the plasma. The beam energy dependence of the various measurements could reveal the structure of QCD phase diagram and possibly the critical point in the diagram, where the properties of phase transition are expected to change drastically. 7. Hadronization via coalescence at RHIC and LHC Minissale V. 2016-01-01 Full Text Available An hadronization model that includes coalescence and fragmentation is used in this work to obtain predictions at both RHIC and LHC energy for light and strange hadrons transverse momentum spectra (π, p, k, Λ and baryon to meson ratios (p/π, Λ/k in a wide range of pT. This is accomplished without changing coalescence parameters. The ratios p/π and Λ/K shows the right behaviour except for some lack of baryon yield in a limited pT range around 6 GeV. This would indicate that the AKK fragmentation functions is too flat at pT < 8 GeV. 8. Code generation of RHIC accelerator device objects Olsen, R.H.; Hoff, L.; Clifford, T. 1995-12-01 A RHIC Accelerator Device Object is an abstraction which provides a software view of a collection of collider control points known as parameters. A grammar has been defined which allows these parameters, along with code describing methods for acquiring and modifying them, to be specified efficiently in compact definition files. These definition files are processed to produce C++ source code. This source code is compiled to produce an object file which can be loaded into a front end computer. Each loaded object serves as an Accelerator Device Object class definition. The collider will be controlled by applications which set and get the parameters in instances of these classes using a suite of interface routines. Significant features of the grammar are described with details about the generated C++ code. 9. Tracking studies in eRHIC energy-recovery recirculator Meot, F. [Brookhaven National Lab. (BNL), Upton, NY (United States); Brooks, S. [Brookhaven National Lab. (BNL), Upton, NY (United States); Ptitsyn, V. [Brookhaven National Lab. (BNL), Upton, NY (United States); Trbojevic, D. [Brookhaven National Lab. (BNL), Upton, NY (United States); Tsoupas, N. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-07-13 Beam and polarization tracking studies in eRHIC energy recovery electron recirculator are presented, based on a very preliminary design of the FFAG lattice. These simulations provide examples of some of the beam and spin optics aspects of the linear FFAG lattice concept and its application in eRHIC, they provide code benchmarking for synchrotron radiation and spin diffusion in addition, and pave the way towards end-to-end 6-D(phasespace)+3D(spin) tracking simulations. 10. Polarization simulations in the RHIC run 15 lattice Meot, F. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Huang, H. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Luo, Y. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Ranjbar, V. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; Robert-Demolaize, G. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept.; White, S. [Brookhaven National Lab. (BNL), Upton, NY (United States). Collider-Accelerator Dept. 2015-05-03 RHIC polarized proton Run 15 uses a new acceleration ramp optics, compared to RHIC Run 13 and earlier runs, in relation with electron-lens beam-beam compensation developments. The new optics induces different strengths in the depolarizing snake resonance sequence, from injection to top energy. As a consequence, polarization transport along the new ramp has been investigated, based on spin tracking simulations. Sample results are reported and discussed. 11. COOLING DYNAMICS STUDIES AND SCENARIOS FOR THE RHIC COOLER. FEDOTOV,A.V.; BEN-ZVI,I.; LITVINENKO, V. 2005-05-16 In this paper, we discuss various electron cooling dynamics studies for RHIC. We also present simulations [1] of various possibilities of using electron cooling at RHIC, which includes cooling at the top energy, pre-cooling at low energy, aspects of transverse and longitudinal cooling and their impact on the luminosity. Electron cooling at various collision energies both for heavy ions and protons is also discussed. 12. MEASURING LOCAL GRADIENT AND SKEW QUADRUPOLE ERRORS IN RHIC IRS. CARDONA,J.; PEGGS,S.; PILAT,R.; PTITSYN,V. 2004-07-05 The measurement of local linear errors at RHIC interaction regions using an ''action and phase'' analysis of difference orbits has already been presented. This paper evaluates the accuracy of this technique using difference orbits that were taken when known gradient errors and skew quadrupole errors were intentionally introduced. It also presents action and phase analysis of simulated orbits when controlled errors are intentionally placed in a RHIC simulation model. 13. Thermal and prompt photons at RHIC and the LHC Paquet, Jean-François [Department of Physics & Astronomy, Stony Brook University, Stony Brook, NY 11794 (United States); Department of Physics, McGill University, 3600 University Street, Montreal, Quebec, H3A2T8 (Canada); Shen, Chun [Department of Physics, McGill University, 3600 University Street, Montreal, Quebec, H3A2T8 (Canada); Denicol, Gabriel [Department of Physics, McGill University, 3600 University Street, Montreal, Quebec, H3A2T8 (Canada); Physics Department, Brookhaven National Laboratory, Upton, NY 11973 (United States); Luzum, Matthew [Universidade de Santiago de Compostela, E-15706 Santiago de Compostela, Galicia-Spain (Spain); Universidade de São Paulo, Rua do Matão Travessa R, no. 187, 05508-090, Cidade Universitária, São Paulo (Brazil); Schenke, Björn [Physics Department, Brookhaven National Laboratory, Upton, NY 11973 (United States); Jeon, Sangyong; Gale, Charles [Department of Physics, McGill University, 3600 University Street, Montreal, Quebec, H3A2T8 (Canada) 2016-12-15 Thermal and prompt photon production in heavy ion collisions is evaluated and compared with measurements from both RHIC and the LHC. An event-by-event hydrodynamical model of heavy ion collisions that includes shear and bulk viscosities is used, along with up-to-date photon emission rates. Larger tension with measurements is observed at RHIC than at the LHC. The center-of-mass energy and centrality dependence of thermal and prompt photons is investigated. 14. Study on the Principle Mechanisms of Heat Transfer for Cryogenic Insulations: Especially Accounting for the Temperature-Dependent Deposition-Evacuation of the Filling Gas (Self-Evacuating Systems) Geisler, Matthias; Vidi, Stephan; Ebert, Hans-Peter 2016-11-01 This study concentrates on the principles of heat transfer within cryogenic insulation systems, especially accounting for self-evacuating systems (deposition-evacuation of the filling gas). These principles allow the extrapolation to other temperatures, gases and other materials with the input of only a few experimentally derived or carefully estimated material properties. The type of gas (e.g. air or CO2) within the porous insulation material dominates the behaviour of the effective thermal conductivity during the cooldown of the cryogenic application. This is due to the specific temperature-dependent saturation gas pressure which determines the contribution of the gas conductivity. The selected material classes include powders, fibrous insulations, foams, aerogels and multilayer insulations in the temperature range of 20 K to 300 K. Novel within this study is an analytical function for the total and the mean thermal conductivity with respect to the temperature, type of gas, external pressure and material class of the insulation. Furthermore, the integral mean value of the thermal conductivity, the so-called mean thermal conductivity, is calculated for a mechanically evacuated insulation material and an insulation material evacuated by deposition-evacuation of the filling gas, respectively. This enables a comparison of the total thermal conductivity of cryogenic insulation materials and their applicability for a self-evacuating cryogenic insulation system. 15. Centrality Dependent Studies of Identified Particle Spectra at RHIC Bekele, Selemon 2008-10-01 We present preliminary results from the BRAHMS experiment on identified particle spectra and ratios at y ˜0 and y ˜3 as a function of centrality for 200 GeV/NN Cu+Cu collisions. By comparing the Cu+Cu data with earlier results for the Au+Au and d+Au systems, it is possible to study how the heavy-ion reaction dynamics for a given number of participants depends on the overall system size. Particle yields, , and particle ratios are studied as a function of the number of participants. Transverse momentum distributions provide information on the final stages of the collision evolution at kinetic freeze-out. The kinetic freeze-out parameters of the Cu+Cu system are studied as a function of centrality by a simultaneous blast-wave model fit to the pion, kaon and (anti)proton spectra. The Cu+Cu results will be compared to other collision systems at RHIC to unravel the dependence on system size. 16. What RHIC Experiments and Theory tell us about Properties of Quark-Gluon Plasma ? Shuryak, E V 2004-01-01 This brief review summarizes the main experimental discoveries made at RHIC and then discusses their implications. The robust collective flow phenomena are well described by ideal hydrodynamics, with the Equation of State (EoS) predicted by lattice simulations. However the transport properties turned out to be unexpected, with rescattering cross section one-to-two orders of magnitude larger than expected from perturbative QCD. These and other theoretical developments indicate that Quark-Gluon Plasma (QGP) produced at RHIC, and probably in a wider temperature regionT_csystems'', (i) the strongly coupled supersymmetric theories studied via Maldacena duality; (ii) trapped ultra-cold atoms with very large scattering length, we return to sQGP and show that there should exist literally hundreds of bound states in it in the RHIC domain, most them... 17. Low-mass Drift Chambers of the PHENIX central spectrometers at RHIC Riabov, Y 2002-01-01 Beginning of regular operation of Relativistic Heavy Ion Collider (RHIC) at Brookhaven National Laboratory opened unique possibilities for the study of nuclear matter at unprecedentedly high energies. PHENIX is one of the two large-scale experiments at RHIC, which is intended for detection and study of new state of matter--the so called quark-gluon plasma. The basic element of the PHENIX Central Tracking System is low-mass focusing Drift Chamber (DC). Distinctive features of ion-ion collisions at RHIC impose specific requirements on the DC which could not be satisfied by any structure developed earlier. DC should provide high efficiency of track reconstruction with use of 'hot' gas mixture under conditions of high charged particle densities, high collision rates and large-scale of the detector (6 m sup 3). Detailed description of the DC and its original wire structure characterized by the absence of left-right ambiguity and controlled charge collection geometry are given. Parameters of the DC during the first... 18. Cryogenic Hybrid Magnetic Bearing Meeks, Crawford R.; Dirusso, Eliseo; Brown, Gerald V. 1994-01-01 Cryogenic hybrid magnetic bearing is example of class of magnetic bearings in which permanent magnets and electromagnets used to suspend shafts. Electromagnets provide active control of position of shaft. Bearing operates at temperatures from -320 degrees F (-196 degrees C) to 650 degrees F (343 degrees C); designed for possible use in rocket-engine turbopumps, where effects of cryogenic environment and fluid severely limit lubrication of conventional ball bearings. This and similar bearings also suitable for terrestrial rotating machinery; for example, gas-turbine engines, high-vacuum pumps, canned pumps, precise gimbals that suspend sensors, and pumps that handle corrosive or gritty fluids. 19. Cryogenic regenerative heat exchangers Ackermann, Robert A 1997-01-01 An in-depth survey of regenerative heat exchangers, this book chronicles the development and recent commercialization of regenerative devices for cryogenic applications. Chapters cover historical background, concepts, practical applications, design data, and numerical solutions, providing the latest information for engineers to develop advanced cryogenic machines. The discussions include insights into the operation of a regenerator; descriptions of the cyclic and fluid temperature distributions in a regenerator; data for various matrix geometries and materials, including coarse and fine bronze, stainless steel-woven wire mesh screens, and lead spheres; and unique operating features of cryocoolers that produce deviations from ideal regenerator theory. 20. Cryogenic Applications of Commercial Electronic Components Buchanan, Ernest D.; Benford, Dominic J.; Forgione, Joshua B.; Moseley, S. Harvey; Wollack, Edward J. 2012-01-01 We have developed a range of techniques useful for constructing analog and digital circuits for operation in a liquid Helium environment (4.2K), using commercially available low power components. The challenges encountered in designing cryogenic electronics include finding components that can function usefully in the cold and possess low enough power dissipation so as not to heat the systems they are designed to measure. From design, test, and integration perspectives it is useful for components to operate similarly at room and cryogenic temperatures; however this is not a necessity. Some of the circuits presented here have been used successfully in the MUSTANG and in the GISMO camera to build a complete digital to analog multiplexer (which will be referred to as the Cryogenic Address Driver board). Many of the circuit elements described are of a more general nature rather than specific to the Cryogenic Address Driver board, and were studied as a part of a more comprehensive approach to addressing a larger set of cryogenic electronic needs. 1. Cryogenics for HL-LHC Tavian, L; Claudet, S; Ferlin, G; Wagner, U; van Weelderen, R 2015-01-01 The discovery of a Higgs boson at CERN in 2012 is the start of a major program of work to measure this particle's properties with the highest possible precision for testing the validity of the Standard Model and to search for further new physics at the energy frontier. The LHC is in a unique position to pursue this program. Europe's top priority is the exploitation of the full potential of the LHC, including the high-luminosity upgrade of the machine and detectors with an objective to collect ten times more data than in the initial design, by around 2030. To reach this objective, the LHC cryogenic system must be upgraded to withstand higher beam current and higher luminosity at top energy while keeping the same operation availability by improving the collimation system and the protection of electronics sensitive to radiation. This chapter will present the conceptual design of the cryogenic system upgrade with recent updates in performance requirements, the corresponding layout and architecture of the system a... 2. Cryogenic hydrogen-induced air liquefaction technologies Escher, William J. D. 1990-01-01 Extensively utilizing a special advanced airbreathing propulsion archives database, as well as direct contacts with individuals who were active in the field in previous years, a technical assessment of cryogenic hydrogen-induced air liquefaction, as a prospective onboard aerospace vehicle process, was performed and documented. The resulting assessment report is summarized. Technical findings are presented relating the status of air liquefaction technology, both as a singular technical area, and also that of a cluster of collateral technical areas including: compact lightweight cryogenic heat exchangers; heat exchanger atmospheric constituents fouling alleviation; para/ortho hydrogen shift conversion catalysts; hydrogen turbine expanders, cryogenic air compressors and liquid air pumps; hydrogen recycling using slush hydrogen as heat sink; liquid hydrogen/liquid air rocket-type combustion devices; air collection and enrichment systems (ACES); and technically related engine concepts. 3. Temperature Sensing Solution for Cryogenic Space Engines Project National Aeronautics and Space Administration — Cryogenic systems, heavily used in rocket ground testing, space station operations, shuttle launch systems, etc, require a large number of temperature sensors for... 4. Cryogenics Research and Engineering Experience Toro Medina, Jaime A. 2013-01-01 Energy efficient storage, transfer and use of cryogens and cryogenic propellants on Earth and in space have a direct impact on NASA, government and commercial programs. Research and development on thermal insulation, propellant servicing, cryogenic components, material properties and sensing technologies provides industry, government and research institutions with the cross-cutting technologies to manage low-temperature applications. Under the direction of the Cryogenic Testing Lab at Kennedy Space Center, the work experience acquired allowed me to perform research, testing, design and analysis of current and future cryogenic technologies to be applied in several projects. 5. Seals For Cryogenic Turbomachines Hendricks, Robert C.; Tam, L. T.; Braun, M. J.; Vlcek, B. L. 1988-01-01 Analysis considers effects of seals on stability. Report presents method of calculation of flows of cryogenic fluids through shaft seals. Key to stability is local average velocity in seal. Local average velocity strongly influenced by effects of inlet and outlet and injection of fluid. 6. High Power Cryogenic Targets Gregory Smith 2011-08-01 The development of high power cryogenic targets for use in parity violating electron scattering has been a crucial ingredient in the success of those experiments. As we chase the precision frontier, the demands and requirements for these targets have grown accordingly. We discuss the state of the art, and describe recent developments and strategies in the design of the next generation of these targets. 7. Preliminary Thermal Design of Cryogenic Radiation Shielding Li, Xiaoyi; Mustafi, Shuvo; Boutte, Alvin 2015-01-01 Cryogenic Hydrogen Radiation Shielding (CHRS) is the most mass efficient material radiation shielding strategy for human spaceflight beyond low Earth orbit (LEO). Future human space flight, mission beyond LEO could exceed one year in duration. Previous radiation studies showed that in order to protect the astronauts from space radiation with an annual allowable radiation dose less than 500 mSv, 140 kgm2 of polyethylene is necessary. For a typical crew module that is 4 meter in diameter and 8 meter in length. The mass of polyethylene radiation shielding required would be more than 17,500 kg. The same radiation study found that the required hydrogen shielding for the same allowable radiation dose is 40 kgm2, and the mass of hydrogen required would be 5, 000 kg. Cryogenic hydrogen has higher densities and can be stored in relatively small containment vessels. However, the CHRS system needs a sophisticated thermal system which prevents the cryogenic hydrogen from evaporating during the mission. This study designed a cryogenic thermal system that protects the CHRS from hydrogen evaporation for one to up to three year mission. The design also includes a ground based cooling system that can subcool and freeze liquid hydrogen. The final results show that the CHRS with its required thermal protection system is nearly half of the mass of polyethylene radiation shielding. 8. Cryogenic Adsorption of Nitrogen and Carbon Dioxide in Activated Carbon Shen, Fuzhi; Liu, Huiming; Xu, Dong; Zhang, Hengcheng; Lu, Junfeng; Li, Laifeng 2017-09-01 Activated carbon have been used for a long time at low temperature for cryogenic applications. The knowledge of adsorption characteristics of activated carbon at cryogenic temperature is essential for some specific applications. However, such experimental data are very scare in the literature. In order to measure the adsorption characteristics of activated carbon under variable cryogenic temperatures, an adsorption measurement device was presented. The experiment system is based on the commercially available PCT-pro adsorption analyzer coupled to a two-stage Gifford McMahon refrigerator, which allows the sample to be cooled to 4.2K. Cryogenic environment can be maintained steadily without the cryogenic liquid through the cryocooler and temperature can be controlled precisely between 5K and 300K by the temperature controller. Adsorption measurements were performed in activated carbon for carbon dioxide and nitrogen and the adsorption isotherm were obtained. 9. The LHC cryogenic operation for first collisions and physics run Brodzinski, K; Benda, V; Bremer, J; Casas-Cubillos, J; Claudet, S; Delikaris, D; Ferlin, G; Fernandez Penacoba, G; Perin, A; Pirotte, O; Soubiran, M; Tavian, L; van Weelderen, R; Wagner, U 2011-01-01 The Large Hadron Collider (LHC) cryogenic system was progressively and successfully run for the LHC accelerator operation period starting from autumn 2009. The paper recalls the cryogenic system architecture and main operation principles. The system stability during magnets powering and availability periods for high energy beams with first collisions at 3.5 TeV are presented. Treatment of typical problems, weak points of the system and foreseen future consolidations will be discussed. 10. Application of computational fluid dynamics for the simulation of cryogenic molecular sieve bed absorber of hydrogen isotopes recovery system for Indian LLCB-TBM Gayathri Devi, V.; Sircar, A.; Sarkar, B. [Institute of Plasma Research, Bhat, Gandhinagar, Gujarar (India) 2015-03-15 One of the most challenging tasks in the design of the fuel cycle system lies in the effective design of Tritium Extraction System (TES) which involves proper extraction and purification of tritium in the fuel cycle of the fusion reactor. Indian Lead Lithium cooled Ceramic Breeder Test Blanket Module (LLCB-TBM) would extract hydrogen isotopes through Cryogenic Molecular Sieve Bed (CMSB) adsorber system. A prototype Hydrogen Isotopes Recovery System (HIRS) is being developed to validate the concepts for tritium extraction by adsorption mass transfer mechanism. In this study, a design model has been developed and analyzed to simulate the adsorption mass transfer kinetics in a fixed bed adsorption column. The simulation leads primarily to effective design of HIRS, which is a state-of-the-art technology. The paper describes the process simulation approach and the results of Computational Fluid Dynamics (CFD) analysis. The effects of different operating conditions are studied to investigate their influence on the hydrogen isotopes adsorption capacity. The results of the present simulation study would be used to understand the best optimized transport phenomenon before realizing the TES as a system for LLCB-TBM. (authors) 11. Polarized proton parameters for the 2015 PP-on-Au setup in RHIC Gardner, C. J. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-08-25 Values are given for RHIC circumference shifts due to snakes for various situations. Relevant parameters are tabulated for polarized protons (PP) in the booster and in AGS and RHIC for PP-on-Au stores. 12. Polarized proton parameters for the 2015 PP-on-Aluminum setup in RHIC Gardner, C. J. [Brookhaven National Lab. (BNL), Upton, NY (United States) 2015-10-02 Values are given for RHIC circumference shifts due to snakes for various situations. Relevant parameters are tabulated for polarized protons (PP) in the booster and in AGS and RHIC for PP-on-Aluminum stores. 13. Energy Efficient Storage and Transfer of Cryogens Fesmire, James E. 2013-01-01 Cryogenics is globally linked to energy generation, storage, and usage. Thermal insulation systems research and development is an enabling part of NASA's technology goals for Space Launch and Exploration. New thermal testing methodologies and materials are being transferred to industry for a wide range of commercial applications. 14. The cryogenic storage ring CSR. von Hahn, R; Becker, A; Berg, F; Blaum, K; Breitenfeldt, C; Fadil, H; Fellenberger, F; Froese, M; George, S; Göck, J; Grieser, M; Grussie, F; Guerin, E A; Heber, O; Herwig, P; Karthein, J; Krantz, C; Kreckel, H; Lange, M; Laux, F; Lohmann, S; Menk, S; Meyer, C; Mishra, P M; Novotný, O; O'Connor, A P; Orlov, D A; Rappaport, M L; Repnow, R; Saurabh, S; Schippers, S; Schröter, C D; Schwalm, D; Schweikhard, L; Sieber, T; Shornikov, A; Spruck, K; Sunil Kumar, S; Ullrich, J; Urbain, X; Vogel, S; Wilhelm, P; Wolf, A; Zajfman, D 2016-06-01 An electrostatic cryogenic storage ring, CSR, for beams of anions and cations with up to 300 keV kinetic energy per unit charge has been designed, constructed, and put into operation. With a circumference of 35 m, the ion-beam vacuum chambers and all beam optics are in a cryostat and cooled by a closed-cycle liquid helium system. At temperatures as low as (5.5 ± 1) K inside the ring, storage time constants of several minutes up to almost an hour were observed for atomic and molecular, anion and cation beams at an energy of 60 keV. The ion-beam intensity, energy-dependent closed-orbit shifts (dispersion), and the focusing properties of the machine were studied by a system of capacitive pickups. The Schottky-noise spectrum of the stored ions revealed a broadening of the momentum distribution on a time scale of 1000 s. Photodetachment of stored anions was used in the beam lifetime measurements. The detachment rate by anion collisions with residual-gas molecules was found to be extremely low. A residual-gas density below 140 cm(-3) is derived, equivalent to a room-temperature pressure below 10(-14) mbar. Fast atomic, molecular, and cluster ion beams stored for long periods of time in a cryogenic environment will allow experiments on collision- and radiation-induced fragmentation processes of ions in known internal quantum states with merged and crossed photon and particle beams. 15. The cryogenic storage ring CSR von Hahn, R.; Becker, A.; Berg, F.; Blaum, K.; Breitenfeldt, C.; Fadil, H.; Fellenberger, F.; Froese, M.; George, S.; Göck, J.; Grieser, M.; Grussie, F.; Guerin, E. A.; Heber, O.; Herwig, P.; Karthein, J.; Krantz, C.; Kreckel, H.; Lange, M.; Laux, F.; Lohmann, S.; Menk, S.; Meyer, C.; Mishra, P. M.; Novotný, O.; O'Connor, A. P.; Orlov, D. A.; Rappaport, M. L.; Repnow, R.; Saurabh, S.; Schippers, S.; Schröter, C. D.; Schwalm, D.; Schweikhard, L.; Sieber, T.; Shornikov, A.; Spruck, K.; Sunil Kumar, S.; Ullrich, J.; Urbain, X.; Vogel, S.; Wilhelm, P.; Wolf, A.; Zajfman, D. 2016-06-01 An electrostatic cryogenic storage ring, CSR, for beams of anions and cations with up to 300 keV kinetic energy per unit charge has been designed, constructed, and put into operation. With a circumference of 35 m, the ion-beam vacuum chambers and all beam optics are in a cryostat and cooled by a closed-cycle liquid helium system. At temperatures as low as (5.5 ± 1) K inside the ring, storage time constants of several minutes up to almost an hour were observed for atomic and molecular, anion and cation beams at an energy of 60 keV. The ion-beam intensity, energy-dependent closed-orbit shifts (dispersion), and the focusing properties of the machine were studied by a system of capacitive pickups. The Schottky-noise spectrum of the stored ions revealed a broadening of the momentum distribution on a time scale of 1000 s. Photodetachment of stored anions was used in the beam lifetime measurements. The detachment rate by anion collisions with residual-gas molecules was found to be extremely low. A residual-gas density below 140 cm-3 is derived, equivalent to a room-temperature pressure below 10-14 mbar. Fast atomic, molecular, and cluster ion beams stored for long periods of time in a cryogenic environment will allow experiments on collision- and radiation-induced fragmentation processes of ions in known internal quantum states with merged and crossed photon and particle beams. 16. The Nuclear Cryogenic Propulsion Stage Houts, Michael G.; Kim, Tony; Emrich, William J.; Hickman, Robert R.; Broadway, Jeramie W.; Gerrish, Harold P.; Doughty, Glen; Belvin, Anthony; Borowski, Stanley K.; Scott, John 2014-01-01 The fundamental capability of Nuclear Thermal Propulsion (NTP) is game changing for space exploration. A first generation Nuclear Cryogenic Propulsion Stage (NCPS) based on NTP could provide high thrust at a specific impulse above 900 s, roughly double that of state of the art chemical engines. Characteristics of fission and NTP indicate that useful first generation systems will provide a foundation for future systems with extremely high performance. The role of the NCPS in the development of advanced nuclear propulsion systems could be analogous to the role of the DC-3 in the development of advanced aviation. Progress made under the NCPS project could help enable both advanced NTP and advanced Nuclear Electric Propulsion (NEP). Nuclear propulsion can be affordable and viable compared to other propulsion systems and must overcome a biased public fear due to hyper-environmentalism and a false perception of radiation and explosion risk. 17. Fulfilling the RHIC mission with sPHENIX Connors, Megan 2016-08-01 RHIC has made groundbreaking contributions to the understanding of QCD under extreme conditions with the discovery of the quark gluon plasma (QGP) as a perfect fluid and first observations of energy loss. It continues to play a crucial role in understanding and quantifying the properties of the QGP as well as mapping out the QCD phase diagram. However, detailed questions concerning partonic energy loss in the QGP remain. There is a need to build a new detector at RHIC to measure important rare probes of the QGP. A new detector will benefit from advances in reconstructing jets in heavy ion collisions and the increased luminosity achievable with RHIC. Constraining models at RHIC and LHC energies are crucial for extracting the temperature dependence of transport properties of the QGP. To measure newly developed observables made at the LHC with high precision at RHIC, a detector with full azimuthal coverage and spanning a pseudorapidity range between -1.1 and 1.1, known as sPHENIX, has been proposed. The capabilities of the new detector will allow for a full understanding of jet energy loss and upsilon suppression. The goals for sPHENIX and route to achieving these goals along with the current status of the detector will be presented on behalf of the new collaboration. 18. PHOBOS at RHIC: Some global observations Alan S Carroll; B B Back; M D Baker; D S Barton; R R Betts; M Ballintijn; A A Bickley; R Bindel; A Budzanowski; W Busza; A Carroll; M P Decowski; E García; N George; K Gulbrandsen; S Gushue; C Halliwell; J Hamblen; G A Heintzelman; C Henderson; D J Hofman; R S Hollis; R Hoyłyński; B Holzman; A Iordanova; E Johnson; J L Kane; J Katzy; N Khan; W Kucewicz; P Kulinich; C M Kuo; W T Lin; S Manly; D McLeod; J Michałowski; A C Mignerey; R Nouicer; A Olszewski; R Pak; I C Park; H Pernegger; C Reed; L P Remsberg; M Reuter; C Roland; G Roland; L Rosenberg; J Sagerer; P Sarin; P Sawicki; W Skulski; S G Steadman; P Steinberg; G S F Stephans; M Stodulski; A Sukhanov; J-L Tang; R Teng; A Trzupek; C Vale; G J van Nieuwenhuizen; R Verdier; B Wadsworth; F L H Wolfs; B Wosiek; K Woźniak; A H Wuosmaa; B Wysłouch; For the PHOBOS Collaboration 2003-11-01 Particle production in Au+Au collisions has been measured in the PHOBOS experiment at RHIC for a range of collision energies for a large span of pseudorapidities, || < 5.4. Three empirical observations have emerged from this data set which require theoretical examination. First, there is clear evidence of limiting fragmentation. Namely, particle production in central Au+Au collisions, when expressed as d/d' (' ≡ -beam), becomes energy independent at high energy for a broad region of ' around '=0. This energy-independent region grows with energy, allowing only a limited region (if any) of longitudinal boost-invariance. Second, there is a striking similarity between particle production in +- and Au + Au collisions (scaled by the number of participating nucleon pairs). Both the total number of produced particles and the longitudinal distribution of produced particles are approximately the same in +- and in scaled Au + Au. This observation was not predicted and has not been explained. Finally, particle production has been found to scale approximately with the number of participating nucleon pairs for $\\langle N_{\\text{part}}\\rangle > 65$. This scaling occurs both for the total multiplicity and for high T particles (3 < T < 4.5 GeV/c). 19. Possibilities for stochastic cooling at RHIC Brennan, J M; Wei, J 2004-01-01 Intra-Beam Scattering (IBS) is the fundamental performance limitation for RHIC. The emittance growth from IBS determines the ultimate luminosity lifetime and the only cure is cooling. Full-energy electron cooling will be installed to not only control emittance growth but also reduce emittances during a store. Before that, stochastic cooling could increase integrated luminosity by momentum cooling. Two significant benefits would follow; the average luminosity in a 10 h store would double, and the problem of coasting beam in the abort gap would be solved. Of course high-frequency bunched beam stochastic cooling is required and previous attempts at this at the Tevatron and SPS were not successful. It appears that the conditions in the heavy ion collider are more favorable. First, the high charge state of ions gives better signal to noise ratio in the Schottky signal. Second, the anomalous coherent components in the pick up signals that caused saturation in the electronics in previous attempts are greatly reduced... 20. Particle Production at RHIC and LHC Energies Tawfik, A; Shalaby, A G 2012-01-01 The production of different particle species is recently measured in $Pb-Pb$ collisions by the ALICE experiment at $\\sqrt{s}=7$TeV. This motivates the use of various bosons and baryons measured at lower center-of-mass energies in comparing their ratios to the hadron resonance (HRG) gas model and PYTHIA event generator. It is found that the particle-to-antiparticle ratios are perfectly reproduce by means of HRG and PYTHIA at RHIC and LHC energies. The kaon-to-pion and proton-to-pion ratios are entirely overestimated by the HRG model. The PYTHIA event generator obviously underestimates the kaon-to-pion ratio and simultaneously reproduces the proton-to-pion ratio, almost perfectly, especially at LHC energy. While matter-to-antimatter and non-strange abundances are partly in line with predictions from the HRG model, it is found in the ALICE experiment that the measured baryon ratios are suppressed by a factor of $\\sim1.5$. The strange abundances are overestimated in the HRG model.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7401677370071411, "perplexity": 6751.571885668572}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257647692.51/warc/CC-MAIN-20180321195830-20180321215830-00694.warc.gz"}
https://www.tutorialspoint.com/How-to-index-and-slice-a-tuple-in-Python
# How to index and slice a tuple in Python? PythonServer Side ProgrammingProgramming To index or slice a tuple you need to use the [] operator on the tuple. When indexing a tuple, if you provide a positive integer, it fetches that index from the tuple counting from the left. In case of a negative index, it fetches that index from the tuple counting from the right. ## example my_tuple = ('a', 'b', 'c', 'd') print(my_tuple[1]) print(my_tuple[-1]) ## Output This will give the output − b d If you want to get a part of the tuple, use the slicing operator. [start:stop:step]. ## example my_tuple = ('a', 'b', 'c', 'd') print(my_tuple[1:]) #Print elements from index 1 to end print(my_tuple[:2]) #Print elements from start to index 2 print(my_tuple[1:3]) #Print elements from index 1 to index 3 print(my_tuple[::2]) #Print elements from start to end using step sizes of 2 ## Output This will give the output − ('b', 'c', 'd') ('a', 'b') ('b', 'c') ('a', 'c') Published on 05-Jan-2018 13:00:43
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.16031336784362793, "perplexity": 4115.179503049931}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320299927.25/warc/CC-MAIN-20220129032406-20220129062406-00716.warc.gz"}
https://artofproblemsolving.com/wiki/index.php/2018_AMC_10A_Problems/Problem_25
# 2018 AMC 10A Problems/Problem 25 ## Problem For a positive integer and nonzero digits , , and , let be the -digit integer each of whose digits is equal to ; let be the -digit integer each of whose digits is equal to , and let be the -digit (not -digit) integer each of whose digits is equal to . What is the greatest possible value of for which there are at least two values of such that ? ## Solution 1 Observe ; similarly and . The relation rewrites as Since , and we may cancel out a factor of to obtain This is a linear equation in . Thus, if two distinct values of satisfy it, then all values of will. Now we plug in and (or some other number), we get and . Solving the equations for and , we get To maximize , we need to maximize . Since and must be integers, must be a multiple of . If then exceeds . However, if then and for an answer of . ## Solution 2 Immediately start trying and . These give the system of equations and (which simplifies to ). These imply that , so the possible pairs are , , and . The first puts out of range but the second makes . We now know the answer is at least . We now only need to know whether might work for any larger . We will always get equations like where the coefficient is very close to being nine times the coefficient. Since the term will be quite insignificant, we know that once again must equal , and thus is our only hope to reach . Substituting and dividing through by , we will have something like . No matter what really was, is out of range (and certainly isn't as we would have needed). ## Solution 3 (Cheating) Notice that and . Setting and , we see works for all possible values of . Similarly, if and , then works for all possible values of . The second solution yields a greater sum of . ~ dolphin7
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9838111996650696, "perplexity": 365.4079651408125}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347435987.85/warc/CC-MAIN-20200603175139-20200603205139-00356.warc.gz"}
https://homework.cpm.org/category/CC/textbook/CCA2/chapter/Ch3/lesson/3.1.2/problem/3-30
### Home > CCA2 > Chapter Ch3 > Lesson 3.1.2 > Problem3-30 3-30. Describe how the graph of $y+3=−2(x+1)^2$ is different from $y=x^2$. Homework Help ✎ Which direction does each parabola open? What is the shape of each parabola? Where is the locator point of each parabola? Describe the translation.
{"extraction_info": {"found_math": true, "script_math_tex": 2, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.41433951258659363, "perplexity": 6539.232651588127}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875145818.81/warc/CC-MAIN-20200223154628-20200223184628-00550.warc.gz"}
https://mathinfocusanswerkey.com/math-in-focus-grade-4-chapter-7-practice-2-answer-key/
# Math in Focus Grade 4 Chapter 7 Practice 2 Answer Key Understanding Hundredths Practice the problems of Math in Focus Grade 4 Workbook Answer Key Chapter 7 Practice 2 Understanding Hundredths to score better marks in the exam. ## Math in Focus Grade 4 Chapter 7 Practice 2 Answer Key Understanding Hundredths Shade the squares to represent each decimal. Each large square represents 1 whole. Example Question 1. Explanation: 0.55 = 0.50 + 0.05. Question 2. Explanation: 0.55 = 0.50 + 0.05. Question 3. Explanation: 1.05 = 1.00 + 0.05. Write a decimal for each place-value chart. Question 4. Explanation: 0 ones 3 tenths 6 hundredths = 0.36. Question 5. Explanation: 5 ones 6 tenths 8 hundredths = 5.68. Question 6. Explanation: 4 ones 5 tenths 0 hundredths = 4.50. Question 7. Explanation: 2 ones 0 tenths 5 hundredths = 2.05. Write the correct decimal in each box. Question 8. Explanation: 0 + 0.03 = 0.03. 0.1 + 0.02 = 0.12. 0.12 + 0.06 = 0.18. 0.2 + 0.05 = 0.25. Mark ✗ t0 show where each decimal is located on the number line. Label its value. Question 9. 0.14 Explanation: 0.1 + 0.04 = 0.14. Question 10. 0.22 Explanation: 0.20 + 0.02 = 0.22. Question 11. 0.27 Explanation: 0.20 + 0.07 = 0.27. Write each of these as a decimal. Question 12. 9 hundredths = ___________ 9 hundredths = 0.09. Explanation: 9 hundredths = 9 ÷ 100 =  0.09. Question 13. 23 hundredths = __________ 23 hundredths = 0.23. Explanation: 23 hundredths = 23 ÷ 100 = 0.23. Question 14. 6 tenths 1 hundredth = ___________ 6 tenths 1 hundredth = 0.61. Explanation: 6 tenths 1 hundredth = 0.61. Question 15. 7 ones and 90 hundredths = __________ 7 ones and 90 hundredths = 7.090. Explanation: 7 ones and 90 hundredths = 7.090. Write each fraction as a decimal. Question 16. $$\frac{5}{100}$$ = _______ $$\frac{5}{100}$$ = 0.05. Explanation: $$\frac{5}{100}$$ = 0.05. Question 17. $$\frac{19}{100}$$ = _______ $$\frac{19}{100}$$ = 0.19. Explanation: $$\frac{19}{100}$$ = 0.19. Question 18. $$\frac{83}{100}$$ = _______ $$\frac{83}{100}$$ = 0.83. Explanation: $$\frac{83}{100}$$ = 0.83. Question 19. $$\frac{70}{100}$$ = _______ $$\frac{70}{100}$$ = 0.70. Explanation: $$\frac{70}{100}$$ = 0.70. Write each fraction or mixed number as a decimal. Question 20. 3$$\frac{17}{100}$$ = _______ 3$$\frac{17}{100}$$ = 3.17. Explanation: 3$$\frac{17}{100}$$ = (300 + 17) ÷ 100 = $$\frac{317}{100}$$ = 3.17. Question 21. 18$$\frac{9}{100}$$ = _______ 18$$\frac{9}{100}$$ = 18.09. Explanation: 18$$\frac{9}{100}$$ = (1800 + 9) ÷ 100 = $$\frac{1809}{100}$$ = 18.09 Write each fraction or mixed number as a decimal. Question 22. $$\frac{233}{100}$$ = _______ $$\frac{233}{100}$$ = 2.33. Explanation: $$\frac{233}{100}$$ = 2.33. Question 23. $$\frac{104}{100}$$ = _______ $$\frac{104}{100}$$ =1.04. Explanation: $$\frac{104}{100}$$ = 1.04. Write each decimal in hundredths. Question 24. 0.07 = __________ hundredths 0.07 = 7 hundredths. Explanation: 0.07 = 7 ÷ 100 = 7 hundredths. Question 25. 2.31 = __________ hundredths 2.31 = 231 hundredths. Explanation: 2.31 = 231 ÷ 100 = 231 hundredths. Question 26. 1.83 = __________ hundredths 1.83 = 183 hundredths. Explanation: 1.83 = 183 ÷ 100 = 183 hundredths. Question 27. 5.09 = __________ hundredths 5.09 = 509 hundredths. Explanation: 5.09 = 509 ÷ 100 = 509 hundredths. Fill in the blanks. Question 28. 0.38 = __________ tenths 8 hundredths 0.38 = 3 tenths 8 hundredths. Explanation: 0.38 = 3 tenths 8 hundredths. Question 29. 2.71 = 2 ones and 7 tenths __________ hundredth 2.71 = 2 ones and 7 tenths 1 hundredth. Explanation: 2.71 = 271 ÷ 100 = 2 ones and 7 tenths 1 hundredth. Question 30. 5.09 = 5 ones and __________ hundredths 5.09 = 5 ones and 9 hundredths. Explanation: 5.09 = 509 ÷ 100 = 5 ones and 9 hundredths. Question 31. 8.86 = 8 ones and 8 tenths ___________ hundredths 8.86 = 8 ones and 8 tenths 6 hundredths. Explanation: 8.86 = 886 ÷ 100 = 8 ones and 8 tenths 6 hundredths. You can write 6.13 in expanded form as 6 + $$\frac{1}{10}$$ + $$\frac{3}{100}$$. Complete in the same way. Question 32. Explanation: 5.24 = 5 + $$\frac{2}{10}$$ + $$\frac{4}{100}$$ = 5 + 0.20 + 0.04. Question 33. Explanation: 8.96 = 8 + $$\frac{9}{10}$$ + $$\frac{6}{100}$$ = 8 + 0.90 + 0.06. You can write 7.45 in expanded form as 7 +0.4 + 0.05. Complete in the same way. Question 34. Explanation: 4.31  = 4 + 0.30 + 0.01. Question 35. Explanation: 9.57 = 9 + 0.50 + 0.01 Fill in the blanks. Question 36. In 0.38, the digit 8 is in the ____ place. In 0.38, the digit 8 is in the hundredths place. Explanation: In 0.38, => 0 ones 3 tenths 8 hundredths. The digit 8 is in the hundredths place. Question 37. In 1 2.67, the digit in the tenths place is ___. In 1 2.67, the digit in the tenths place is 6. Explanation: In 1 2.67, => 1 tens 2 ones 6 tenths 7 hundredths. The digit in the tenths place is 6. Question 38. In 3.45, the value of the digit 5 is ____. In 3.45, the value of the digit 5 is hundredths(0.05). Explanation: In 3.45, => 3 ones 4 tenths 5 hundredths. The value of the digit 5 is hundredths(0.05). Question 39. In 5.02, the value of the digit 2 is ___________ hundredths. In 5.02, the value of the digit 2 is 0.02 hundredths. Explanation: In 5.02, => 5 ones 0 tenths 2 hundredths. The value of the digit 2 is 0.02 hundredths. Write each amount in decimal form. Question 40. 75 cents = $_______ Answer: 75 cents =$ 0.75. Explanation: 75 cents = $?? Conversion: 1 dollar = 100 cents. => 75 cents = 75 ÷100 =$0.75. Question 41. 40 cents = $_______ Answer: 40 cents =$0.40. Explanation: 40 cents = $?? Conversion: 1 dollar = 100 cents. => 40 cents = 40 ÷100 =$0.40. Question 42. 5 cents = $____ Answer: 5 cents =$0.05. Explanation: 5 cents = $?? Conversion: 1 dollar = 100 cents. => 5 cents = 5 ÷100 =$0.05. Question 43. 130 cents = $____ Answer: 130 cents =$ 1.30. Explanation: 130 cents = $?? Conversion: 1 dollar = 100 cents. => 130 cents = 130 ÷100 =$ 1.30. Question 44. 10 dollars and 25 cents = $____ Answer: 10 dollars and 25 cents =$ 10.25. Explanation: 10 dollars and 25 cents = $?? Conversion: 1 dollar = 100 cents. => 10 dollars and 25 cents =$10 +$(25 ÷100) =$10 + $0.25 =$10.25. Question 45. 28 dollars = $____ Answer: 28 dollars =$28.00. Explanation: 28 dollars = $??. Conversion: 1 dollar = 100 cents. => 28 dollars =$28.00. Question 46. 1 dollar and 9 cents = $____ Answer: 1 dollar and 9 cents =$1.09. Explanation: 1 dollar and 9 cents = $?? Conversion: 1 dollar = 100 cents. => 1 dollar and 9 cents =$1 + $(9 ÷ 100) =>$1 + $0.09 =>$1.09.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.43765223026275635, "perplexity": 9759.170764878798}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573029.81/warc/CC-MAIN-20220817153027-20220817183027-00421.warc.gz"}
http://tex.stackexchange.com/questions/17839/xelatex-inserts-blank-double-page-compared-to-pdflatex
# XeLaTeX inserts blank double page compared to pdfLaTeX I've tried recompiling some complex documents meant for pdfLaTeX using XeLaTeX. After a few fixes and `\ifxetex` in the preamble (regarding input, hyperref and font encodings), it went through... but for some reason the newly produced document was 117 pages long instead of the original 115! The reason is that a blank double page (well, blank except for running headers) is inserted between chapter 3 and 4. Obviously, this is something difficult to reproduce - it did not occur between all chapters, and it possibly depends on the height of the last page. Does this resemble some known phenomenon? - It looks as if you have something invisible after the equation which moves to the next page. If you can't find what's responsable: Try `\enlargethispage{2\baselineskip}` on this page. –  Ulrike Fischer May 10 '11 at 7:34 It was a \index command. It looks like having a \index command after a displayed equation creates something that takes up vertical space. Moving it at the end of a text paragraph solved the issue. Now why it does it with xelatex and not pdflatex remains a mystery; it may be related to hyperref. –  monniaux May 10 '11 at 8:52 Compare the documents: The pages are different (which is not surprising as different fonts are use and other things can differ too). The xelatex page is longer, probably just long enough to move the index to the next page.possible that the –  Ulrike Fischer May 10 '11 at 10:56 @Ulrike I'd re-write your comment as an answer with the brief explanation given above so this question can be closed. –  Edd May 10 '11 at 11:06 \index creates a "wotsit" node in the output list. it would seem that xetex plays slightly different with wotsits in this case (they're notorious for producing odd effects). my guess would be that your index command got detached from the rest of the text. –  wasteofspace May 10 '11 at 11:36 It turned out that an `\index` command after a displayed equation created additional vertical space, which effected the blank pages. Moving the command to the end of a text paragraph solved the issue. A possible reason for why this behaviour only occured with XeLaTeX is that different fonts were used and the XeLaTeX page was longer, probably just long enough to move the `\index` command to the next page.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9274208545684814, "perplexity": 2169.3796172975945}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-27/segments/1435375097199.58/warc/CC-MAIN-20150627031817-00256-ip-10-179-60-89.ec2.internal.warc.gz"}
http://www.jiskha.com/search/index.cgi?query=what+is+the+molarity+of+a+solution+made+by+dissolving+3.4g+of+Ba(OH)+in+enough+water+to+make+350+mL+solution%3F+Assume+that+Ba(OH)2+ionizes+completely+in+water+toBa2%2B+and+OH-+ions.+What+is+the+pH+of+the+solution%3F&page=8
Sunday November 29, 2015 # Search: what is the molarity of a solution made by dissolving 3.4g of Ba(OH) in enough water to make 350 mL solution? Assume that Ba(OH)2 ionizes completely in water toBa2+ and OH- ions. What is the pH of the solution? Number of results: 77,712 chemistry......plz help A solution is prepared by dissolving 10.00g of a monobasic salt in a total volume of 1.00litres. Th pH of this solution was measured at 13.07 and a conductivity experiment determined that this salt was completely ionized in solution. What is the molecular weight of this ... March 24, 2009 by tomi AP Chem An aqueous solution contains 4.50 g of calcium chloride, CaCl2, per liter. What is the molarity of CaCl2? When calcium chloride dissolves in water, the calcium ions, Ca^+2, and chloride ions, Cl^-, in the crystal go into the solution. What is the molar concentration of each ... November 3, 2008 by Anna chemistry If 34.7 mL of 0.210 M KOH is required to completely neutralize 30.0 mL of a CH3COOH solution, what is the molarity of the acetic acid solution? May 3, 2012 by Heny Calculate the freezing point of an aqueous solution formed by dissolbing 4.22 g of KCl in 25.00 mL of water. Kf for water is (-1.86 degree C/m). Assume density of water is 1.00 g/mL. (Hint: Use the ideal van't Hoff factor, i). July 10, 2013 by Kelly Chemistry Provide equations for each of the following: a. Dissociation of a strong base in water solution b. Ionization of a strong acid in water solution: c. Ionization of a weak acid in water solution: d. Ionization of a weak base in water solution: e. Autoionization of water: April 29, 2014 by ASAP!! chemistry I have no idea how to solve this problem, please help me out.... thanks a) A liter of a solution saturated at 25degree with calcium oxalate, CaC2O4, is evaporated to dryness, giving a 0.0061gm residue of CaC2O4. calculate the concentrations of the ions, and molar solubility ... April 6, 2007 by wendy science In the lab you need to make 1.00 L of a 0.300 M NaCl solution. You had previously made a 0.950 M NaCl solution that you can dilute to form your new solution. How many mL of the concentrated solution would you use to make the dilute solution? November 2, 2011 by liz Chemistry 12 15. The concentration of hydroxide ions in a solution of nitric acid 0.2 M is: a. 2 x 10-14 M b. 5 x 10-14 M c. 5 x 10-7 M d. 2 x 10-1 M 16. Aqueous solutions of strong acids and strong bases: a. are good electricity conductors b. make litmus paper become red c. have a pH of 7... December 3, 2012 by Need Help Chemistry What is the molality of a solution formed by dissolving 1.95 mol of KCl in 17.0 mol of water (MM of water = 18.02 g/mol), answer in 3 sigfigs? November 26, 2012 by Jake Chemistry hellppp please i have a test on this on mondayy im so screwed A solid mixture contains MgCl2 and NaCl. When 0.5000 g of this solid is dissolved in enough water to form 1.000 L of solution, the osmotic pressure at 25C is observed to be 0.4010 atm. What is the mass percent of ... November 29, 2008 by yvonne chemistry A student placed 11.5 of glucose () in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.- mark on the neck of the flask was reached. The flask was then shaken until the solution was uniform. A 25.0... January 31, 2010 by sam Chemistry A student placed 14.0 of glucose () in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.- mark on the neck of the flask was reached. The flask was then shaken until the solution was uniform. A 55.0... February 8, 2010 by Sherryl Chemistry Industrial Sulfuric Acid (H2SO4) may be purchased as 97 % (weight percent solution). Find the concentration of this solution of H2SO4 in units of mg/L. Also find the molarity and normality. Sulfuric Acid (100%) has a specific gravity of 1.839. Assume that the temperature of ... December 12, 2010 by Amanda chem (a) When a 4.75-g sample of solid ammonium nitrate dissolves in 60.0 g of water in a coffee-cup calorimeter (see figure), the temperature drops from 23.0C to 16.6C. Calculate H (in kJ/mol NH4NO3) for the solution process NH4NO3(s) NH4+(aq) + NO3−(aq). Assume ... November 6, 2012 by hannah chemistry A stock solution was prepared by dissolving exactly 0.4000 g of pure ASA (180.16 g/mol) in 10.00 mL of NaOH and heating the solution to a gentle boil. After cooling to room temperature, the solution was poured into a 250-mL volumetric flask and diluted to the mark with DI ... January 28, 2010 by Tasha chemistry A student placed 12.5 of glucose () in a 100.- volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the calibration mark on the neck of the flask was reached. The flask was then shaken until the solution was ... March 12, 2009 by tomi Chemistry A student placed 12.0g of glucose (C6H1206) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.-ml mark on the neck of the flask was reached. The flask was then shaken until the solution was ... September 29, 2010 by Steve chemistry A 60.0-g sample of NaOH is dissolved in water, and the solution is diluted to give a final volume of 1.50 L. The molarity of the final solution is ? December 10, 2009 by Anonymous chemistry If 5 g of solid sodium hydroxide is dissolved in 100 mL of water, what is the molarity of the solution in moles of NaOH/L of solution (moles/L)? September 29, 2011 by Sarah chemistry If 150.0 Ml of HCL solution are required to react completely with .35 moles of solid NaOH whats the molarity of the solution? October 12, 2009 by Aleander Azumah Chemistry What is the percent by volume concentration of a solution containing 300.mL of isopropyl alcohol and enough water to give 400.ml of solution?Which component is the solute and which is the solvent? March 16, 2014 by Hannah chemistry If 215.0 mL of water are added to a 237 mL sample of a 1.702 M dextrose solution, what is the new molarity of dextrose in the resulting solution? December 2, 2011 by B chemistry A sample containing .187 g cl is dissolved in 50.0 mL water. How many moles of cl ion are in the solution? what is the molarity of the cl ion in the solution? March 1, 2011 by kristen Chemistry Calculate the mass of HONH2 required to dissolve in enough water to make 255.5 mL of solution having a pH of 10.02 (Kb = 1.110−8). February 27, 2013 by Margaret Chemistry Calculate the mass of HONH2 required to dissolve in enough water to make 255.5 mL of solution having a pH of 10.02 (Kb = 1.110−8). February 28, 2013 by Margaret chemistry A student placed 19.0g of glucose (c6h12o6) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100ml.- mark on the neck of the flask was reached. The flask was then shaken until the solution was ... February 28, 2010 by Heather A student placed 17.0 g of glucose (C6H1206) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.-ml mark on the neck of the flask was reached. The flask was then shaken until the solution was ... June 4, 2011 by Christy chem A student placed 13.5g of glucose (\rm C_6H_{12}O_6) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100. \rm mL mark on the neck of the flask was reached. The flask was then shaken until the ... February 15, 2013 by Anonymous chemistry If you add a metal piece to excess acid, what of listed factors below will increase, decrease, or have no impact on how fast a reaction happens? a) lower molarity (concentration) of a solution b) higher molarity (concentration)of a solution c) increase volume of a solution (... May 22, 2013 by K science-biology In lab, you use a special balloon that is permeable to water but not sucrose to make an "artificial cell". The balloon is filled with a solution of 20% sucrose and 80% water and is immersed in a beaker containing a solution of 40% sucrose and 60% water. Which of the folloing ... November 19, 2010 by Tracee Chemistry Find the pH of a solution prepared by taking a 50.0 mL aliquot of a solution prepared by dissolving 12.25g of NaOH and 250.0 mL of water and diluting that aliquot to 1.00L? January 20, 2011 by Audrey Chemistry Using Volumetric Glassware to Measure, Dilute,and Titrate Acid Solution 1)For each of the following procedural errors, indicate whether the calculated molarity of the HCl solution would be higher or lower than the actual molarity. Briefly explain. a)The NaOH titrant is ... April 1, 2011 by Anonymous chemistry 8.058g x 10-2 kg of washing soda na2co3 10h20 is dissolved in water to obtain 1L of a solution of density 1077.2kg/m3. Calculate the molarity, molality and mole fraction of na2c03 in the solution? who helps January 1, 2013 by Fai Chemistry HELP! A student placed 10.5g of glucose (C6H12O6) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.- mL mark on the neck of the flask was reached. The flask was then shaken until the solution was ... February 12, 2009 by anonymous chem A student placed 12.5 of glucose (C6H12O6) in a 100ml volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the calibration mark on the neck of the flask was reached. The flask was then shaken until the solution ... March 12, 2009 by tomi A student placed 11.0 of glucose (C6H12O6) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.-ml mark on the neck of the flask was reached. The flask was then shaken until the solution was uniform... October 5, 2009 by Rose RE: Water If I boil water at home, when and how do I know I've boiled the water long enough to make distilled water? January 1, 2008 by Diana Chemistry When 5.00 g of NaOH(s) are added to 100 g of water using a calorimeter, the temperature rises from 25.0 to 37.5 C. Calculate the molar heat of solution. NaOH(s)--->Na(aq)+OH(aq) Assume that the specific heat capacity of water is 4.18 J/gK and that of the NaOH(aq) solution ... Chem II What mass of ethanol must be added to 10.0L of water to five a solution that freezes at -10.0 degrees C. Assume the density of water is q.00 g/mL. Kf of water is 1.86 degrees C/m. June 6, 2008 by Ken Chemistry A 2.7 M solution of phosphoric acid (H3PO4) is to be reacted with a 7.5 M solution of sodium hydroxide to make sodium phosphate and water. How many L of the H3PO4 solution will react with 0.662 L of the NaOH solution April 25, 2012 by John chemistry A solution is prepared by dissolving 10.00 grams of a monobasic salt in a total volume of 1.00 litres. The pH of this solution was measured at 13.07 and a conductivity experiment determined that this salt was completely ionized in solution. What is the molecular weight of this... January 29, 2014 by Temmi chemistry A total of 2.00mol of a compound is allowed to react with water in a foam coffee cup and the reaction produces 185g of solution. The reaction caused the temperature of the solution to rise from 21.0 to 24.7C . What is the enthalpy of this reaction? Assume that no heat is lost ... November 25, 2011 by Yuli BLD 414 A liquid solution of formaldehyde is 37.0% pure. This solution has a specific gravity of 1.037. The molecular weight of formaldehyde is 30.0. to 1.00 ml of this impure stock solution is added quantity sufficient to take the total volume to 500ml. If 1.00 ml of this solution is... March 19, 2012 by Amber` Chemistry Concentrated sulfuric acid is 18.4 molar and has a density of 1.84 grams per milliliter. After dilution with water to 5.20 molar the solution has a density of 1.38 grams per milliliter. This concentration of H2SO4 is often used in lead storage batteries(like a car battery). 1.... January 9, 2011 by Kevin A student placed 10.5g of glucose (C6H12O6) in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100.- mL mark on the neck of the flask was reached. The flask was then shaken until the solution was ... February 12, 2009 by anonymous chemistry Compute the vapor pressure of water above a solution formed by dissolving 25.00 grams of magnesium sulfate, MgSO4 (MM=120.3) in 95.00 grams of water at 338K. The vapor pressure of pure water at 338K is 188.0 torr. February 23, 2009 by Aimee chemistry the density at 20 degrees celcius of a .258m solution of glucose in water is 1.0173 g/ml, and the molar mass of glucose is 180.2 g/mol. What is the MOLARITY of the solution? October 20, 2008 by Eric GReenblatt chemistry 20mL ethanol(density=o.789g/mol) is dissolved in sufficient water to produce a 200mL solution. the density of the solution is 0.982g/mL. what is the molarity? March 11, 2013 by mj chemistry A 25.0-ml sample of 1.00M KBr and a 75.0-ml are mixed. The solution is then heated to evaporate water until the total volume is 50.0 ml. What is the Molarity of the KBr in the final solution February 9, 2014 by Mary Biochemistry Can someone help me calculate protein concentration using the Bradford Assay? If you could show me how to do tube 2, that'd be great. Tube 1: 0 ul Protein Stock Solution, 800 ul Water, 0 ug Protein, Absorbance @595nm = 0 Tube 2: 10 ul Protein Stock Solution, 790 ul Water, 1.4 ... October 7, 2013 by OR chemistry A solution was made by dissolving 5.15 mg of hemoglobin in water to give final volume of 1.00 mL. The osmotic pressure of this solution was 1.97 x 10^-3 atm at 25C. Calculate the molar mass of hemoglobin. pi = iMRT i = 1 since hemoglobin is nonelectrolyte and doesn't ... September 14, 2012 by Wiped Out Chemistry A solution is made by mixing 30.0 mL of ethanol, C2H6O, and 70.0 mL of water. Assuming ideal behavior, what is the vapor pressure of the solution at 20 C? I got an answer of 15.45 torr and it marked me incorrect and said "Your answer is the the partial vapor pressure ... February 18, 2012 by Lauren chem an aq. solution(NaCl) made by using 133g diluted to a total solution volume of 1 L.Calculate the molarity, molality & mass % of the solution(density of solution = 1.08g/mL) [ans: M=1.92M, m=2.0m, 10.4 %] i need the calculation work bcus my ans is contradict from the real ans. July 21, 2011 by lily holy school H2 S + HNO3 --------> NO + H2 O + S 150 g of hydrogen sulfide are dissolved in 3500 mL of water. a. What is the molarity of the hydrogen sulfide solution? b. How many grams of sulfur can be produced from 50 mL of this solution? c. 22.4 L of nitrogen monoxide gas are ... March 3, 2013 by mamta A solution is prepared by dissolving 3.44 g of sucrose, C12H22O11, in 118 g of water. a) What is the molality of the resulting solution? b) What is the mole fraction of water in this solution? B) moles of h2o=118g/18(g/mol)=6.5556 mol moles of c12h22O11= 3.44g/342.3(g/mol)=0.... January 16, 2014 by Kate Chemistry A1g sample of enriched water, a mixture of H2O AND D2O, reacted completely with Cl2 to give a mixture of HCl and DCl. The HCl and DCl were then dissolved in pure H2O to make a 1.00 L solution. A 25.00 mL, sample of the 1 L solution was reacted with excess AgNO3 and 0.3800g of ... September 21, 2013 by Anirudh Chemistry A 500.0-mL sodium iodide (NaI) solution was prepared by dissolving 37.7g of solid NaI in water. Chlorine gas was bubbled through a 100.0-mL sample of this solution until the reaction was complete. How many grams of molecular iodine were produced? November 6, 2009 by L Chem Lab Math! A liquid solution of formaldehyde is 37.0% pure. This solution has a specific gravity of 1.037. The molecular weight of formaldehyde is 30.0. to 1.00 ml of this impure stock solution is added quantity sufficient to take the total volume to 500ml. If 1.00 ml of this solution is... March 18, 2012 by Morgan M The vapor pressure of pure water at 25 C is 23.8 torr. Determine the vapor pressure (torr) of water at 25 C above a solution prepared by dissolving 35 g of urea (a nonvolatile, non-electrolyte, MW = 60.0 g/mol) in 75 g of water. im not sure about the steps for problems like ... September 15, 2010 by BEN chemistry What is the molarity of a salt solution in which 117.0 grams of salt (NaCl) is dissolved in water to give 5.00 L of solution? Show all work including units. August 23, 2011 by len chem A sample of potassium hydrogen oxalate, KHC204, weighing 0.717 g, was dissolved in water and titrated with 18.47 mL of an NaOH solution. Calculate the molarity of the NaOH solution. November 21, 2012 by amanda78 chemistry A sample of potassium hydrogen oxalate, KHC2O4, weighing 0.717 g, was dissolved in water and titrated with 18.47 mL of an NaOH solution. Calculate the molarity of the NaOH solution. April 27, 2015 by mary A/P Chem A student wants to make a 5.00% solution of potassium chloride using .0377 g of the substance.What mass of water will be needed to make the solution? July 24, 2011 by Autum chemistry determine the molarity of each of the following solutions: a. 20.0 g NaOH in enough H2O to make 2.00 L of solution is the answer .250 M yes May 16, 2007 by beckz Chemistry A 0.885 M solution of sulfuric acid (H2SO4) is to be reacted with a 0.552 M solution of sodium hydroxide to make sodium sulfate and water. How many L of the H2SO4 solution will react with 0.842 L of the NaOH solution? Give your answer to 3 decimal places. April 29, 2015 by Devin AP Chem A student placed 17.5 g of glucose in a volumetric flask, added enough water to dissolve the glucose by swirling, then carefully added additional water until the 100. mL mark on the neck of the flask was reached. The flask was then shaken until the solution was uniform. A 60.0... October 19, 2009 by Ana chem- easy What is the total mass of a solution prepared by dissolving 4.00 of in 1.00 of water? February 10, 2009 by - Chemistry what is the molality of a solution prepared by dissolving 13.0 g of Ba(NO3)2 in 450 g of water? chemistry 35.00mL of 0.250M KOH solution is required to completely neutralize 10.50mL of H2SO4 solution. what is the molarity of the sulfuric acid solution KOH(aq)+ H2SO4 -------> K2SO4(aq) +H2O(l March 15, 2012 by brina Chemistry If 43.2 mL of 2.46 M HClO4 solution completely reacts with a 75.0 mL sample of a sodium hydroxide solution, what is the molarity of the sodium hydroxide solution? (hint: write the complete balanced rxn) October 10, 2014 by Diane AP Chem What is the freezing point in degrees Celsius of a solution prepared by dissolving 11.3 (g) of calcium nitrate in 115 (g) of water? K of f for water is 1.86 degrees Celsius/m January 22, 2013 by Anonymous Science The vapor pressure of water at 25 oC is 23.8 torr. Calculate the vapor pressure of a solution prepared by dissolving 35.0 g of urea [(NH2)2CO] in 75.0 g of water at 25 oC. February 17, 2013 by Carlos chemistry step by step thank a lot 342gm of 20% by mass of ba (OH)2 solution sq gr 0.57 is reacted with 1200ml of 2M hn03. If the final density is same as pure water then molarity of the ion in resulting solution by nature of the above solution is dentified 0.888m Who solve it for me step by step thank a lot July 6, 2014 by ken Chemistry Help? Thanks At what temperature do KCL and KNO3 have the same solubility If a saturated solution of KNO3 at 20.0 C is heated to 80.0 how much more could be dissolved? If a saturated solution of KCL at 90.0 C is colled to 30.0 C how much of the solid will precipitate? How much NaCL will ... September 23, 2012 by Hannah CHEMISTRY At what temperature do KCL and KNO3 have the same solubility If a saturated solution of KNO3 at 20.0 C is heated to 80.0 how much more could be dissolved? If a saturated solution of KCL at 90.0 C is colled to 30.0 C how much of the solid will precipitate? How much NaCL will ... September 23, 2012 by Mike Chemistry Under an O2(g) pressure of 1.00 atm , 28.31mL of O2 (g) dissolves in 1.00L H2O at 298K. What will be the molarity of O2 in the saturated solution at 298K when the pressure is 3.86atm ? (Assume that the solution volume remains at 1.00L .) I used PV=nRT and found moles of solute... February 22, 2010 by Amber Chemistry Under an O2(g) pressure of 1.00 atm , 28.31mL of O2 (g) dissolves in 1.00L H2O at 298K. What will be the molarity of O2 in the saturated solution at 298K when the pressure is 3.86atm ? (Assume that the solution volume remains at 1.00L .) I used PV=nRT and found moles of ... Chemistry Chemistry - Danielle , Wednesday, February 11, 2009 at 7:24pm The density of a 22.0% by mass ethylene glycol (C2H6O2) solution in water is 1.03 g/mL. Find the molarity of the solution. Answer in M February 11, 2009 by Danielle Chemistry What is the pH and pOH of a solution that was made by adding 400mL of water to 350mL of 5.0x10^-3 M NaOH solution? May 12, 2011 by Jackie chemistry A potassium iodide solution is made by mixing 0.520 g of potassium iodide with 120.0 mL of water. Calculate the molality of potassium iodide. Assume that the density of water is 1.000 g/mL April 19, 2012 by hereyago Chemistry: Freezing point An aqueous solution is prepared by dissolving 1.08 g of human serum albumin, a protein obtained from blood plasma, in 50.0 mL of water. The solution has an osmotic pressure of 0.00770 atm at 298 K. What is the molar mass of the albumin? Answer: 6.86 x 10^4 g/mol February 25, 2013 by JJ Chemistry 1. Imagine you have a saturated solution of Ca(OH)2, at equilibrium, that has some undissolved Ca(OH)2 precipitated on the bottom of the test tube. If you add an additional scoop of Ca(OH)2 to the solution will concentration of [OH-] in the solution (increase, decrease, or be ... March 16, 2015 by Caroline Chemistry calculate the molarity of cl- ions in a solution made from mixing 50.0 ml of a 0.100 m/l solution of ferric chloride with 50.0 ml of a 0.500 mol/l solution of ammonium chloride March 29, 2015 by Amy chemistry Consider 18.0 M H2SO4, if you need to make 250.0 mL of a 3.0 M solution of H2SO4, how would you do this? 18.0 M H2SO4 has a density of 1.84g/mL what is the molality of solution? Mass % of H2SO4 in the solution and the mol fraction of H2SO4 and water in the solution? April 29, 2013 by Marie Chemistry Consider 18.0 M H2SO4, if you need to make 250.0 mL of a 3.0 M solution of H2SO4, how would you do this? 18.0 M H2SO4 has a density of 1.84g/mL what is the molality of solution? Mass % of H2SO4 in the solution and the mol fraction of H2SO4 and water in the solution? April 30, 2013 by Marie chemistry How would you prepare a solution of about 8mg/100ml of ascorbic acid in water (25ml)? I'm not sure if they mean to use that ratio of ascorbic acid in water to make a 25 ml solution or if the wording means something else. February 22, 2014 by solution chemistry calculate the molarity of cl- ions in a solution made from mixing 50.0 ml of a 0.100 m/l solution of ferric chloride with 50.0 ml of a 0.500 mol/l solution of ammonium chloride March 28, 2015 by Sim physics an antifreeze solution is made by mixing ethylene glycol (density= 1116 kg/m^3) with water. suppose that the specific gravity of such a solution is 1.0730. assuming that the total volume of the solution is the sum of its parts, determine the volume percentage of ethylene ... November 20, 2010 by kai chemistry A solution is prepared by adding 1.00 g of solid \rm NaCl to 50.0 mL of 0.100 M \rm {CaCl_2}. What is the molarity of chloride ion in the final solution? Assume that the volume of the final solution is 50.0 m October 2, 2011 by jared chemistry - vapor pressure the vapor pressure of pure water at 25C is 23.8 torr. determine the vapor pressure (torr) of water at 25 C above a solution prepared by dissolving 35g of urea (a nonvolatile, non-electrolyte, MW = 60.0G/MOL) IN 76G of water. answer:21 i just want to know the steps bc im ... September 13, 2010 by billy Chemistry 102 What is the freezing point in ∘C of a solution prepared by dissolving 6.60 g of MgCl2 in 110 g of water? The value of Kf for water is 1.86 (∘C⋅kg)/mol, and the van't Hoff factor for MgCl2 is i = 2.7. September 23, 2015 by Dee Chemistry A solution is made by diluting 18.2 g glucose (C6H12O6) in 0.150 kg water. What is the mass percentage of solute in this solution? August 3, 2010 by Jake Chemistry A solution is made by diluting 18.2 g glucose (C6H12O6) in 0.150 kg water. What is the mass percentage of solute in this solution? August 9, 2013 by Anonymous Chemistry what is the pH and pOH of a solution made by adding water to 15 grams of HCl until the volume of the solution is 2500mL? April 5, 2014 by Lena Chemistry what is the pH and pOH of a solution made by adding water to 15 grams of HCl until the volume of the solution is 2500mL? April 5, 2014 by Anonymous chemistry if A(aq) + B(aq) = AB(aq), and the molarity of A in the solution is 0.60 M and the molarity of B in solution is 0.60M, calculate the enthalpy of the reaction, delta Hrxn, for the formation of 1 mole of AB solution. November 1, 2010 by alex chemistry if A(aq) + B(aq) = AB(aq), and the molarity of A in the solution is 0.60 M and the molarity of B in solution is 0.60M, calculate the enthalpy of the reaction, delta Hrxn, for the formation of 1 mole of AB solution. November 1, 2010 by alex Chemistry A 1.00 g sample of enriched water, a mixture of H2O and D2O, reacted completely with Cl2 to give a mixture of HCl and DCl. The HCl and DCl were then dissolved in pure H2O to make a 1.00 L solution. A 25.00 mL sample of the 1.00 L solution was reacted with excess AgNO3 and 0....
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.826089084148407, "perplexity": 3740.715028020384}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398456975.30/warc/CC-MAIN-20151124205416-00101-ip-10-71-132-137.ec2.internal.warc.gz"}
https://asmedigitalcollection.asme.org/IDETC-CIE/proceedings-abstract/IDETC-CIE2013/55966/V07AT10A064/253281?redirectedFrom=PDF
This paper proposes a measurement method of cubic non-linear elasticity. The measurement system consists of a vibrator and a control circuit. We apply a nonlinear feedback to actuate the vibrator for inducing van der Pol type self-excited oscillation, so that the response amplitude of the oscillation can be arbitrarily set by changing the nonlinear feedback gain. We focus on the fact that the nonlinear elasticity causes a natural frequency shift related to the vibration amplitude of the object. We can set the response amplitude various values by changing the nonlinear feedback gain and measure the shift of the response frequency depending on the magnitude of the response amplitude. As a result, the bend of the backbone curve reflecting the nonlinear elasticity of the object is obtained. This content is only available via PDF.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8926467299461365, "perplexity": 364.6311888263915}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337853.66/warc/CC-MAIN-20221006155805-20221006185805-00115.warc.gz"}
https://socratic.org/questions/how-do-you-multiply-4-8-x-10-2-2-101-x-10-3
Algebra Topics # How do you multiply (4.8 x 10^2) * (2.101 x 10^3)? Apr 14, 2015 $\left(4.8 \times {10}^{2}\right) \cdot \left(2.101 \times {10}^{3}\right)$ $= \left(4.8 \times 2.101\right) \times \left({10}^{2} \times {10}^{3}\right)$ $= 1.00848 \times {10}^{5}$ Remember ${b}^{p} \times {b}^{q} = {b}^{p + q}$ ##### Impact of this question 236 views around the world You can reuse this answer
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 4, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9915692210197449, "perplexity": 19007.376920509316}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986710773.68/warc/CC-MAIN-20191020132840-20191020160340-00413.warc.gz"}
https://chemistry.stackexchange.com/questions/72577/calculations-for-making-a-buffer-from-a-weak-base-and-strong-acid
# Calculations for making a buffer from a weak base and strong acid How many moles of $\ce{HCl}$ must be added to $\pu{100 ml}$ of a $\pu{0.100 M}$ solution of methylamine ($\mathrm pK_\mathrm b = 3.36$) to give a buffer having a $\mathrm{pH}$ of $10.00$? (The answer is supposed to be $\pu{8.1 mmol}$.) My thought process: First, I found the $\mathrm pK_\mathrm a$ of methylammonium ion: \begin{align} \mathrm pK_\mathrm a(\ce{CH3NH3+}) &= 14 - \mathrm pK_\mathrm b(\ce{CH3NH2}) \\ &= 10.64 \end{align} Substituting this into the Henderson–Hasselbalch equation \begin{align} \mathrm{pH} &= \mathrm pK_\mathrm a + \log\left(\frac{[\ce{CH3NH2}]}{[\ce{CH3NH3+}]}\right) \\ 10.00 &= 10.64 + \log\left(\frac{\pu{10 mmol}}{x}\right) \end{align} where $x$ is the amount of $\ce{HCl}$ that must be added. However, when I solve for $x$ I find $x = \pu{43.7 mmol}$, nowhere near the correct answer. What am I missing? Since $\mathrm{p}K_\mathrm{a} + \mathrm{p}K_\mathrm{b} = 14$, you get $\mathrm{p}K_\mathrm{a} = 10.64$ for the methylammonium cation. $\ce{HCl}$ protonates methylamine. The amount of methylammonium increases by the same amount methylamine decreases. So $$\mathrm{p}K_\mathrm{a} = \mathrm{pH} + \log\left(\frac{[\ce{HA}]}{[\ce{A-}]}\right) = \mathrm{pH} + \log\left(\frac{x}{10~\pu{mmol} - x}\right)$$ lets you calculate the value $x$ of $\ce{HCl}$ needed to obtain the desired $\mathrm{pH}$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 2, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9999915361404419, "perplexity": 1513.019025221922}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439737204.32/warc/CC-MAIN-20200807143225-20200807173225-00147.warc.gz"}
http://openstudy.com/updates/51f487dbe4b0c8f6bc362af8
## A community for students. Sign up today Here's the question you clicked on: 55 members online • 0 viewing ## anonymous 3 years ago Which expression is the simplified form of the square root 338 times w to the eighteenth power.? 13w3the square root 2 times w. 2w9the square root 13 times w. 2w3the square root 13 times w. 13w9the square root of 2. Delete Cancel Submit • This Question is Closed 1. anonymous • 3 years ago Best Response You've already chosen the best response. 0 how many times does 2 go in to 18 ? 2. anonymous • 3 years ago Best Response You've already chosen the best response. 0 i.e. what is $$18\div 2$$ ? 3. anonymous • 3 years ago Best Response You've already chosen the best response. 0 this is the easy part, the hard part is finding $$\sqrt{338}$$ 4. anonymous • 3 years ago Best Response You've already chosen the best response. 0 9 5. anonymous • 3 years ago Best Response You've already chosen the best response. 0 ok so $$\sqrt{w^{18}}=w^9$$ is a start 6. anonymous • 3 years ago Best Response You've already chosen the best response. 0 as for $$\sqrt{338}$$ it turns out that $$338=13\times 13\times 2$$ or $$13^2\times 2$$ that means $\sqrt{338}=\sqrt{13^2\times 2}=\sqrt{13^2}\sqrt{2}=13\sqrt{2}$ 7. anonymous • 3 years ago Best Response You've already chosen the best response. 0 giving you a final answer of $$13w^9\sqrt{2}$$ 8. anonymous • 3 years ago Best Response You've already chosen the best response. 0 which i hope is one of your choices 9. anonymous • 3 years ago Best Response You've already chosen the best response. 0 so a 10. anonymous • 3 years ago Best Response You've already chosen the best response. 0 no 11. anonymous • 3 years ago Best Response You've already chosen the best response. 0 then what ? 12. anonymous • 3 years ago Best Response You've already chosen the best response. 0 looks like A is $$13w^3\sqrt{2}$$ which is not right 13. anonymous • 3 years ago Best Response You've already chosen the best response. 0 find the one that looks like $$13w^9\sqrt{2}$$ 14. anonymous • 3 years ago Best Response You've already chosen the best response. 0 d 15. anonymous • 3 years ago Best Response You've already chosen the best response. 0 okay pick that one 16. anonymous • 3 years ago Best Response You've already chosen the best response. 0 thanks 17. Not the answer you are looking for? Search for more explanations. • Attachments: Find more explanations on OpenStudy ##### spraguer (Moderator) 5→ View Detailed Profile 23 • Teamwork 19 Teammate • Problem Solving 19 Hero • You have blocked this person. • ✔ You're a fan Checking fan status... Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.999957799911499, "perplexity": 12738.206574487225}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988725470.56/warc/CC-MAIN-20161020183845-00344-ip-10-171-6-4.ec2.internal.warc.gz"}
https://math.stackexchange.com/questions/288886/probability-of-an-item-being-selected-in-a-multiset
# Probability of an item being selected in a multiset I would like to know that probability of an item occurring in a multiset (a combination of selections with repetitions). Given a set $S = \{x_1,x_2,...,x_n\}$ the number of possible unordered subsets of size $k$ that can be chosen with repetitions is: $${\left(n\choose k\right)} = {n + k - 1\choose k}$$ So if I want to know the probability that any given $x$ occurs in at least one of these multisets, how do I find it? I'm trying to come up with a random variable $X_i$ which represents the probability that $x_i$ occurs at least once in a multiset. If you pick one element uniformly at random from $S$, there is a probability of $\frac{n-1}n$ of not getting a fixed element $x \in S$. You do this $k$ times, the probability of not getting any $x$'s is $\left(\frac{n-1}{n}\right)^k$. Subtract this from $1$ to get the probability of getting at least one $x$. If you assume the distribution of multisets is uniform, then you would divide the number of sets having $x$ by the number of sets not having $x$. The number of multisubsets of $S$ of size $k$ not having $x$ is the same as the number of multisubsets of $S - \{x\}$ of size $k$, which is ${n + k - 2 \choose k}$. Therefore, the probability of not having $x$ in the multiset is $\frac{n + k - 2 \choose k}{n + k - 2 \choose k} = \frac{n-1}{n+k-1}$, and the answer to the question (the probability of finding $x$ at least once in the multiset) is $1-\frac{n-1}{n+k-1}=\frac k{n+k-1}$. • But the mentioned procedure is not at all one of selecting a size $k$ multiset on $S$ uniformly at random! For instance there are $\binom54=5$ size $4$ multiset of $\{A,B\}$, all but one of which contain $A$, so you sohuld get $80$% in this case. – Marc van Leeuwen Jan 28 '13 at 12:53 • i accepted the first answer b/c i think i may have been misguided in using multisets in the first place. the original problem was "there are k ppl on an elevator on the ground floor, each wants to get off at a random floor of the n upper floors, so what's the expected # of lift stops." the best approach is w/ a rand var $X_i$ for the probability that >= 1 ppl get off at floor $x$. i saw it as a multiset problem, though as the first part of the answer shows, it's more like picking elements uniformly at random. – aaronstacy Jan 29 '13 at 12:38
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8895246982574463, "perplexity": 151.33588365352293}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986672723.50/warc/CC-MAIN-20191017045957-20191017073457-00015.warc.gz"}
http://mathhelpforum.com/differential-geometry/192763-arithmetic-progression-print.html
# Arithmetic progression Let $E \in \mathbb{R}$ be a Lebesgue measurable set with strictly positive measure. Show that $E$ contains an arithmetic progression of length three.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 2, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9403766989707947, "perplexity": 107.42036691117467}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500835822.36/warc/CC-MAIN-20140820021355-00100-ip-10-180-136-8.ec2.internal.warc.gz"}
https://cimmul.fsg.ulaval.ca/event/seminaire-du-cimmul-marina-vegue-llorente/
• Cet évènement est passé # Séminaire du CIMMUL – Marina Vegué Llorente ## Firing rate and synaptic weight distributions in plastic networks of spiking neurons ### Marina Vegué Llorente Stagiaire postdoctorale, Dynamica Département de physique, génie physique et d’optique, Université Laval Résumé Networks of spiking neurons have been widely used as models to represent neuronal activity in the brain. These models are reasonably realistic but they are also difficult to treat analytically. Mean-field theory has nevertheless proven to be successful as a method for deriving some of their statistical properties at equilibrium, such as the distribution of firing rates, both in homogeneous networks and in networks which exhibit a large heterogeneity in their structure. However, these models lack realism in the sense that they assume a fixed connectivity, whereas the connection strengths in brain networks evolve in time according to plasticity rules that depend on the neuronal activity. In this talk I will present a way to extend the mean-field formalism to networks with synaptic weights that are prone to plastic, activity-dependent modulation. The plasticity in our model is mediated by the introduction of spike traces. A trace associated to one neuron represents a chemical signal that is released every time the neuron emits a spike and which is degraded over time. Mathematically, it is a stochastic variable that can be rescaled to be an approximation to the neuron’s firing rate. The temporal evolution of the trace is controlled by its degradation speed (i.e., a measure of how fast the spiking “memory” is lost) and by the mean temporal separation between consecutive spikes (i.e., the inverse of the neuron’s firing rate). These time scales jointly determine a shift from a regime characterized by highly noisy traces to a regime of accurate traces. Our mean-field formalism provides a set of equations whose solution specifies the firing rate and synaptic weight distributions at equilibrium. The equations are exact in the limit of traces with infinite memory but I will show that they already provide accurate results when the degradation speed is of the order of a few seconds. Overall, this work offers a new perspective to explore and better understand the way in which plasticity shapes both activity and structure in neural networks. ## Détails Date: mars 25 Heure : 13 h 00 min - 14 h 00 min En ligne
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8769499063491821, "perplexity": 1059.7659482967356}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652663035797.93/warc/CC-MAIN-20220529011010-20220529041010-00072.warc.gz"}
http://mathhelpforum.com/advanced-math-topics/223432-integrals-involving-polar-coordinate-conversion.html
Integrals involving Polar Coordinate Conversion The following is part of a proof that I am studying from the book Partial Differential Equations by Lawrence Evans. Given that $u \in C^{1}(\mathbb{R}^{n}) with t \in (0,s)$ and $w \in \partial B(0,1)$. Consider the integral $\int_{0}^{s}\int_{\partial B(0,1)} |Du(x+tw)|dSdt = \int_{0}^{s}\int_{\partial B(0,1)}|Du(x+tw)|\frac{t^{n-1}}{t^{n-1}}dSdt$. How does it follow then that if we let $y=x+tw$ so that $t = |x-y|$. Then converting from polar coordinates, we have the inequality: $\int_{\partial B(0,1)}|u(x+sw) - u(x)|dS \leq \int_{B(x,s)}\frac{|Du(y)|}{|x-y|^{n-1}}dy$ How does this final inequality follow? How is it a result of polar coordinate conversion? Let me know if anything is unclear. Thanks.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 6, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9990575909614563, "perplexity": 135.64991748701368}, "config": {"markdown_headings": false, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-26/segments/1466783397748.48/warc/CC-MAIN-20160624154957-00127-ip-10-164-35-72.ec2.internal.warc.gz"}
http://math.stackexchange.com/users/8134/jaime?tab=activity&sort=all&page=2
Jaime Reputation 381 Top tag Next privilege 500 Rep. Access review queues Dec7 answered Derivatives of Functions Dec6 answered Exponential Distribution Probability question Dec5 answered Probabilities for unknown finite population from sample? Dec5 answered Finding a basis for the plane with the equation Oct2 comment How to complete the argument to find the solution of the following non-linear O. D. E.? Methinks you are missing a $'$ in the RHS of your original equation (*) Sep30 awarded Teacher Sep29 answered Orthogonality of Bessel functions Mar11 awarded Autobiographer Mar11 awarded Supporter
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.890325129032135, "perplexity": 5335.348577577183}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207925201.39/warc/CC-MAIN-20150521113205-00267-ip-10-180-206-219.ec2.internal.warc.gz"}
https://www.gamedev.net/topic/645414-legal-issues-for-software-tools/?setlanguage=1&langurlbits=topic/645414-legal-issues-for-software-tools/&langid=1
$15 ### Image of the Day Submit IOTD | Top Screenshots ### The latest, straight to your Inbox. Subscribe to GameDev.net's newsletters to receive the latest updates and exclusive content. Sign up now ## Legal issues for software tools Old topic! Guest, the last post of this topic is over 60 days old and at this point you may not reply in this topic. If you wish to continue this conversation start a new topic. 8 replies to this topic ### #1Dark_Oppressor Members Posted 14 July 2013 - 12:37 AM This isn't about a game, but a tool related to games, so I hope that's OK. I've recently had an idea about a software tool that would provide a new way to look at one's Steam profile. I'm wondering if there are legal issues to creating such a tool. Is this sort of thing common? Is it totally out of the question, perfectly legitimate, or questionable? I really have no idea, and I've never even considered selling software that isn't a game (my company is called Cheese and Bacon Games!) but if it is an OK thing, it would not take long to create it and try it out (famous last words...) ### #2bschmidt1962 Members Posted 14 July 2013 - 07:19 PM It's hard without knowing the details, but in general you are free to reverse engineer things and do what you like with the data. One interesting case is Sega vs Accolada (http://www.law.cornell.edu/copyright/commentary/chn95t1.htm). Accolade reverse engineered the Sega Genesis security code so they could make Genesis games without Sega's imprimatur. Sega sued and eventually lost on appeal. Another interesting case was the Palm Pre. The engineers at Plam reverse engineered the format and data transfer protocol for the iPod, so that when you plugged in your Palm Pre, your computer would think you were plugging in an iPod. The US copyright code (http://www.copyright.gov/title17/92chap12.html), specifically allows that type of reverse engineering. One thing to note-- although they didn't sue, Apple was able to very easily thwart Palm by simply changing their protocol from time to time, making Palm's reverse-engineering not work.... (Note, that there was a lot of bad blood between Apple and Palm, particularly since many of Palm's engineers were former Apple engineers...) Brian Schmidt Executive Director, GameSoundCon: GameSoundCon 2016:September 27-28, Los Angeles, CA Founder, Brian Schmidt Studios, LLC Music Composition & Sound Design Audio Technology Consultant ### #3Dark_Oppressor Members Posted 14 July 2013 - 08:59 PM It's hard without knowing the details, but in general you are free to reverse engineer things and do what you like with the data. That's great news, thank you! The Apple/Palm thing sounds like a funny story, heh. What about naming? For instance, could I call it something like "Steam Sorting Thingy"? That seems like it would be problematic, but if so, how would anyone ever know what it was? ### #4bschmidt1962 Members Posted 14 July 2013 - 09:07 PM What about naming? For instance, could I call it something like "Steam Sorting Thingy" Steam is trademarked, so you can't use it without explicit permission. Calling it "Steam Sorting Thingy" would probably be a no-no. The issue is whether or not someone might be confused to believe that your thingy was created or endorsed or licensed by Valve. There are a few cases where you can use a trademarked name-- such as a Drug-store version of a brand name drug that can say on the label "Compare to the active ingredent in Claritin™", but not in naming your own product. Brian Schmidt Executive Director, GameSoundCon: GameSoundCon 2016:September 27-28, Los Angeles, CA Founder, Brian Schmidt Studios, LLC Music Composition & Sound Design Audio Technology Consultant ### #5Servant of the Lord Members Posted 14 July 2013 - 10:25 PM Seeing that Valve post the Steam Web API online, I don't think they'd be antagonistic towards you providing "a new way to look at one's Steam profile." See here for the community-sourced information (also hosted on Valve-controlled web servers). You have seen things like Steam Gauge, right? Apparently I have$2278.59 worth of games, that take a total of 468.14 GB, and that have been played for 2605.9 hours. I've never used Steam Gauge specifically before, but there are dozens of sites like that. Never give them your passwords, obviously, but most just ask for your Steam account name, which is already public. It's perfectly fine to abbreviate my username to 'Servant' or 'SotL' rather than copy+pasting it all the time. All glory be to the Man at the right hand... On David's throne the King will reign, and the Government will rest upon His shoulders. All the earth will see the salvation of God. Of Stranger Flames - ### #6Dark_Oppressor  Members Posted 14 July 2013 - 11:03 PM Oh wow, I did not know that API existed, thanks! Heh, I've just been gathering all the info I need from their site. I.e., you give it a Steam username, and it builds a list of your games from the profile page's HTML, then it looks at each game's store page for more info on it. It actually works really well but this might simplify some things :-) So the only real issue is with the name. It looks like Steam Gauge just has a disclaimer on the page saying that Steam is Valve's thing. Is that enough, or should I be safer/more paranoid and not have Steam anywhere in my program's name? Bear in mind I am thinking of selling it (on mobile, maybe), in case that is relevant. Also, the number of hours you've played Awesomenauts makes me sad, Servant of the Lord, it's a great game! :-( ### #7Servant of the Lord  Members Posted 15 July 2013 - 10:26 AM Valve is very community-friendly. Legally, using the name Steam would be a no-go, but Valve doesn't go lawsuit crazy - they have a reputation to protect with their fanbase. When a bunch of modders were making a remake of Halflife 1 with the Source engine, Valve encouraged it. When they called it "Black Mesa Source", Valve politely sent them an email asking them not to use the name "Source", so Valve can use it to mark official Valve products (Day of Defeat: Source, Halflife: Source, CounterStrike: Source, etc...), and suggested just naming it "Black Mesa"... but let them continue to use their "blackmesasource.com" domain name so they wouldn't have to spend \$10 on another domain. I think the disclaimer would be fine - but shoot Valve's community management team an email, explain your project, and ask them if "Bob's Steam Profile Tool" is an acceptable name to use. Uh, I'm not sure about you commercially selling the product, but you can email them about that too. Maybe make it donation-based, and ask Valve if they're cool with that. Re: Awesomenauts I got it in a bundle (one of the Humble Indie Bundles, iirc) only a month ago, so I've never played it. It's perfectly fine to abbreviate my username to 'Servant' or 'SotL' rather than copy+pasting it all the time. All glory be to the Man at the right hand... On David's throne the King will reign, and the Government will rest upon His shoulders. All the earth will see the salvation of God. Of Stranger Flames - ### #8Stormynature  GDNet+ Posted 15 July 2013 - 11:43 AM Uh, I'm not sure about you commercially selling the product, but you can email them about that too. Don't forget Steam Greenlight also does programs/tools as well games as one possible avenue. ### #9Dark_Oppressor  Members Posted 15 July 2013 - 05:08 PM Thanks a ton for all this advice, guys. It all actually sounds less scary than I expected. I'll definitely try to get in touch with Valve and just ask how they'd feel about it. I think I might make it even if they don't want it sold, because I think it could be useful to people. But I'd sure not mind making some money off of it, haha! Donations is a cool idea, that might be a good way to go. I really only had plans for a mobile version, but my prototype is a little C++ Windows program so I could probably do a desktop version if I wanted. Old topic! Guest, the last post of this topic is over 60 days old and at this point you may not reply in this topic. If you wish to continue this conversation start a new topic.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1951974481344223, "perplexity": 2783.9855426506233}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501171163.39/warc/CC-MAIN-20170219104611-00576-ip-10-171-10-108.ec2.internal.warc.gz"}
http://ejde.math.txstate.edu/Volumes/2005/05/abstr.html
Electron. J. Diff. Eqns., Vol. 2005(2005), No. 05, pp. 1-20. Multiplicity of symmetric solutions for a nonlinear eigenvalue problem in Daniela Visetti Abstract: In this paper, we study the nonlinear eigenvalue field equation where is a function from to with , is a positive parameter and . We find a multiplicity of solutions, symmetric with respect to an action of the orthogonal group : For any we prove the existence of finitely many pairs solutions for sufficiently small, where is symmetric and has topological charge . The multiplicity of our solutions can be as large as desired, provided that the singular point of and are chosen accordingly. Submitted October 22, 2004. Published January 2, 2005. Math Subject Classifications: 35Q55, 45C05. Key Words: Nonlinear Schrodinger equations; nonlinear eigenvalue problems. Show me the PDF file (334K), TEX file, and other files for this article. Daniela Visetti Dipartimento di Matematica Applicata "U. Dini", Università degli studi di Pisa via Bonanno Pisano 25/B, 56126 Pisa, Italy email: visetti@mail.dm.unipi.it
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9500190019607544, "perplexity": 1227.6460792524765}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500816424.18/warc/CC-MAIN-20140820021336-00186-ip-10-180-136-8.ec2.internal.warc.gz"}
https://openstax.org/books/university-physics-volume-2/pages/12-2-magnetic-field-due-to-a-thin-straight-wire
University Physics Volume 2 # 12.2Magnetic Field Due to a Thin Straight Wire University Physics Volume 212.2 Magnetic Field Due to a Thin Straight Wire ### Learning Objectives By the end of this section, you will be able to: • Explain how the Biot-Savart law is used to determine the magnetic field due to a thin, straight wire. • Determine the dependence of the magnetic field from a thin, straight wire based on the distance from it and the current flowing in the wire. • Sketch the magnetic field created from a thin, straight wire by using the second right-hand rule. How much current is needed to produce a significant magnetic field, perhaps as strong as Earth’s field? Surveyors will tell you that overhead electric power lines create magnetic fields that interfere with their compass readings. Indeed, when Oersted discovered in 1820 that a current in a wire affected a compass needle, he was not dealing with extremely large currents. How does the shape of wires carrying current affect the shape of the magnetic field created? We noted in Chapter 28 that a current loop created a magnetic field similar to that of a bar magnet, but what about a straight wire? We can use the Biot-Savart law to answer all of these questions, including determining the magnetic field of a long straight wire. Figure 12.5 shows a section of an infinitely long, straight wire that carries a current I. What is the magnetic field at a point P, located a distance R from the wire? Figure 12.5 A section of a thin, straight current-carrying wire. The independent variable $θ θ$ has the limits $θ 1 θ 1$ and $θ 2 . θ 2 .$ Let’s begin by considering the magnetic field due to the current element $Idx→Idx→$ located at the position x. Using the right-hand rule 1 from the previous chapter, $dx→×r^dx→×r^$ points out of the page for any element along the wire. At point P, therefore, the magnetic fields due to all current elements have the same direction. This means that we can calculate the net field there by evaluating the scalar sum of the contributions of the elements. With $|dx→×r^|=(dx)(1)sinθ,|dx→×r^|=(dx)(1)sinθ,$ we have from the Biot-Savart law $B=μ04π∫wireIsinθdxr2.B=μ04π∫wireIsinθdxr2.$ 12.5 The wire is symmetrical about point O, so we can set the limits of the integration from zero to infinity and double the answer, rather than integrate from negative infinity to positive infinity. Based on the picture and geometry, we can write expressions for r and $sinθsinθ$ in terms of x and R, namely: $r=x2+R2sinθ=Rx2+R2.r=x2+R2sinθ=Rx2+R2.$ Substituting these expressions into Equation 12.5, the magnetic field integration becomes $B=μoI2π∫0∞Rdx(x2+R2)3/2.B=μoI2π∫0∞Rdx(x2+R2)3/2.$ 12.6 Evaluating the integral yields $B=μoI2πR[ x(x2+R2)1/2]0∞.B=μoI2πR[ x(x2+R2)1/2]0∞.$ 12.7 Substituting the limits gives us the solution $B=μoI2πR.B=μoI2πR.$ 12.8 The magnetic field lines of the infinite wire are circular and centered at the wire (Figure 12.6), and they are identical in every plane perpendicular to the wire. Since the field decreases with distance from the wire, the spacing of the field lines must increase correspondingly with distance. The direction of this magnetic field may be found with a second form of the right-hand rule (illustrated in Figure 12.6). If you hold the wire with your right hand so that your thumb points along the current, then your fingers wrap around the wire in the same sense as $B→.B→.$ Figure 12.6 Some magnetic field lines of an infinite wire. The direction of $B → B →$ can be found with a form of the right-hand rule. The direction of the field lines can be observed experimentally by placing several small compass needles on a circle near the wire, as illustrated in Figure 12.7. When there is no current in the wire, the needles align with Earth’s magnetic field. However, when a large current is sent through the wire, the compass needles all point tangent to the circle. Iron filings sprinkled on a horizontal surface also delineate the field lines, as shown in Figure 12.7. Figure 12.7 The shape of the magnetic field lines of a long wire can be seen using (a) small compass needles and (b) iron filings. ### Example 12.3 #### Calculating Magnetic Field Due to Three Wires Three wires sit at the corners of a square, all carrying currents of 2 amps into the page as shown in Figure 12.8. Calculate the magnitude of the magnetic field at the other corner of the square, point P, if the length of each side of the square is 1 cm. Figure 12.8 Three wires have current flowing into the page. The magnetic field is determined at the fourth corner of the square. #### Strategy The magnetic field due to each wire at the desired point is calculated. The diagonal distance is calculated using the Pythagorean theorem. Next, the direction of each magnetic field’s contribution is determined by drawing a circle centered at the point of the wire and out toward the desired point. The direction of the magnetic field contribution from that wire is tangential to the curve. Lastly, working with these vectors, the resultant is calculated. #### Solution Wires 1 and 3 both have the same magnitude of magnetic field contribution at point P: $B1=B3=μoI2πR=(4π×10−7T⋅m/A)(2A)2π(0.01m)=4×10−5T.B1=B3=μoI2πR=(4π×10−7T⋅m/A)(2A)2π(0.01m)=4×10−5T.$ Wire 2 has a longer distance and a magnetic field contribution at point P of: $B2=μoI2πR=(4π×10−7T⋅m/A)(2A)2π(0.01414m)=3×10−5T.B2=μoI2πR=(4π×10−7T⋅m/A)(2A)2π(0.01414m)=3×10−5T.$ The vectors for each of these magnetic field contributions are shown. The magnetic field in the x-direction has contributions from wire 3 and the x-component of wire 2: $Bnetx=−4×10−5T−2.83×10−5Tcos(45°)=−6×10−5T.Bnetx=−4×10−5T−2.83×10−5Tcos(45°)=−6×10−5T.$ The y-component is similarly the contributions from wire 1 and the y-component of wire 2: $Bnety=−4×10−5T−2.83×10−5Tsin(45°)=−6×10−5T.Bnety=−4×10−5T−2.83×10−5Tsin(45°)=−6×10−5T.$ Therefore, the net magnetic field is the resultant of these two components: $Bnet=Bnetx2+Bnety2Bnet=(−6×10−5T)2+(−6×10−5T)2Bnet=8×10−5T.Bnet=Bnetx2+Bnety2Bnet=(−6×10−5T)2+(−6×10−5T)2Bnet=8×10−5T.$ #### Significance The geometry in this problem results in the magnetic field contributions in the x- and y-directions having the same magnitude. This is not necessarily the case if the currents were different values or if the wires were located in different positions. Regardless of the numerical results, working on the components of the vectors will yield the resulting magnetic field at the point in need. Check Your Understanding Using Example 12.3, keeping the currents the same in wires 1 and 3, what should the current be in wire 2 to counteract the magnetic fields from wires 1 and 3 so that there is no net magnetic field at point P? Order a print copy As an Amazon Associate we earn from qualifying purchases.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 19, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9152059555053711, "perplexity": 256.91838692198235}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320305242.48/warc/CC-MAIN-20220127072916-20220127102916-00706.warc.gz"}
https://arxiv.org/abs/1112.5761
cs.PL # Title:Semantics and Algorithms for Parametric Monitoring Authors:Grigore Rosu (University of Illinois at Urbana-Champaign), Feng Chen (University of Illinois at Urbana-Champaign) Abstract: Analysis of execution traces plays a fundamental role in many program analysis approaches, such as runtime verification, testing, monitoring, and specification mining. Execution traces are frequently parametric, i.e., they contain events with parameter bindings. Each parametric trace usually consists of many meaningful trace slices merged together, each slice corresponding to one parameter binding. This gives a semantics-based solution to parametric trace analysis. A general-purpose parametric trace slicing technique is introduced, which takes each event in the parametric trace and dispatches it to its corresponding trace slices. This parametric trace slicing technique can be used in combination with any conventional, non-parametric trace analysis technique, by applying the later on each trace slice. As an instance, a parametric property monitoring technique is then presented. The presented parametric trace slicing and monitoring techniques have been implemented and extensively evaluated. Measurements of runtime overhead confirm that the generality of the discussed techniques does not come at a performance expense when compared with existing parametric trace monitoring systems. Comments: This paper will appear in LMCS. It is an extended version of a paper presented in TACAS'09 Subjects: Programming Languages (cs.PL); Logic in Computer Science (cs.LO); Software Engineering (cs.SE) ACM classes: D.1.5, D.2.1, D.2.4, D.2.5, D.3.1, F.3.1, F.3.2 Journal reference: Logical Methods in Computer Science, Volume 8, Issue 1 (February 23, 2012) lmcs:710 DOI: 10.2168/LMCS-8(1:9)2012 Cite as: arXiv:1112.5761 [cs.PL] (or arXiv:1112.5761v2 [cs.PL] for this version) ## Submission history From: Grigore Rosu [view email] [via LMCS proxy] [v1] Sun, 25 Dec 2011 00:02:26 UTC (75 KB) [v2] Tue, 21 Feb 2012 21:58:42 UTC (77 KB)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8232088685035706, "perplexity": 5597.146609835011}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540496492.8/warc/CC-MAIN-20191207055244-20191207083244-00104.warc.gz"}
https://migerusblog.wordpress.com/
WordPress supports LaTeX. Who knew? $e^{i\pi}+1=0$ They’ve created a monster… Esta tarde a las 18h en la plaza de Callao hay una concentración conmemorativa del 60 aniversario del Tratado de Roma. Según la Comisión Europea, las asociaciones convocantes son (en orden alfabético): Estudiantes por Europa, Eurocitizens, Europeístas, Europa en Suma, Jóvenes Federalistas Europeos, Valor Europa. ¿Quiénes son estas asociaciones? Cross-posted on European Tribune. In his satire Candide, published in 1759, Voltaire pokes fun at the way the Portuguese Inquisition persecuted jews who had falsely converted to Catholicism: After the earthquake had destroyed three-fourths of Lisbon, the sages of that country could think of no means more effectual to prevent utter ruin than to give the people a beautiful auto-da-fe; for it had been decided by the University of Coimbra, that the burning of a few people alive by a slow fire, and with great ceremony, is an infallible secret to hinder the earth from quaking. In consequence hereof, they had seized on a Biscayner, convicted of having married his godmother, and on two Portuguese, for rejecting the bacon which larded a chicken they were eating[7]; after dinner, they came and secured Dr. Pangloss, and his disciple Candide, the one for speaking his mind, the other for having listened with an air of approbation. Fast-forward to 2016, and Sarkozy’s extremism is indistinguishable from Voltaire’s satire. Se ha vertido mucha tinta sobre los pactos en la constitución de la mesa del primer parlamento cuasi-post-bipartidista desde la Transición. Por primera vez se constituyen las Cortes españolas sin que esté nada claro el color del futuro gobierno, ni siquiera si habrá futuro gobierno sin tener que repetir las elecciones. Por tanto los posibles pactos para constituir la mesa del Congreso no están necesariamente ligados a un pacto de gobierno. Cross-posted from European Tribune. Prominent heterodox economist James Galbraith, who enjoyed an inside view of the last five months of Greek negotiations as an advisor to Yanis Varoufakis, writes the following for a mainstream American audience: Greece, Europe, and the United States (Harper’s, July 16, 2015) What will become of Europe? Clearly the hopes of the pro-European, reformist left are now over. That will leave the future in the hands of the anti-European parties, including UKIP, the National Front in France, and Golden Dawn in Greece. These are ugly, racist, xenophobic groups; Golden Dawn has proposed concentration camps for immigrants in its platform. The only counter, now, is for progressive and democratic forces to regroup behind the banner of national democratic restoration. Which means that the left in Europe will also now swing against the euro. The parallel between the Greek crisis and the Prague Spring, with a ruthless mainstream left crushing the hopes of an idealist left in defence of a system, is illustrated with poetic irony by the following tweet by a Social-Democrat finance minister from the former Czechoslovakia: Meanwhile, in an interview with Jacobin Magazine which we have already been discussing in previous threads on European Tribune, Left Platform Syriza MP Stathis Kouvelakis says the following about the ideology of “left-Europeanism”: Greece: The Struggle Continues (Jacobin, July 14, 2015) I think that in this case we can clearly see what the ideology at work here is. Although you don’t positively sign up to the project and you have serious doubts about the neoliberal orientation and top-down structure of European institutions, nevertheless you move within its coordinates and can’t imagine anything better outside of its framework. I imagine that you could have written the same of Communist parties in the 1960s and their support for the Soviet Union. Out of the disappointment of the Prague Spring (on top of the invasion of Hungary a decade earlier) was born the Eurocommunist strand of the 1970s. Ptochotrapezocracy, or rule by bankrupt banks, is a term used (and, as far as I can tell, coined) by Yanis Varoufakis to describe the system of capitalism post-crisis [1]. While the European council applies itself to the task of apportioning blame for the clusterfuck that is the Greek debt crisis (though now it seems the Council President Donald Tusk wants nothing to do with the mess) it seems to me that policy in and around Greece will in the next days and weeks be determined by the state of the Greek banks, with all concern for “Europe” thrown to the wind in the face of more pressing demands. A summary of the argument below the fold: • Lifting trade credit and import restrictions should be a priority for the Greek government, more than lifting restrictions on cash withdrawals • The greek banks are still solvent, but the ECB won’t allow ELA to be relaxed for the purpose of financing imports unless the ongoing bank run is stopped • The bank run can only be stopped by freezing deposits until the economy is stabilised, as done in Cyprus two years ago • If the ECB pulls ELA entirely, in addition to a deposit freeze there will have to be a deposit haircut, of maybe 35% of the €65bn of domestic time deposits (€60bn of domestic overnight deposits would be preserved, but frozen) • In order to avoid a bail-in of deposits even in the event of a recapitalization, the ECB’s supervisory arm SSM should have restructured the banks as early as February 4, when in fact the ECB shifted half of its liquidity provision to ELA • This is because €31bn of liquidity flowed out of the Greek banks in the month of January, the worst month of the 9-month bank run At today’s press conference (see my full transcript below the fold) between the Eurogroup and his one-on-one meeting with Mario Draghi, Yanis Varoufakis made a very significant observation on bank recapitalization: The second reason concerned the financing. The proposal for financing the 5-month extension period of the current agreement – the funding proposal that was on the table – by the Institutions was technically inadequate: the numbers simply didn’t add up. And when we asked questions about, in particular, the utilisation of parts of the HFSF 10.9 billion – which was there in order to recapitalise banks and strengthen them in view of the problem with the NPLs – we received an answer that involved the ESM and that, as we know, effectively baked into this short-term arrangement a new program. Both technically and legally that did not make much sense and I’m sure that other member states would have a problem with it, Germany in particular. What Yanis Varoufakis is referring to is the following. In the framework of the European Bank Recovery and Resolution Directive, the European Stability Mechanism has established a direct bank recapitalisation instrument. From the FAQ [PDF]: Until the creation of the direct recapitalisation instrument, the ESM could only recapitalise financial institutions indirectly. In this case, the ESM provides a loan to the government of a euro area Member State. With these funds the government then recapitalises the financial institutions, which is how the ESM provided assistance to Spain. However, such assistance adds to the beneficiary country’s public debt, which could have a negative impact on market sentiment. … Until the creation of the direct recapitalisation instrument, the ESM could only recapitalise financial institutions indirectly. In this case, the ESM provides a loan to the government of a euro area Member State. With these funds the government then recapitalises the financial institutions, which is how the ESM provided assistance to Spain. However, such assistance adds to the beneficiary country’s public debt, which could have a negative impact on market sentiment. So far, so good. But crucially (my emphasis) There will be conditions applying to the recapitalised institution, established under EU state aid rules. In addition, the ESM, in liaison with the Commission and the ECB, can add additional institution-specific conditions. These can include rules on the governance of the institution, remuneration of management and bonuses. Other policy conditions may be related to the general economic policies of the ESM Member concerned. They will be included in the Memorandum of Understanding (MoU) attached to the financial assistance. This is something that the Greek government will apparently not stand for. But this has more immediate implications. It’s looking likely that the ECB Governing Council may withdraw emergency liquidity assistance (ELA) from the Greek banks no later than Wednesday (the current “second programme” expires at the end of Tuesday), if not earlier (though I think it could also freeze ELA now and withdraw it on Wednesday). At that point, the Single Supervisory Mechanism could declare the four major Greek banks insolvent and set into motion the procedures foreseen by the Bank Recovery and Resolution Directive. Given the insolvency of the Greek state, “recovery” would involve ESM recapitalisation as outlined in the FAQ above. But, as Varoufakis hints, the creditors would attempt to impose policy conditionality on the Greek state, at which point the Greek government would refuse recapitalisation. This would leave resolution (i.e., liquidation) as the only solution for the Greek banks. And it is quite likely that resolution would involve a bail-in of Greek depositors as Greek banks have very little by way of long-term funding other than time deposits. If the ECB withdraws emergency liquidity from the Greek banks on Wednesday, a decision to wind down the Greek banks could be taken as early as Friday, two days before the referendum.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 1, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3302246630191803, "perplexity": 6780.092633139044}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676589980.6/warc/CC-MAIN-20180718002426-20180718022426-00425.warc.gz"}
https://astarmathsandphysics.com/index.php?option=com_content&view=article&id=5090:the-cauchy-product&catid=148&Itemid=1787
## The Cauchy Product Take two power series $\sum_i a_ix^i, \; \sum_j b_jx^j$ which converge on intervals $I_1, \; I_2$ respectively. The Cauchy product of the two power series is \begin{aligned} (\sum_i a_ix^i)(\sum_j b_jx^j) &= \sum_i \sum_j a_ib_jx^{i+j} \\ &= \sum_{i,j, \; i+j=k} c_k x^k \end{aligned} . Then $c_0=a_0b_0$ $c_1=a_0b_1+a_1b_0$ $c_2=a_0b_2+a_1b_1+a_2b_0$ $c_3=a_0b_3+a_1b_2+a_2b_1+a_3b_0$ The product of the two series will only converge if both series converge, i.e. on the intersection of $I_1, \; I_2$ .
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 1, "x-ck12": 0, "texerror": 0, "math_score": 1.0000100135803223, "perplexity": 3479.453900355205}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141176049.8/warc/CC-MAIN-20201124082900-20201124112900-00606.warc.gz"}
http://mathoverflow.net/questions/67359/the-number-of-pairings-between-multisets?sort=oldest
# The number of pairings between multisets Given two multisets $A$ and $B$ of the same finite cardinality $n$, how many ways are there of pairing the two sets together? If both sets consist of distinct elements, the answer is $n!$: there are $n$ ways to pair the first element of $A$ with something from $B$, $n-1$ for the second element, etc. If one of the sets has distinct elements and the other is allowed to have repeated elements, again the answer is well-understood. If $A$ has distinct elements and the elements of $B$ have multiplicities $b_1,\dots,b_s$ with $b_1+\dots+b_s=n$, then the number of pairings is $n!/b_1!\dots b_s!$. What's not obvious to me is what happens when both sets are allowed to have repeated elements. As a simple example, suppose $A=\{1,2,3\}$ and $B=\{a,a,b\}$. Either per the above formula or by simple counting, one sees that there are 3 pairings - $[1a,2a,3b],[1a,2b,3a]$, and $[1b,2a,3a]$. However, if $A=\{1,1,2\}$ and $B=\{a,a,b\}$, then there are only 2 pairings - $[1a,1a,2b]$ and $[1a,1b,2a]$. This example is noteworthy in that it shows that the number of pairings doesn't have to divide $n!/b_1!\dots b_s!$. In particular, if $a_1,\dots,a_r$ are the multiplicities of the elements of $A$, the number of pairings is not $n!/a_1!\dots a_r! b_1! \dots b_s!$, a quantity which does not even have to be an integer. For my purposes, I'd like to have a way to write this in terms of fairly simple combinatorial objects (multinomial coefficients, Bell or Stirling numbers, etc.), but I'm not convinced this is possible, at least without resorting to a heinous sum. In fact, I only care about the parity of this count, so even a characterization of the $a_i$ and $b_i$ which make this even or odd would be of use to me. The only restriction I have on $A$ and $B$ is that at least one $b_i$, say, must be 1, but I'm not sure how to take advantage of that here. - What do you mean by "pairing"? –  Qiaochu Yuan Jun 9 '11 at 17:26 Sorry, that is perhaps unclear. I included the examples to try to clear things up, but I guess I was unsuccessful. I mean a pairing to be a way to associate to each element of A a unique element of B. Perhaps one-to-one correspondence would be better? –  rlo Jun 9 '11 at 17:56 If the multiplicities of the elements of the first multiset are $a_1,a_2,\dots$ and of the second $b_1,b_2,\dots$, then you are asking for the number of matrices $A=(A_{ij})_{i,j\geq 1}$ of nonnegative integers with row-sum vector $(a_1,a_2,\dots)$ and column-sum vector $(b_1,b_2,\dots)$. These are very well-studied numbers, but in general there is no simple formula. Their computation is in fact #P-complete. One reference is Chapter 7 of Enumerative Combinatorics, vol. 2. See for instance Corollary 7.12.3.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8412861227989197, "perplexity": 78.85045418050578}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207926736.56/warc/CC-MAIN-20150521113206-00288-ip-10-180-206-219.ec2.internal.warc.gz"}
https://cs.stackexchange.com/questions/103675/is-this-counting-language-context-free
# Is this counting language context free? Let $$\Sigma = \left\{ 0,\,1,\,2\right\}$$. I want to look at the following language: $$L=\left\{ xyz \, | \, |x|_0 + |z|_0 = |x|_2 +|z|_2 \wedge y \in \left\{ 1 \right\} ^{*} \right\}$$. I would like to prove or disprove $$L$$ being context-free I'm having a very hard time to construct a context-free grammer for $$L$$, so I attempted to build a non-deterministic PDA. My attempt goes in the direction of: If i want to sum the zeroes from $$x$$ and $$z$$, and fron this sum to substract the sum of $$2$$s from both $$x$$ and $$z$$, I can do it differently: I can count the number of $$0$$s from $$x$$ by inserting a #, then remove one # for every $$2$$ I count, if negative I'll push @ . Then I will do the same with $$z$$. The points are: 1. I'm not sure if this idea will work? 2. If this language is CFL, I have no idea how to begin constructing a CFL grammer for it? • @dkaeae That's a pretty invasive edit that seems to go beyond clarification into the terrain of stylistic adaptions. Your first line of action should be to ask the OP for clarification, and then to only edit if they don't and/or there are obvious typos or formatting mishaps. The style of the presentation should be in the OP's responsibility. – Raphael Jan 31 at 16:52 I can count the number of 0s from $$x$$ by inserting a #, then remove one # for every 2 I count, if negative I'll push @ . Then I will do the same with $$𝑧$$. This PDA is pretty good. You can certainly continue to convert this PDA to a context-free grammar according to some algorithm such as the one given in the proof of the language of a PDA is context-free, although that procedure is usually rather heavy and less enlightening. If this language is CFL, I have no idea how to begin constructing a CFL grammer for it? I would agree that it might not be easy. I would recommend that you should read this answer to how to produce a context-free grammar by Hendrik Jan if you have not yet. Here is the simplest context-free grammar for $$L$$. $$S \to 0S2S \mid 2S0S \mid 1S \mid \epsilon$$ Note that $$L=\{ xz \mid |x|_0 + |z|_0 = |x|_2 +|z|_2\}=\{ w \mid |w|_0 = |w|_2 \}$$. That is, $$L$$ is the language of words with equal number of 0s and 2s, as you have observed. Here is a brief reasoning that shows the language of above grammar is $$L$$. • It is immediate that all words generated by $$S$$ has equal number of $$0$$'s and $$2$$'s. • Suppose $$1^kw\in L$$ for some $$w$$ that does not start with 1. • Suppose $$w$$ starts with 0. If we count the number of 0s and the number of 2s in $$w$$ starting from the starting 0, there will be a time that the number of 2s catches up with the number of 0s. So $$w=0w_12w_2$$ for some $$w_1, w_2\in L$$. • Suppose $$w$$ starts with 2. Similarly, $$w=2w_10w_2$$ for some $$w_1, w_2\in L$$. Combining the above two cases with the base case $$1^k$$ for $$k\ge0$$, which can be generated, we see by mathematical induction that all words in $$L$$ can be generated. 1. Your idea is correct. After all, $$|x|_0 + |z|_0 = |x|_2 + |z|_2$$ if and only if $$|x|_0 + |z|_0 - (|x|_2 + |z|_2) = |xz|_0 - |xz|_2 = 0$$. Note, however, you must (ab)use nondeterminism correctly so as to guess where $$y$$ is so as to know what belongs to $$x$$ and what to $$z$$. 2. Producing a CFG from a PDA is, unfortunately, not easy in general. There are cases in which, given a CFL, it is (much) easier to construct one than it is to construct the other. Having experience in these kind of exercises might give you a sudden flash of inspiration; failing that, you can always resort to the standard construction. In this particular case, your "flash of inspiration" could have come from taking a closer look at the equation on the number of zeroes and two's. The idea is the equation holds for the empty word and we can produce symbols while conserving it; that is, for example, if you generate a zero for $$x$$ (i.e., $$|x|_0$$ is increased by one), then you also want to generate a $$2$$ somewhere so the equation still holds. Hence, a CFG with the following set of productions suffices: \begin{align*} S &\to X_0 S X_2 \mid X_2 S X_0 \mid X S X \mid Y \mid \varepsilon, \\ X_0 &\to X 0 \mid 0 X, \\ X_2 &\to X 2 \mid 2 X, \\ X &\to 0 X 2 \mid 2 X 0 \mid 1 X \mid X 1 \mid \varepsilon, \\ Y &\to 1 Y \mid \varepsilon \end{align*} A brief reasoning to make sure every word $$w = x y z \in L$$ is generated: The ones in $$x$$ and $$z$$ are generated by $$X$$, so we can exclude them. Further, any neighboring pair $$02$$ or $$20$$ may be generated by $$X$$ (and we include $$X$$ everywhere it is possible for such a pair to appear). Removing all such pairs, the only possible word left is of the form $$x = 0^n$$ and $$z = 2^n$$ or $$x = 2^n$$ and $$z = 0^n$$, and we can generate $$w$$ by using the appropriate rules for $$S$$ containing $$X_0$$ and $$X_2$$. • Nice answer. Note that $1012\in L$ is not generated. – Apass.Jack Feb 1 at 2:56 • @Apass.Jack Whoops. I was under the impression $x, z \in \{ 0, 2 \}^\ast$. (It seems there is much more nondeterminism at work than I'd expected.) Patched by adding production rules for ones everywhere. Still, your answer looks far more elegant :) – dkaeae Feb 1 at 8:20
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 60, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9832813143730164, "perplexity": 294.523225270836}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627998879.63/warc/CC-MAIN-20190619003600-20190619025600-00303.warc.gz"}
https://stke.sciencemag.org/content/2000/14/tw8
Editors' ChoiceCell Biology See allHide authors and affiliations Science's STKE  11 Jan 2000: Vol. 2000, Issue 14, pp. tw8 DOI: 10.1126/stke.2000.14.tw8 The list of key signaling events that are regulated by controlled ubiquitin-dependent degradation of proteins continues to grow. Brondello et al. report that MAP kinase phosphatase-1 (MKP-1), which dephosphorylates and inactivates the p42 and p44 mitogen-activated protein (MAP) kinases, is degraded by the proteasome. The phosphatase itself is a target of the p42 MAPK and p44 MAPK enzymes that it inactivates. This phosphorylation of MKP-1 appears not to regulate its enzymatic activity, but rather to decrease its ubiquitin-dependent degradation. The MAP kinases also increase transcription of the gene encoding MKP-1, thus providing two signals that increase the abundance of MKP-1. Such regulation would appear to limit prolonged activation of the p42 and p44 MAP kinases. Brondello, J-M., Pouysségur, J., and McKenzie, F.R. (1999) Reduced MAP kinase phosphatase-1 degradation after p42/p44MAPK-dependent phosphorylation. Science 286: 2514-2517. [Abstract] [Full Text]
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9439298510551453, "perplexity": 26586.411904021545}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153980.55/warc/CC-MAIN-20210730185206-20210730215206-00173.warc.gz"}
http://rin.io/tag/complex-geometry/
## Calculating the Period Matrix of a Shiga Curve, $y^3 = x^4-1$. Thanks to Dami Lee for patiently walking through how to compute the period matrix of this 12-fold cyclic cover of a thrice punctured sphere, and thanks Matthias Weber for showing me how to write the 3-fold cover of a 5-punctured sphere as a 12-fold cyclic cover of a thrice punctured sphere. Most of the figures are either hand-drawn or made using Geogebra. Note that I will sometimes use $\tau$ to denote $2 \pi$. All errors are mine and mine alone. Motivational Sidenote: This is part of my project in attempting to understand the notion of height (in formal group law theory) in terms of the symmetry of the underlying variety. Though it may seem disparate, this post is the computation of the Jacobian of a Shiga curve shown to have height 3 properties by Sebastien Thyssen and Hanno van Woerden.  Continue reading Calculating the Period Matrix of a Shiga Curve, $y^3 = x^4-1$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8664513826370239, "perplexity": 776.8273709671557}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948520218.49/warc/CC-MAIN-20171213011024-20171213031024-00106.warc.gz"}
http://www.mathnet.ru/php/archive.phtml?wshow=paper&jrnid=tmf&paperid=5581&option_lang=eng
RUS  ENG JOURNALS   PEOPLE   ORGANISATIONS   CONFERENCES   SEMINARS   VIDEO LIBRARY   PACKAGE AMSBIB General information Latest issue Archive Impact factor Subscription Guidelines for authors License agreement Submit a manuscript Search papers Search references RSS Latest issue Current issues Archive issues What is RSS TMF: Year: Volume: Issue: Page: Find TMF, 1992, Volume 91, Number 2, Pages 334–345 (Mi tmf5581) The problem of an expanding shell in the relativistic theory of gravitation A. A. Vlasov, O. V. Monovsky M. V. Lomonosov Moscow State University Abstract: The gravitational field of an expanding shelf matched in a continuously differentiable manner at its boundaries is obtained in the linear approximation of the relativistic theory of gravitation. Analysis of the second approximation of the problem indicates that there is convergence of the employed expansion with respect to the gravitational constant. Full text: PDF file (1072 kB) References: PDF file   HTML file English version: Theoretical and Mathematical Physics, 1992, 91:2, 561–569 Bibliographic databases: Citation: A. A. Vlasov, O. V. Monovsky, “The problem of an expanding shell in the relativistic theory of gravitation”, TMF, 91:2 (1992), 334–345; Theoret. and Math. Phys., 91:2 (1992), 561–569 Citation in format AMSBIB \Bibitem{VlaMon92} \by A.~A.~Vlasov, O.~V.~Monovsky \paper The problem of an expanding shell in the relativistic theory of gravitation \jour TMF \yr 1992 \vol 91 \issue 2 \pages 334--345 \mathnet{http://mi.mathnet.ru/tmf5581} \mathscinet{http://www.ams.org/mathscinet-getitem?mr=1182543} \transl \jour Theoret. and Math. Phys. \yr 1992 \vol 91 \issue 2 \pages 561--569 \crossref{https://doi.org/10.1007/BF01018854} \isi{http://gateway.isiknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&DestLinkType=FullRecord&DestApp=ALL_WOS&KeyUT=A1992KG55600011} • http://mi.mathnet.ru/eng/tmf5581 • http://mi.mathnet.ru/eng/tmf/v91/i2/p334 SHARE: Citing articles on Google Scholar: Russian citations, English citations Related articles on Google Scholar: Russian articles, English articles This publication is cited in the following articles: 1. O. V. Monovsky, “The equation of motion of the spherically symmetric shell in relativistic gravitational theory”, Theoret. and Math. Phys., 107:2 (1996), 698–705 2. A. A. Vlasov, “Gravitational field of expanding spherically symmetric shell in RTG in linear velocity approximation”, Theoret. and Math. Phys., 108:1 (1996), 912–915 • Number of views: This page: 198 Full text: 91 References: 16 First page: 2
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.20612534880638123, "perplexity": 11512.767770413097}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703511903.11/warc/CC-MAIN-20210117081748-20210117111748-00108.warc.gz"}
https://brilliant.org/problems/math-archive-10/
# Math archive (10) Calculus Level 4 If $$f:\mathbb R^{+}\rightarrow\mathbb R$$ and $$f(x)f(y)=f(xy)+2\left(\frac{x+y}{xy}+1\right)$$ for all positive $$x,y$$. Find $$3$$ times the sum of all possible values of $$f\left(\frac{1}{2}\right)$$. Try my set Math Archive × Problem Loading... Note Loading... Set Loading...
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9323921799659729, "perplexity": 2306.4475756241263}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676591718.31/warc/CC-MAIN-20180720154756-20180720174756-00233.warc.gz"}
https://link.springer.com/article/10.1007/s00500-017-2615-6
Soft Computing , Volume 22, Issue 12, pp 4047–4070 # Convergence analysis of standard particle swarm optimization algorithm and its improvement • Weiyi Qian • Ming Li Methodologies and Application ## Abstract Standard particle swarm optimization (PSO) algorithm is a kind of stochastic optimization algorithm. Its convergence, based on probability theory, is analyzed in detail. We prove that the standard PSO algorithm is convergence with probability 1 under certain condition. Then, a new improved particle swarm optimization (IPSO) algorithm is proposed to ensure that IPSO algorithm is convergence with probability 1. In order to balance the exploration and exploitation abilities of IPSO algorithm, we propose the exploration and exploitation operators and weight the two operators in IPSO algorithm. Finally, IPSO algorithm is tested on 13 benchmark test functions and compared with the other algorithms published in the recent literature. The numerical results confirm that IPSO algorithm has the better performance in solving nonlinear functions. ## Keywords Standard particle swarm optimization Analysis of algorithm Convergence in probability Global optimization ## Notes ### Acknowledgements This work is partly supported by the National Natural Science Foundation of China (11371071) and Scientific Research Foundation of Liaoning Province Educational Department (L2013426). ### Conflict of interest The authors declare that there is no conflict of interests regarding the publication of this paper. ### Ethical approval This article does not contain any studies with human participants or animals performed by any of the authors. ## References 1. Adewumi AO, Arasomwan AM (2016) An improved particle swarm optimizer based on swarm success rate for global optimization problems. J Exp Theor Artif Intell 28(3):441–483 2. Afshar MH (2012) Large scale reservoir operation by constrained particle swarm optimization algorithms. J Hydro-Environ Res 6(1):75–87 3. Ali MM, Kaelo P (2008) Improved particle swarm algorithms for global optimization. Appl Math Comput 196(2):578–593 4. Arasomwan AM, Adewumi AO (2014) An investigation into the performance of particle swarm optimization with various chaotic maps. Math Probl Eng pp 1–17, Article ID 178959. doi: 5. Behnamian J, Fatemi Ghomi SMT (2010) Development of a PSO-SA hybrid metaheuristic for a new comprehensive regression model to time-series forecasting. Expert Syst Appl 37(2):974–984 6. Chen CH, Chen YP (2011) Convergence time analysis of particle swarm optimization based on particle interaction. Adv Artif Intell pp. 1–7, Article ID 204750. doi: 7. Chuang LY, Yang CH, Li JC (2011) Chaotic maps based on binary particle swarm optimization for feature selection. Appl Soft Comput 11(1):239–248 8. Clerc M, Kennedy J (2002) The particle swarm: explosion, stability, and convergence in a multidimensional complex space. IEEE Trans Evol Comput 6(1):58–73 9. Ding J, Liu J, Chowdhury KR, Zhang W, Hu Q, Lei J (2014) A particle swarm optimization using local stochastic search and enhancing diversity for continuous optimization. Neurocomputing 137:261–267 10. Emara HM, Fattah AHA (2004) Continuous swarm optimization technique with stability analysis. In: Proceedings of American control conference, Boston, MA, pp 2811–2817Google Scholar 11. Feng Y, Teng GF, Wang AX, Yao YM (2007) Chaotic inertia weight in particle swarm optimization. In: Proceedings of the 2nd international conference on innovative computing, information and control, KumamotoGoogle Scholar 12. Fernandez-Martinez JL, Carcia-Gonzalo E, Saraswathi S, Jernigan R, Kloczkowski A (2011) Particle swarm optimization: a powerful family of stochastic optimizers. Analysis, design and application to inverse modelling. In: Proceedings of the 2nd international conference on advances in swarm intelligence. Springer, Berlin, pp 1–8Google Scholar 13. Franken N, Engelbrecht A (2005) Particle swarm optimization approaches to coevolve strategies for the iterated prisoner’s dilemma. IIEEE Trans Evol Comput 9(6):562–579 14. Gandomi AH, Yun GJ, Yang XS, Talatahari S (2013) Chaos-enhanced accelerated particle swarm optimization. Commun Nonlinear Sci Numer Simul 18(2):327–340 15. Jiang M, Luo YP, Yang SY (2007) Stochastic convergence analysis and parameter selection of the standard particle swarm optimization algorithm. Inf Process Lett 102(1):8–16 16. Jin XL, Ma LH, Wu TJ, Qian JX (2007) Convergence analysis of the particle swarm optimization based on stochastic processes. Acta Autom Sin 33(12):1263–1268 17. Kadirkamanathan V, Selvarajajah K, Fleming PJ (2006) Stability analysis of the particle dynamics in particle swarm optimizer. IEEE Trans Evol Comput 10(3):45–255 18. Kennedy J, Eberhart RC (1995) Particle swarm optimization. In: IEEE international conference on neural networks, Piscataway, NJ, pp 1942–1948Google Scholar 19. Li N, Sun DB, Zou T, Qin YQ, Wei Y (2006) An analysis for a particle’s trajectory of PSO based on difference equation. Chin J Comput 29(11):2052–2061Google Scholar 20. Lim WH, Isa NAM (2014) Teaching and peer-learning particle swarm optimization. Appl Soft Comput 18:39–58 21. Malik RF, Rahman TA, Hashim SZM, Ngah R (2007) New particle swarm optimizer with sigmoid increasing inertia weight. Int J Comput Sci Secur 1(2):35–44Google Scholar 22. Nickabadi A, Ebadzadeh MM, Safabakhsh R (2011) A novel particle swarm optimization algorithm with adaptive inertia weight. Appl Soft Comput 11:3658–3670 23. Ozcan, E, Mohan CK (1998) Analysis of a simple particle swarm optimization system. In: Intelligent engineering systems through artificial neural networks, pp 253–258Google Scholar 24. Ozcan, E, Mohan CK (1999) Particle swarm optimization: surfing the waves. In: Congress on evolutionary computation. IEEE Press, Washington, DC, pp 1939–1944Google Scholar 25. Pan F, Li XT, Zhou Q, Li WX, Gao Q (2013) Analysis of standord particle swarm optimization algorithm based on Markov chain. Acta Autom Sin 39(4):281–289Google Scholar 26. Rapaic MR, Kanovic Z (2009) Time-varying PSO-convergence analysis, convergence-related parameterization and new parameter adjustment schemes. Inf Process Lett 109(11):548–552 27. Shen X, Chi Z, Yang J, Chen C, Chi Z (2010) Particle swarm optimization with dynamic adaptive inertia weight. In: International conference on challenges in environmental science and computer engineering, Wuhan, pp 287–290Google Scholar 28. Shi YH, Eberhart RC (1999) Empirical study of particle swarm optimization. In: IEEE international conference on evolutionary computation, Washington, DC, pp 1945–1950Google Scholar 29. Trelea IC (2003) The particle swarm optimization algorithm: convergence analysis and parameter selection. Inf Process Lett 85(6):317–325 30. van den Bergh F (2002) An analysis of particle swarm optimization. PhD thesis, Department of Computer Science, University of PretoriaGoogle Scholar 31. van den Bergh F, Engelbrecht AP (2006) A study of particle swarm optimization particle trajectories. Inf Sci 176(8):937–971 32. Victoire A, Jeyakumar AE (2005) Reserve constrained dynamic dispatch of units with valve-point effects. IEEE Trans Power Syst 20(3):1273–1282 33. Wang L, Yang B, Chen Y (2014) Improving particle swarm optimization using multi-layer searching strategy. Inf Sci 274:70–94 34. Yao X, Liu Y, Lin G (1999) Evolutionary programming made faster. IEEE Trans Evol Comput 3(2):81–102Google Scholar 35. Yu X, Zhang X (2014) Enhanced comprehensive learning particle swarm optimization. Appl Math Comput 242:265–276 36. Zou F, Wang L, Hei X, Chen D, Yang D (2014) Teaching-learning-based optimization with dynamic group strategy for global optimization. Inf Sci 273:112–131
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8914298415184021, "perplexity": 15943.30753733436}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583660175.18/warc/CC-MAIN-20190118151716-20190118173716-00322.warc.gz"}
https://www.physicsforums.com/threads/simplification-on-tangens.129627/
# Simplification on Tangens 1. Aug 24, 2006 ### philipp2020 hi i have a question on a simplification tan (45° + b) - tan (45° -b) Then I put on both sides the Theorems for Addition. But I am not sure if my result is right. What is most possible simplification here? is it 1 + 4 tan b / 1 - tan^2 b ??? Thanks very much for an answer. Greetings Philipp Last edited: Aug 24, 2006 2. Aug 24, 2006 ### BobG Double check the tangent sum and difference identities. Your denominator is good, but I can't figure out what mistake you could have made to get the numerator that you did. If you get the numerator correct, there's another identity that allows you to simplify the fraction. 3. Aug 25, 2006 ### philipp2020 ah so the numerator is only 4 tan b and not plus 1 yes... oh ok... i see then at the end it will be just 2 * tan 2b Similar Discussions: Simplification on Tangens
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9226621389389038, "perplexity": 1868.5569879200625}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128329344.98/warc/CC-MAIN-20170629135715-20170629155715-00550.warc.gz"}
http://mathhelpforum.com/calculus/85443-pplication-double-integrals.html
# Thread: pplication of double integrals 1. ## pplication of double integrals i need help with this. evaluate : Double integral x dA by changing the polr coordinates, where D is the region in the pfirst quadrant that lies between the circles x^2+y^2=4 and x^2+y^2=2x Thank you 2. Originally Posted by fearless901 i need help with this. evaluate : Double integral x dA by changing the polar coordinates, where D is the region in the first quadrant that lies between the circles x^2+y^2=4 and x^2+y^2=2x Thank you The first circle is centred at the origin and has radius 2. Complete the square for the other circle to write it as $(x-1)^2+y^2=1$. So it has radius 1 and is centred at (1,0). Now draw a diagram. You should be able to see from the diagram that for a given value of θ (between 0 and π/2), the region between the two circles corresponds to r going from $2\cos\theta$ to 2. Thus $\iint_Dx\,dA = \int_0^{\pi/2}\!\!\int_{2\cos\theta}^2r\cos\theta\,rdrd\theta$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 3, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9803274869918823, "perplexity": 475.5924991092541}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988720967.29/warc/CC-MAIN-20161020183840-00560-ip-10-171-6-4.ec2.internal.warc.gz"}
https://www.physicsforums.com/threads/matlab-diary-function-help.555914/
Matlab diary function help 1. Dec 1, 2011 jemma I am using the diary function to store simulation results into a txt file, e.g. diary('TextLog.txt') however I expect to have around 50,000 rows of data and so not all my data will be stored. Is there anything I can do to specify the size of the file? Am I right in thinking the file will only hold what is displayed in the Command Window? If so, is there a way to change this? Unless there is a different command I can use? Thanks if you can help! 2. Dec 1, 2011 3. Dec 1, 2011 jemma Yes thanks, I've been trying to do this using fprintf. With the example here: http://www.mathworks.com/help/techdoc/ref/fprintf.html B = [8.8 7.7 ; ... 8800 7700]; fprintf('X is %4.2f meters or %8.3f mm\n', 9.9, 9900, B) MATLAB displays: X is 9.90 meters or 9900.000 mm X is 8.80 meters or 8800.000 mm X is 7.70 meters or 7700.000 mm What does the %4.2f and %8.3f do apart from rounding values to 2 and 3 decimal places respectively? i.e. what's the 4 and 8? Thanks! 4. Dec 2, 2011 jhae2.718 The 4 and 8 specify the width of the printed output. So, for %8.3f, if you has 12.345, it would print: Code (Text): 12.345 and if you had 0.1236 it would print: Code (Text): 0.124 padding with whitespace so that the column width is 8. Try this code to see how that works: Code (Text): number = 12.3456789; [COLOR="Blue"]for[/COLOR] i=1:10; format = sprintf([COLOR="DarkOrchid"]'%%%i.3f\n'[/COLOR], i); fprintf(format, number); [COLOR="blue"]end[/COLOR] It should produce: Code (Text): 12.346 12.346 12.346 12.346 12.346 12.346 12.346 12.346 12.346 12.346 Last edited: Dec 2, 2011 Similar Discussions: Matlab diary function help
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8455817699432373, "perplexity": 4113.465869060793}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084890514.66/warc/CC-MAIN-20180121100252-20180121120252-00374.warc.gz"}
http://swmath.org/software/16633
# GAPDoc GAP package GAPDoc. This package describes a document format for writing GAP documentation. The idea is to define a sufficiently abstract markup language for GAP documentation which can be (relatively easily) converted into different output formats. We used XML to define such a language. ## References in zbMATH (referenced in 1 article ) Showing result 1 of 1. Sorted by year (citations)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8523483276367188, "perplexity": 5785.515634410925}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698543782.28/warc/CC-MAIN-20161202170903-00229-ip-10-31-129-80.ec2.internal.warc.gz"}
https://zenodo.org/record/4389238/export/dcite4
Working paper Open Access # What Is the Concise explanation of a Hilarious Essay? Jeremy Barlow ### DataCite XML Export <?xml version='1.0' encoding='utf-8'?> <identifier identifierType="DOI">10.5281/zenodo.4389238</identifier> <creators> <creator> <creatorName>Jeremy Barlow</creatorName> </creator> </creators> <titles> <title>What Is the Concise explanation of a Hilarious Essay?</title> </titles> <publisher>Zenodo</publisher> <publicationYear>2020</publicationYear> <subjects> <subject>writing</subject> </subjects> <dates> <date dateType="Issued">2020-12-23</date> </dates> <resourceType resourceTypeGeneral="Text">Working paper</resourceType> <alternateIdentifiers> <alternateIdentifier alternateIdentifierType="url">https://zenodo.org/record/4389238</alternateIdentifier> </alternateIdentifiers> <relatedIdentifiers> <relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.5281/zenodo.4389237</relatedIdentifier> </relatedIdentifiers> <rightsList> <rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights> </rightsList> <descriptions> <description descriptionType="Abstract">&lt;p&gt;&lt;strong&gt;What Is A Amusing Essay?&lt;/strong&gt;&lt;/p&gt; &lt;p&gt;A hilarious essay, unlike other essay papers, is a document that is supposed to be comic. It might be an entertaining way to prove a stage or even be described as a humorous essay. A hilarious all handles the presentation of the jokes and also the structure of your respective paper. To very carefully take your factors to the kitchen table within a amusing way, you need to do research and fully grasp your matter. First, if you will find instructions give you need to realize what is required individuals by your coach very carefully. Using this method, your essay will likely be direct to the stage and also relevant.&lt;/p&gt; &lt;p&gt;&lt;strong&gt;Do you know the actions you should get?&lt;/strong&gt;&lt;/p&gt; &lt;p&gt;Humor can be brought out in a lot of ways, and although this is taken into account, you should also remember that some lessons may sound offensive to particular age bracket inside the masses. Therefore, you must study on who you are showing your pieces of paper to this way your cracks might make feeling rapidly and at the same time develop a funny setting round the area.&lt;/p&gt; &lt;p&gt;Pick a thesis declaration.&lt;/p&gt; &lt;p&gt;Professional&amp;nbsp;&lt;a href="https://www.masterpapers.com/"&gt;paper writer&lt;/a&gt; believe that the assertion you settle on must be short and simple although not some thing as well cliche or foreseeable. You could use a sarcastic matter when your theses assertion; that way, you easily produce suspense within your masses and start your demonstration funnily. Your thesis assertion will also help you with your investigation about comic points you may also use in your paper. Understand that the feeling of humour of individuals is distinct and the humor you think of will not be hilarious ample to others., though don&amp;rsquo;t let this discourage you There are numerous study options that you may use to assist you develop amusing records.&lt;/p&gt; &lt;p&gt;Choose a Scenario&lt;/p&gt; &lt;p&gt;The circumstance you select is the narrative you notify inside your essay. It needs to be exciting and a shirt, but at the same time, it needs to impress your audience.&lt;/p&gt; &lt;p&gt;There is absolutely no principal construction to use when producing a amusing essay. Nevertheless, the normal kind of all documents is definitely used to have a fair demonstration in the paper: an intro, a system along with a summary paragraph. Your intro continue to must be catching and very humorous. It requires to notify the competition that and this is what you are going to focus on. The section should be short and to the stage. Don&amp;rsquo;t use long sentences because it becomes hard for the audience to keep up; instead, simplify your issues to be short and still bring out the same information. Your whole body sentences need to state your case narrative. It needs to be full of sarcasm and humour. This can be and also hardwearing. Market interested.&lt;/p&gt; &lt;p&gt;The past section, the actual final outcome must say thanks to your viewers with regard to their some time and also conclude funnily. Lastly, proofread your essay and pick out the faults manufactured. It will always be recommended setting away the papers for a while and later on on buy it for editing and enhancing. This way, you are able to develop upgrades, and it can be simpler to pick up the errors. You can even go through it out to a few of your friends to discover the impression they have in the pieces of paper.&lt;/p&gt;</description> </descriptions> </resource> 98 2 views
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2322341352701187, "perplexity": 1633.6218456719466}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618039626288.96/warc/CC-MAIN-20210423011010-20210423041010-00112.warc.gz"}
https://forums.mydigitallife.net/threads/removewat-install-message-at-startup.24861/
# RemoveWAT install message at startup Discussion in 'Windows 7' started by shishaboy, Feb 16, 2011. 1. ### shishaboy MDL Novice Feb 16, 2011 3 0 0 #1 Last edited: Feb 22, 2011 Hi guys, I have got this problem with removewat.exe. When windows starts up I get the window "would you like install removewat.exe". whether I choose yes or no it still runs in the background and forces alg.exe to start firefox and it runs in the background and steals about 10% of the cpu. When I try to end the firefox process in windows task manager I get the same original message "would you like install removewat.exe". What I have done to fix the problem: 1. visited the removewat threat and followed the wat.fix and removewat tool, but they didn't fix the problem. 2. I went to the C:\Users\name\AppData\Local\Temp folder where the window says its there and tried to delete the removewat.exe but it didn't help as the file was replaced over and over again after every deletion. 3. went to msconfig but the file is not there and therefore could not disable start-up. oo and my windows is not activated and I am running windows 7 enterprise. I would be grateful if someone has any suggestions on how to deal with this problem. Thanks in advance. the problem was fixed, it was a virus which was detected after installing various anti-viruses but norton did the trick and the whole thing just vanished.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8366687893867493, "perplexity": 2210.4967702135245}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891814872.58/warc/CC-MAIN-20180223233832-20180224013832-00288.warc.gz"}
https://www.arxiv-vanity.com/papers/cond-mat/0503686/
arXiv Vanity renders academic papers from arXiv as responsive web pages so you don’t have to squint at a PDF. Read this paper on arXiv.org. # Entropy production along a stochastic trajectory and an integral fluctuation theorem Udo Seifert II. Institut für Theoretische Physik, Universität Stuttgart, 70550 Stuttgart, Germany ###### Abstract For stochastic non-equilibrium dynamics like a Langevin equation for a colloidal particle or a master equation for discrete states, entropy production along a single trajectory is studied. It involves both genuine particle entropy and entropy production in the surrounding medium. The integrated sum of both is shown to obey a fluctuation theorem for arbitrary initial conditions and arbitrary time-dependent driving over a finite time interval. ###### pacs: 05.40.-a, 05.70-a Introduction. – Can the notions appearing in the first and second law of thermodynamics consistently be applied to mesoscopic non-equilibrium processes like dragging a colloidal particle through a viscous fluid Wang et al. (2002); Carberry et al. (2004); Mazonka and Jarzynski (1999); van Zon and Cohen (2003a)? Concerning the first law, Sekimoto interpreted the terms in the standard overdamped Langevin equation as dynamical increments for applied work, internal energy and dissipated heat Sekimoto (1998). For the second law and, in particular, entropy, a proper formulation and interpretation is more subtle. Entropy might be considered as an ensemble property and therefore not to be applicable to a single trajectory. On the other hand, the so-called fluctuation theorem Evans et al. (1993); Evans and Searles (1994); Gallavotti and Cohen (1995); Kurchan (1998); Lebowitz and Spohn (1999); Searles and Evans (1999); Crooks (1999, 2000); Jarzynski (2000); Maes (2003); Gaspard (2004); Seifert (2005) quite generally relates the probability of entropy generating trajectories to those of entropy annihilating ones which requires obviously a definition of entropy on the level of a single trajectory. While for a colloidal particle immersed in a heat bath it is pretty clear what the entropy change of the bath is, it is less obvious whether or not one should assign an entropy to the particle itself as well. The purpose of this paper is to show that consequent adaption of a previously introduced stochastic entropy Crooks (1999); Qian (2001) to the trajectory of a colloidal particle together with the present original discussion of its equation of motion yields a consistent interpretation of entropy production along a single stochastic trajectory. Moreover, it leads to a lucid and concise identification of boundary terms in fluctuation relations. In fact, we show for arbitrary time-dependent driving that the total entropy production obeys an integral fluctuation theorem which is related to but different from Jarzynski’s non-equilibrium work relation Jarzynski (1997). The present definition of entropy also implies that the known steady-state fluctuation theorem holds for finite times rather than in the long-time limit only as previously in stochastic dynamics Kurchan (1998); Lebowitz and Spohn (1999). In a final step, this approach is generalized to arbitrary driven stochastic dynamics governed by a master equation with time-dependent rates. Entropy along a trajectory. – As a paradigm, we consider overdamped motion of a particle with mobility along a one-dimensional coordinate in the time-interval subject to a force F(x,λ)=−∂xV(x,λ)+f(x,λ). (1) This force can arise from a conservative potential and/or be applied to the particle directly as . Both sources may be time-dependent through an external control parameter varied according to some prescribed experimental protocol from to . The motion is governed by the Langevin equation ˙x=μF(x,λ)+ζ, (2) with stochastic increments modelled as Gaussian white noise with where is the diffusion constant. In equilibrium, and are related by the Einstein relation where is the temperature of the surrounding medium. We assume this relation to persist even in a non-equilibrium situation. Throughout the paper we set Boltzmann’s constant to unity such that entropy becomes dimensionless. For a definition of entropy along the trajectory, we consider first the corresponding Fokker-Planck equation for the probability to find the particle at at time as ∂τp(x,τ)=−∂xj(x,τ)=−∂x(μF(x,λ)−D∂xp(x,τ)). (3) This partial differential equation must be augmented by a normalized initial distribution . It will become crucial to distinguish the dynamical solution of this Fokker-Planck equation, which depends on this given initial condition, from the solution for which the rhs of eq. (3) vanishes at any fixed . The latter corresponds either to a steady state for or to equilibrium for , respectively. The common definition of a non-equilibrium Gibbs entropy S(τ)≡−∫dxp(x,τ)lnp(x,τ)≡⟨s(τ)⟩ (4) suggests to define a trajectory-dependent entropy for the particle (or “system”) s(τ)=−lnp(x(τ),τ) (5) where the probability obtained by solving the Fokker-Planck equation is evaluated along the stochastic trajectory . Obviously, for any given trajectory , the entropy depends on the given initial data and thus contains information on the whole ensemble. For an equilibrium Boltzmann distribution at fixed , this definition assigns an entropy s(x)=(V(x,λ)−F(λ))/T, (6) with the free energy . The definition (5) has been used previously by Crooks for stochastic microscopically reversible dynamics Crooks (1999) and by Qian for stochastic dynamics of macromolecules Qian (2001). Neither work, however, discusses the equation of motion for this stochastic entropy. Entropy production. – The rate of change of the entropy (5) is given by ˙s(τ) = −∂τp(x,τ)p(x,τ)|x(τ)−∂xp(x,τ)p(x,τ)|x(τ)˙x = −∂τp(x,τ)p(x,τ)|x(τ)+j(x,τ)Dp(x,τ)|x(τ)˙x−μF(x,λ)D|x(τ)˙x. The first equality identifies the explicit and the implicit time-dependence. The second one uses the Fokker-Planck equation. The third term in the second line can be related to the rate of heat dissipation in the medium ˙q(τ)=F(x,λ)˙x≡T˙sm(τ) (8) where we identify the exchanged heat with an increase in entropy of the medium at temperature . Then (Entropy production along a stochastic trajectory and an integral fluctuation theorem) can be written as a balance equation for the trajectory-dependent total entropy production ˙stot(τ)=˙sm(τ)+˙s(τ)=−∂τp(x,τ)p(x,τ)|x(τ)+j(x,τ)Dp(x,τ)|x(τ)˙x, (9) which is our first central result. The first term on the rhs signifies a change in which can be due to a time-dependent or, even at fixed , due to relaxation from a non-stationary initial state . Upon averaging, the total entropy production rate has to become positive as required by the second law. This ensemble average proceeds in two steps. First, we average over all trajectories which are at time at a given leading to ⟨˙x|x,τ⟩=j(x,τ)/p(x,τ). (10) Second, we average over all with as ˙Stot(τ)≡⟨˙stot(τ)⟩=∫dxj(x,τ)2Dp(x,τ)≥0, (11) where equality holds in equilibrium only. Averaging the increase in entropy of the medium along similar lines leads to ˙Sm(τ) ≡ ⟨˙sm(τ)⟩=⟨F(x,τ)˙x⟩/T (12) = ∫dxF(x,τ)j(x,τ)/T. (13) Hence upon averaging, the increase in entropy of the system itself becomes . On the ensemble level, this balance equation for the averaged quantities has previously been derived directly from the ensemble definition (4) Qian (2001). The key point of our approach is that we have defined entropy production (or annihilation) along a single stochastic trajectory splitting it up into a medium part and a part of the particle (system). Beyond the conceptual advantage, this identification facilitates a discussion of fluctuation theorems. Fluctuation theorem. – Fluctuation theorems derive from the behaviour of the weight of a trajectory under “time-reversal” which associates with each protocol a reversed one and a reversed trajectory For a given initial value and final value , the ratio of probabilities of the forward path and of the backward path can easily be calculated in the path integral representation of the Langevin equation as Kurchan (1998) lnp[x(τ)|x0]~p[~x(τ)|~x0]=∫t0F(x,τ)˙x dτ/T=Δsm. (14) If this quantity is combined with arbitrary normalized distributions for initial and final value and , respectively, according to R[x(τ),λ(τ);p0,p1] ≡ lnp[x(τ)|x0] p0(x0)~p[~x(τ)|~x0] p1(~x0) (15) = Δsm+lnp0(x0)p1(xt), (16) one easily derives the integral fluctuation relation Maes (2003) ⟨e−R⟩ ≡ ∑x(τ),x0p[x(τ)|x0] p0(x0)e−R (17) = ∑~x(τ),~x0~p[~x(τ)|~x0] p1(~x0)=1. Here, the average is over both initial values drawn from the (in principle arbitrary) initial distribution and trajectories determined by the noise history . Since the normalized distribution is arbitrary, there are, even for fixed , an infinity of choices for which obey the constraint (17) and its implication . At least two choices of have physical meaning in the present context. First, for which is the solution of the Fokker-Planck equation for the given initial distribution , the definition (5) implies that the last term in (16) becomes the entropy change of the particle along the trajectory. Hence (17) implies the integral fluctuation theorem ⟨e−Δstot⟩=1 (18) which is our second main result. This integral theorem for is truely universal since it holds for any kind of initial condition (not only for ), any time-dependence of force and potential, with (for ) and without (for ) detailed balance at fixed , and any length of trajectory without the need for waiting for relaxation. Crucial for this universality is our identification of the boundary term in (16) as the change in entropy of the particle. As a second important choice, for and , one recovers Jarzynski’s relation Jarzynski (1997) since in this case R=Δsm+[V(xt,λt)−V(x0,λ0)−F(λt)+F(λ0)]/T=wd/T (19) where is the part of the work which is irreversibly lost as dissipated into the medium. The difference between the two choices for is subtle but important. In the first case, the fluctuation theorem holds for the total entropy change along the trajectory evaluated at the very end of the protocol. For Jarzynski’s relation, one takes the distribution corresponding to equilibrium at the final value of . The difference arises from relaxation of the system towards the final equilibrium state at constant which further increases the averaged entropy of the particle. In fact, is the one choice which leads to the smallest among all possible . For a steady state at constant and constant force like for motion along a ring with periodic boundary conditions, by choosing in (16), one obtains the stronger fluctuation relation Crooks (1999); Maes (2003) p(−R)/p(R)=e−R. (20) Since with the definition (5) the last term in (16) is again the change in entropy of the system , the quantity becomes the total entropy change . Hence, one recovers the fluctuation theorem for the total change in entropy as p(−Δstot)/p(Δstot)=e−Δstot (21) even for a finite length of the trajectories. In contrast, previous derivations of this genuine fluctuation theorem within stochastic dynamics Kurchan (1998); Lebowitz and Spohn (1999) hold in the long-time limit only since they implicitly ignore what we call and consider only . Since the former is bounded for finite potentials, the latter will always win in the long run. Generalizations. – It is obvious that the present discussion holds as well for systems with more than one degree of freedom obeying overdamped coupled Langevin equations. Rather than spelling out the notational details, we will now discuss a more general stochastic dynamics on a discrete set of states. Again, we aim at a consistent definition of an entropy along a trajectory without having available any a priori notion of heat contrary to the colloidal case above which facilitated the identification of entropy production in the medium there. Let a transition between discrete states and occur with a rate , which depends on an externally controlled time-dependent parameter . The master equation for the time-dependent probability then reads ∂τpn(τ)=∑m≠n[wmn(λ)pm(τ)−wnm(λ)pn(τ)]. (22) For a solution, an initial distribution must be specified as well. As above, the system is driven externally from to according to a protocol . For any fixed , there is a steady state which may or may not obey detailed balance . A stochastic trajectory starts at and jumps at times from to ending up at . As entropy along this trajectory, we define s(τ)≡−lnpn(τ)(τ) (23) where is the solution of the master equation (22) for a given initial distribution taken along the specific trajectory . As in the colloidal case, this entropy will depend on the chosen initial distribution. The entropy becomes time-dependent due to two sources. First, even if the system does not jump, can be time-dependent either for time-independent rates due to possible relaxation from a non-stationary initial state or, for time-dependent rates, due to the explicit time-dependence of . Including the jumps, the change of system entropy reads ˙s(τ)=−∂τpn(τ)(τ)pn(τ)(τ)−∑jδ(τ−τj)lnpn+jpn−j. (24) Here, and in the remainder, we suppress notationally the time-dependence of both and the rates in the jump terms. We now split up the rhs into a total entropy production and one of the medium as follows ˙stot(τ)≡−∂τpn(τ)(τ)pn(τ)(τ)−∑jδ(τ−τj)lnpn+jwn+jn−jpn−jwn−jn+j (25) and ˙sm(τ)≡−∑jδ(τ−τj)lnwn+jn−jwn−jn+j (26) such that the balance holds. The rational behind the identification (26) for the increase in entropy of the medium becomes clear after averaging over many trajectories. For this average, we need the probability for a jump to occur at from to which is . Hence, one gets ˙Sm(τ)≡⟨˙sm(τ)⟩=∑n,kpnwnklnwnkwkn, (27) ˙Stot(τ)≡⟨˙stot(τ)⟩=∑n,kpnwnklnpnwnkpkwkn (28) and ˙S(τ)≡⟨˙s(τ)⟩=∑n,kpnwnklnpnpk (29) such that the global balance with is valid. By averaging our stochastic expressions, we thus recover and generalize established results for the non-equilibrium ensemble entropy balance available so far for the steady state only Schnakenberg (1976); Lebowitz and Spohn (1999); foot2 . For the fluctuation theorems, the stochastic quantity is derived from the probability of a trajectory to occur under the protocol and the probability for the reversed trajectory to occur under the reversed protocol . With an arbitrary initial distribution and an arbitrary final distribution it becomes R[n(τ),λ(τ);p0n,p1n]≡lnP[n(τ)|n0]p0n0~P[~n(τ)|~n0]p1~n0=Δsm+lnp0n0p1nt. (30) From the infinity of possible fluctuation relations , we choose two important ones. First, for ) and , the last term in (30) becomes the increase in system entropy and the total entropy change. Hence, we have again the integral theorem (18). Second, the choice of that corresponds to Jarzynski’s relation in the colloidal case above is implied in the theorem derived in Ref. Seifert (2004). Finally, in a steady state for time-independent rates, by choosing , one has the detailed version (21) for the total entropy change for any finite length of the trajectory as exemplified for a molecular motor or enzym in Ref. Seifert (2005). Summarizing perspective. – We have expressed the entropy production along a single stochastic trajectory as a sum of an entropy production of the system and of the medium both for a colloidal particle and for general stochastic dynamics obeying a master equation. The total entropy production obeys an integral fluctuation theorem for arbitrary time-dependent driving, for arbitrary initial conditions and any length of trajectories. This theorem and the Jarzynski relation are both shown to be special cases of an infinity of possible fluctuation relations. With the present definition of entropy, the detailed fluctuation theorem valid in steady states for the total entropy production holds even for trajectories of finite length. The trajectory dependent entropy of the particle could be measured experimentally for a time-dependent protocol by first recording over many trajectories the probability distribution from which the entropy of each trajectory can be inferred. With such data, one could also test the new integral fluctuation theorem (18) and compare it to Jarzynski’s relation for the same protocol. It will be interesting to derive, both experimentally and theoretically, the probability distribution of these entropy changes and to see, e.g., whether they are Gaussian for slow driving as is the dissipated work appearing in Jarzynski’s relation Speck and Seifert (2004). ## References Want to hear about new tools we're making? Sign up to our mailing list for occasional updates.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9780206084251404, "perplexity": 657.4149640655717}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439737225.57/warc/CC-MAIN-20200807202502-20200807232502-00526.warc.gz"}
http://mathhelpforum.com/advanced-statistics/207945-probability-statistics-homework-problem-question.html
Thread: Probability and Statistics homework problem question 1. Probability and Statistics homework problem question Example 1: Computers, fine arts, economics - this one requires using system of equations There are 500 seniors. 210 are enrolled in computers 80 do not need any of the 3 80 are taking only fine arts. 180 are taking fine arts. 36 taking only economics and computers 10 taking only fine arts and computers 220 taking economics Find: 1.) P(only taking economics) 2.) P(economics and fine arts) 3.) P(taking all 3 classes) 2. Re: Probability and Statistics homework problem question (View the picture to understand how i got the system) So, you have the system 1) x + 10 + y + 36 = 210 2) 80 + 10 + z + y = 180 3) 36 + z + y + t = 220 4) x + 10 + y + 36 + t + z + 80 = 420 4 Systems, 4 variables, solve Attached Thumbnails 3. Re: Probability and Statistics homework problem question how would i go about and solve this....
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8244529366493225, "perplexity": 2817.873202193808}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463612502.45/warc/CC-MAIN-20170529165246-20170529185246-00004.warc.gz"}
http://scicomp.stackexchange.com/questions/4889/schurs-complement-and-inverse-of-block-matrices
Schur's Complement and Inverse of Block Matrices Assume that we are given a block matrix of the form: $$M = \left[ \begin{array}{cc} A & b \\ b^T & c \\ \end{array} \right]$$ where $b$ is a column vector. and $c$ is a scalar. Schur's complement of $A$ in $M$ is given by: $$s = c - b^T A^{-1} b$$ Assume that we know $A$ is invertible and we want to check if $M$ is invertible. Is it true that $M$ is invertible if, and only if $s$ is different than zero? Thank you - Yes. Under the assumption that $A^{-1}$ exists, then $\det (M) = \det (A) s.$ Since $\det(A) \neq 0$, $\det (M) \neq 0$ if and only if $s \neq 0$. Thus $M$ is inveritible if and only if $s \neq 0$. This theorem on block determinants is standard material that can be found in many linear algebra textbooks. See for example Carl Meyer's textbook, "Matrix Analysis and Linear Algebra." The proof is pretty simple- write $M$ as $M=\left[ \begin{array}{cc} I & 0 \\ b^{T}A^{-1} & 1 \\ \end{array} \right] \left[ \begin{array}{cc} A & b \\ 0 & c-b^{T}A^{-1}b \\ \end{array} \right]$ Then apply the product rule and the rule for determinants of (block) diagonal matrices. - Mind you, it's not a very good idea to use this in computational practice to check for invertibility of $M$. – Brian Borchers Dec 20 '12 at 18:17 I think practicality hinges on the difficulty of solving $Au=b$, from which computation of $s$ will easily follow. The phrase bordered matrix may be a useful search term. – hardmath Dec 23 '12 at 15:39 Thank you for you reply. You mention that it is not a very good idea to use this in computational practice to check for invertibility of $M$. It would obviously be hard to get $s$ = exact zero, but at least $s$ should in the order of $10^{10}$ or even less if $M$ is singular, right? Can't this be used as a check? I can't seem to find a 100% reliable singularity check. – Mohammad Fawaz Dec 23 '12 at 22:22 The determinant is a very bad measure of singularity in practice, because it isn't scale invariant. For example, suppose that I tell you a matrix $M$ has $\det(M)=1.0 \times 10^{-15}$. You might argue that this means the matrix is singular or nearly singular. Now, let me tell you that this particular matrix is of size 15 by 15, and $M=0.1I$. Clearly, $M$ is nowhere near singular. The proper measure of singularity to use is the condition number, but this requires some knowledge of the eigenvalues of $M$. – Brian Borchers Dec 25 '12 at 3:31
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9510371685028076, "perplexity": 206.65025516846706}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368708690512/warc/CC-MAIN-20130516125130-00033-ip-10-60-113-184.ec2.internal.warc.gz"}
https://stats.meta.stackexchange.com/questions/5235/why-was-the-only-thread-with-the-reversal-badge-deleted
I notice that the Reversal badge has been awarded just a single time, to @Macallan in 2015. I was rather sad to see that for something this rare - an answer apparently that good to a question apparently that terrible - that the question was subsequently deleted. But this question from @Glen_b says "It looks like OPs can only delete a question with an answer if there's only one answer and the answer isn't upvoted". Does this mean that the question was deleted by mods? Is it okay to ask what happened there? And what pearls of wisdom I missed out on because that answer of >+20 was deleted? • I wrote about this topic once in Main Meta Stack Exchange with some suggestions: meta.stackexchange.com/questions/88709/…. The “update” part in the question body it was also me; in that time I came to the conclusion that most (or at least many) of “reversal” posts get deleted with time. – Andre Silva Apr 19 '18 at 13:45 • The question was a rant (likely to be deleted), but for the purpose of keeping as much content as possible alive ( in order to keep the back-end of the website, which new visitors can't see, as open as possible), it would be nice if the links to deleted content would still be available and the censorship only removes the content of the posts or whatever else is considered to be deleted. – Sextus Empiricus Apr 19 '18 at 15:50 • @AndreSilva thanks for that read, and I like your suggestions. – mkt Apr 19 '18 at 20:26 • mkt -- keep in mind the possibility that my comment that you quote may not be quite correct either. I do make mistakes. – Glen_b Apr 20 '18 at 0:32 The question you mention is How to deal with a fake scientist?, it got the -30 score and was deleted by the community, as voted by amoeba, kjetil b halvorsen, Xi'an on Nov 4 '17 at 13:25, neither of whom is or was a mod at that time. You gain the more privileges on the site with reputation, so at some point users are able to cast delete votes. Please notice that the number of upvotes does not need to be a good measure of the quality of a question, or an answer. In this case, the answer is just a quote and a reference. I guess that it got so many upvotes because people upvoted it to say "I agree that the question is wrong", rather then to reward it for an exceptional quality. • Thanks - I appreciate the explanation and the clarification. – mkt Apr 19 '18 at 7:59 • To satisfy the morbid curiosity of any who don't have the rep required to view deleted questions, an archive is available. It's only missing a few un-expanded comments and some votes versus the version at deletion (when the question was at -30 and the answer at +36). – Danica Apr 19 '18 at 12:41 The question was deleted because enough votes were cast to delete it (none by a diamond moderator, as it happens). It was a rant, written in response to a frustrating experience on our site, not a real question; & perhaps by now rather embarrassing to the user that posted it—who will be, after all, four years older. The answer took the question at face value: though it received many up-votes, it consisted of just a reference with a brief quotation. Davidian, M. and Louis, T. A. (2012), Why Statistics? Science, vol. 336, no. 6077, p. 12. • Thanks! Nice to have that cleared up. And I appreciate the citation. – mkt Apr 19 '18 at 8:00 • I upvoted the answer, and elected not to delete that thread, because it was a remarkably gentle and effective response to a question that (probably unintentionally) was little more than a trolling expedition. – whuber Apr 19 '18 at 12:34
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4738234281539917, "perplexity": 1024.5839661648238}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153531.10/warc/CC-MAIN-20210728060744-20210728090744-00201.warc.gz"}
https://bibli.cirm-math.fr/listRecord.htm?list=link&xRecord=19283190157910013729
m • E F Nous contacter 0 # Documents  Rinot, Assaf | enregistrements trouvés : 2 O P Q Déposez votre fichier ici pour le déplacer vers cet enregistrement. ## Distributive Aronszajn trees Rinot, Assaf | CIRM H Post-edited Research talks;Logic and Foundations It is well-known that the statement "all $\aleph_1$-Aronszajn trees are special'' is consistent with ZFC (Baumgartner, Malitz, and Reinhardt), and even with ZFC+GCH (Jensen). In contrast, Ben-David and Shelah proved that, assuming GCH, for every singular cardinal $\lambda$: if there exists a $\lambda^+$-Aronszajn tree, then there exists a non-special one. Furthermore: Theorem (Ben-David and Shelah, 1986) Assume GCH and that $\lambda$ is singular cardinal. If there exists a special $\lambda^+$-Aronszajn tree, then there exists a $\lambda$-distributive $\lambda^+$-Aronszajn tree. This suggests that following stronger statement: Conjecture. Assume GCH and that $\lambda$ is singular cardinal. If there exists a $\lambda^+$-Aronszajn tree, then there exists a $\lambda$-distributive $\lambda^+$-Aronszajn tree. The assumption that there exists a $\lambda^+$-Aronszajn tree is a very mild square-like hypothesis (that is, $\square(\lambda^+,\lambda)$). In order to bloom a $\lambda$-distributive tree from it, there is a need for a toolbox, each tool taking an abstract square-like sequence and producing a sequence which is slightly better than the original one. For this, we introduce the monoid of postprocessing functions and study how it acts on the class of abstract square sequences. We establish that, assuming GCH, the monoid contains some very powerful functions. We also prove that the monoid is closed under various mixing operations. This allows us to prove a theorem which is just one step away from verifying the conjecture: Theorem 1. Assume GCH and that $\lambda$ is a singular cardinal. If $\square(\lambda^+,<\lambda)$ holds, then there exists a $\lambda$-distributive $\lambda^+$-Aronszajn tree. Another proof, involving a 5-steps chain of applications of postprocessing functions, is of the following theorem. Theorem 2. Assume GCH. If $\lambda$ is a singular cardinal and $\square(\lambda^+)$ holds, then there exists a $\lambda^+$-Souslin tree which is coherent mod finite. This is joint work with Ari Brodsky. See: http://assafrinot.com/paper/29 It is well-known that the statement "all $\aleph_1$-Aronszajn trees are special'' is consistent with ZFC (Baumgartner, Malitz, and Reinhardt), and even with ZFC+GCH (Jensen). In contrast, Ben-David and Shelah proved that, assuming GCH, for every singular cardinal $\lambda$: if there exists a $\lambda^+$-Aronszajn tree, then there exists a non-special one. Furthermore: Theorem (Ben-David and Shelah, 1986) Assume GCH and that $\lambda$ is singular ... Déposez votre fichier ici pour le déplacer vers cet enregistrement. ## Chain conditions, unbounded colorings and the $C$-sequence spectrum Rinot, Assaf | CIRM H Post-edited Research talks;Logic and Foundations The productivity of the $\kappa$-chain condition, where $\kappa$ is a regular, uncountable cardinal, has been the focus of a great deal of set-theoretic research. In the 1970’s, consistent examples of $kappa-cc$ posets whose squares are not $\kappa-cc$ were constructed by Laver, Galvin, Roitman and Fleissner. Later, ZFC examples were constructed by Todorcevic, Shelah, and others. The most difficult case, that in which $\kappa = \aleph{_2}$, was resolved by Shelah in 1997. In the first part of this talk, we shall present analogous results regarding the infinite productivity of chain conditions stronger than $\kappa-cc$. In particular, for any successor cardinal $\kappa$, we produce a ZFC example of a poset with precaliber $\kappa$ whose $\omega ^{th}$ power is not $\kappa-cc$. To do so, we introduce and study the principle $U(\kappa , \mu , \theta , \chi )$ asserting the existence of a coloring $c:\left [ \kappa \right ]^{2}\rightarrow \theta$ satisfying a strong unboundedness condition. In the second part of this talk, we shall introduce and study a new cardinal invariant $\chi \left ( \kappa \right )$ for a regular uncountable cardinal $\kappa$ . For inaccessible $\kappa$, $\chi \left ( \kappa \right )$ may be seen as a measure of how far away $\kappa$ is from being weakly compact. We shall prove that if $\chi \left ( \kappa \right )> 1$, then $\chi \left ( \kappa \right )=max(Cspec(\kappa ))$, where: (1) Cspec$(\kappa)$ := {$\chi (\vec{C})\mid \vec{C}$ is a sequence over $\kappa$} $\setminus \omega$, and (2) $\chi \left ( \vec{C} \right )$ is the least cardinal $\chi \leq \kappa$ such that there exist $\Delta\in\left [ \kappa \right ]^{\kappa }$ and b : $\kappa \rightarrow \left [ \kappa \right ]^{\chi }$ with $\Delta \cap \alpha \subseteq \cup _{\beta \in b(\alpha )}C_{\beta }$ for every $\alpha < \kappa$. We shall also prove that if $\chi (\kappa )=1$, then $\kappa$ is greatly Mahlo, prove the consistency (modulo the existence of a supercompact) of $\chi (\aleph_{\omega +1})=\aleph_{0}$, and carry a systematic study of the effect of square principles on the $C$-sequence spectrum. In the last part of this talk, we shall unveil an unexpected connection between the two principles discussed in the previous parts, proving that, for infinite regular cardinals $\theta< \kappa ,\theta \in Cspec(\kappa )$ if there is a closed witness to $U_{(\kappa ,\kappa ,\theta ,\theta )}$. This is joint work with Chris Lambie-Hanson. The productivity of the $\kappa$-chain condition, where $\kappa$ is a regular, uncountable cardinal, has been the focus of a great deal of set-theoretic research. In the 1970’s, consistent examples of $kappa-cc$ posets whose squares are not $\kappa-cc$ were constructed by Laver, Galvin, Roitman and Fleissner. Later, ZFC examples were constructed by Todorcevic, Shelah, and others. The most difficult case, that in which $\kappa = \aleph{_2}$, ... #### Filtrer ##### Audience Titres de périodiques et e-books électroniques (Depuis le CIRM) Ressources Electroniques Books & Print journals Recherche avancée 0 Z
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9631686806678772, "perplexity": 1029.9714375185565}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178367790.67/warc/CC-MAIN-20210303200206-20210303230206-00335.warc.gz"}
https://physics.stackexchange.com/questions/364126/the-usage-of-covariant-coordinates-in-relativistic-field-theories
# The usage of covariant coordinates in relativistic field theories In the opening chapters of typical QFT books, the covariant coordinates $x_\mu = g _{\mu\nu}x^\nu$ $x^\mu = (t,x,y,z)$ and the differential operator $\partial^\mu = \frac{\partial}{\partial x_{\mu}}=(\frac{\partial}{\partial t}, -\nabla)$ are introduced to facilite the evaluation of spacetime interval $x^2$ and the magnitude square of a four gradient of some scalar field $\partial ^\mu \phi \partial _\mu \phi$. On the other hand, we do not introduce the $x_\mu$ in general relativity where the spacetime is curved and our coordinate is curvilinear, in general. My questions are the followings: 1. What is the reason for the absence of $x_{\mu}$ in general relativity where curvilinear coordinates are necessary for describing a curved spacetime? 2.Does $x_\mu = g_{\mu\nu}x^\nu$ make sense only in the flat spacetime where the coordinate $x^\mu$ themselves are contravariant component of some vector, and the $x^\mu$ transform like $x^{\prime\mu} = \frac{\partial x^{\prime\mu}}{\partial x^\nu}x^\nu$ in this special circumstance? I would like to share some of my opinions for further discussion. (Please evaluate and point out the mistakes of my arguments.) In fact, $\partial^\mu$ is fully equivalent to $g^{\mu\nu}\partial_\nu$ in flat spacetime characterized by a constant metric. ($x_\mu = g_{\mu\nu}x^\nu$, $x^\mu = g^{\mu\nu}x_\nu$ so $\partial^{\mu} = \frac{\partial}{\partial x_{\mu}}=\frac{\partial x^{\nu}}{\partial x_{\mu}}\frac{\partial}{\partial x^{\nu}}=g^{\mu\nu}\partial_{\nu}$ We can always work only with $x^\mu$ and $\partial_\mu$ at the expense of always carrying the inverse metric $g^{\mu\nu}$ in the kinetic term of the Lagrangian. However, if $x^\mu$ is a set of curvilinear coordinates describing a curved spacetime, then the coordinates $x^\mu$ cease to carry any contravariant vectorial significance and they are merely a label of a point in spacetime without any vectorial meaning. (Euclidean analogy: $(x,y,z)$ are the components of a position vector $\vec{r}$ but a general curvilinear coordinate describing the same Euclidean space are not components of a vector, in general.) Consequently, it no longer makes sense to define $x_{\mu} = g_{\mu\nu}x^{\nu}$ for a general coordinate $x^\mu$ describing a curved spacetime. (We could do this in a Euclidean Space with $x,y,z$ but this is redundant as the metric is identity. On the other hand, $x_\mu$ is not equal to $x^\mu$ in a flat-spacetime due to the pseudo-Euclidean nature of Minkowskian spacetime. This is the reason why we need $x_{\mu}$ and $x^{\mu}$ in relativistic QFT.) On the other hand, a vector $\vec{V}$ always have two kinds of components $V^\mu$ and $V_\mu$, regardless of the spacetime geometry and the coordinate system describing it. Now the covariant components of the gradient of a scalar field $\phi$ is naturally obtained by $\frac{\partial \phi}{\partial x^\mu}$. If there is no such thing as $x_\mu$ for a set of general coordinate $x^ \mu$ describing a curved spacetime, then the only way to obtain the contravariant component of the gradient is to employ the inverse metric $g^{\mu\nu}\frac{\partial}{\partial x^ \nu}$. (And we have to raise the lower index that comes from taking derivative with respect to coordinate $x^\mu$ in this manner using the inverse metric when manipulating other field variables.) This might be the reason why there is no $x_\mu$ in general relativity. In summary, the introduction and usage of $x_\mu$ depend on the contravariant vectorial nature of $x^{\mu}$ which depends on the flatness of spacetime and linearity of the coordinate system describing it. (Am I correct?) • You are essentially entirely correct, which makes an answer somewhat redundant. – Erik Jörgenfelt Oct 21 '17 at 10:23
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8992013335227966, "perplexity": 157.85155453623366}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141197278.54/warc/CC-MAIN-20201129063812-20201129093812-00179.warc.gz"}
http://www.creationevolutionuniversity.com/science/?cat=1
Poisson Nachman Crowell Even aside from Basener and Sanford, others including Nobel Prize winner Hermann Muller pointed out the human race cannot tolerate very many mutations per individual per generation. The number Muller arrived at was about 1 bad mutation per generation per individual as the limit the human genome can tolerate. Additionally, so what if an individual has a good mutation if he has 10 bad to go with it. This is like have a slight increase in intelligence while having 10 heritable diseases to go with it. You go one step forward and ten steps back. Can natural selection arrest the problem? Only if there are enough reproductive resources relative to the number of offspring per couple. For human populations there was something published by Nachman and Crowell and Eyre-Walker and Keightley using a Poisson distribution as reasonable model for the probability of a eugenically clean individual appearing in the face of various mutation rates. If it is improbable that an eugenically clean kid can be reproduced by a couple, this makes it hard to weed out the bad. So this is an alternative way to arrive at Muller’s conclusions, which are also Sanford and Basener’s conclusions, and really everyone else’s conclusions as summarized by Dan Gruar: “If ENCODE is right, evolution is wrong.” This is a simpler argument than the one Basener and Sanford put forward, but to Sanford’s credit, he’s also put the simpler version in his book Genetic Entropy, although the following derivation isn’t in his book, it’s something I ginned up myself. 🙂 So how can we estimate the probability a kid can be born with no defective mutations? The following derivation was confirmed in Kimrua’s paper (see eqn. 1.4) https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1211299/pdf/1337.pdf which Nachman and Crowell, and Eyre-Walker and Keightley reference as well. So now the details: let U = mutation rate (per individual per generation) P(0,U) = probability of individual having no mutation under a mutation rate U (eugenically the best) P(1,U) = probability of individual having 1 mutation under a mutation rate U P(2,U) = probability of individual having 2 mutations under a mutation rate U etc. The wiki definition of Poisson distribution is: $\huge f(k,\lambda ) = e^{-\lambda }\frac{\lambda^k }{k!}$ to conform the wiki formula with evolutionary literature let $\lambda = U$ and $f = P$ Because P(0,U) = probability of individual having no mutation under a mutation rate U (eugenically the best), we can find the probability the eugenically best individual emerges by letting: $k = 0$ which yields $\large \large P(k,U) = P(0,U) = \frac{U^0 e^{-U }}{0!} = e^{-U}$ Given the Poisson distribution is a discrete probability distribution, the following idealization must hold: $\large \sum_{n}P_n =\sum_{i=0}^{\infty}P(i,U) = 1$ thus $\large \large P(0,U) + \sum_{i=1}^{\infty}P(i,U) = 1$ thus subtracting P(0,U) from both sides $\large \large P(0,U) + \sum_{i=1}^{\infty}P(i,U) -P(0,U) = 1 - P(0,U)$ thus simplifying $\large \sum_{i=1}^{\infty}P(i,U) = 1 - P(0,U)$ On inspection, the left hand side of the above equation must be the percent of offspring that have at least 1 new mutation. Noting gain that $P(0,U) = e^{-U}$, the above equation reduces to the following: $\sum_{i=1}^{\infty}P(i,U) = 1 - P(0,U) = 1- e^{-U}$ which is in full agreement with Nachman and Crowell’s equation in the very last paragraph and in full agreement with an article in Nature: High genomic deleterious mutation rates in homonids by Eyre-Walker and Keightley, paragraph 2. http://www.lifesci.sussex.ac.uk/CSE/members/aeyrewalker/pdfs/EWNature99.pdf The simplicity and elegance of the final result is astonishing, and simplicity and elegance lend force to arguments. So what does this mean? If the bad mutation rate is 6 per individual per generation (more conservative than Gruar’s estimate if ENCODE is right), using that formula, the chances that a eugenically “ideal” offspring will emerge is: $\large \large P(0,6) = e^{-6} = 0.25\%$ This would imply each parent needs to procreate the following number of kids on average just to get 1 eugenically fit kid: $\frac{1}{e^{-U}} = \frac{1}{e^{-6}} = 403.42$ Or equivalently each couple needs to procreate the following number of kids on average just to get 1 eugenically fit kid: $\large \large 2 * \frac{1}{e^{-U}} = 2 * \frac{1}{e^{-6}} \approx 807$ For humanity to survive, even after each couple has 807 kids on average, we still have to make the further utterly unrealistic assumption that the eugenically “ideal” offspring are the only survivors of a selective process. Hence, it is absurd to think humanity can purge the bad out of its populations — the bad just keeps getting worse. In truth, since most mutations are of nearly neutral effect, most of the damaged offspring will reproduce, and the probability of a eugenically ideal line of offspring approaches zero over time. Muller’s number of only 1 new bad mutations per generation per individual. So if anything I understated my case. There are some “fixes” to the problem suggested by Crow and Kondrashov. I suggested my fix. But the bottom line is to look at what is actually happening to the human genome over time. Are we getting dumber and sicker? I think so. It’s sad. We can test Basener and Sanford’s prediction by observing whether human heritable diseases continue to increase with each generation. Whether their derivation is right or not, some of their conclusions are observationally and experimentally testable. On some level, I suppose even Basener and Sanford wished it were not so because it is a tragic conclusion. Absolute Fitness in Theoretical Evolutionary Genetics Joe Felsenstein, like other population geneticists, holds a special place in the Creation/Evolution controversy because his works are regarded highly by many creationists who are familiar with genetics. This is a thread for all of us (myself included) to try to learn and understand one of the key concepts in his book Theoretical Evolutionary Genetics, namely absolute fitness. He has generously made his book available on his website (a book of this calibre could sell for hundreds of dollars). Continue reading Absolute Fitness in Theoretical Evolutionary Genetics Hello world! Welcome to WordPress. This is your first post. Edit or delete it, then start blogging! $i\hbar\frac{\partial}{\partial t}\left|\Psi(t)\right>=h\left|\Psi(t)\right>$ Hey Brother! The Lord bless you. An historical footnote — the great irony of history is that as much as Einstein hated QM, he won the Nobel Prize for his contribution to quantum Mechanics!!! Here is one of his famous equations about E that he shares with Max Planck, the famous Plank-Einstein relation of Quantum Mechanics: $E = h \nu$ https://en.wikipedia.org/wiki/Planck%E2%80%93Einstein_relation of course Einstein’s other famous equation about E 🙂 is: $E = mc^2$ As much as Einstein was famous for his theories of QM (which he ironically hated but also pioneered), he was more famous for Special and General Relativity. What I found very bizarre was that in Relativity we often drop the measurement of time in terms of seconds but use METERS! That is units of distance. Below is my General Relativity book, by Bernard Shutz. During the Thanksgiving/Christmas season last year, Phoodoo began a thread on relativity, and when I started going through the ideas again, I could no longer resist and just had to revisit my old books… And then you were the only one in the last several years who was interested in QM’s connection to ID and that also helped spark a revival in me of old ideas I was almost losing to forgetfulness. It was in the process of shaking the dust off Griffiths book that I finally saw his discussion of the realist, the mentalist (aka Copenhagen), and agnostic interpretations of QM. When I studied QM, all the theological and philosophical implication was sanitized out of by the professor since he was focused on the math. Many of my classmates were more interested in QM’s implications for lasers and semiconductors and chemistry. Half the US economy is based on QM, so the theological implications were thrown by the wayside while I studied. But then, because of your interest, I revisted some of my old essays and then Griffiths book and then it came alive in a way I had not appreciated previously, especially the last chapter which I never learned in his book because the professor didn’t cover it 7 years ago! But that was the best chapter, on Bell’s theorem that establishes the “mentalist” (aka Copenhagen) interpretation. So to what you said: maybe on subatomic level, time, distance past and future don’t matter or they simply don’t exist… this is what Einstein said: Einstein’s belief in an undivided solid reality was clear to him, so much so that he completely rejected the separation we experience as the moment of now. He believed there is no true division between past and future, there is rather a single existence. His most descriptive testimony to this faith came when his lifelong friend Besso died. Einstein wrote a letter to Besso’s family, saying that although Besso had preceded him in death it was of no consequence, “…for us physicists believe the separation between past, present, and future is only an illusion, although a convincing one.” http://everythingforever.com/einstein.htm Although perhaps Einstein was saying this because of relativity, it is moreso true because of QM. And in my experience, if Physicists had to choose which theory takes precedence over all others, it would be QM. And now it makes sense why a prominent professor at my school said: “The ultimate cause of atheism, Newton asserted, is ‘this notion of bodies having, as it were, a complete, absolute and independent reality in themselves.’” The 1925 discovery of quantum mechanics solved the problem of the Universe’s nature. Bright physicists were again led to believe the unbelievable — this time, that the Universe is mental. …. According to Sir James Jeans: “the stream of knowledge is heading towards a non-mechanical reality; the Universe begins to look more like a great thought than like a great machine. Mind no longer appears to be an accidental intruder into the realm of matter…we ought rather hail it as the creator and governor of the realm of matter.” …. The Universe is immaterial — mental and spiritual. Richard Conn Henry Nature 2005, vol 436, The Mental Universe http://henry.pha.jhu.edu/The.mental.universe.pdf Now it’s coming back to me, and now it’s making more sense than ever. There is a God. Praise be!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 17, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5992661714553833, "perplexity": 2064.588994870182}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247481428.19/warc/CC-MAIN-20190217010854-20190217032854-00432.warc.gz"}
https://cp4space.hatsya.com/2013/08/23/new-aperiodic-monotile/
# New aperiodic monotile In three dimensions, an aperiodic monotile is a solid capable of tiling space, but not in such a way that admits translational symmetry. The question of existence of aperiodic monotiles stems from a weaker question, which formed part of Hilbert’s eighteenth problem: Does there exist a polyhedron capable of tiling space, but not tile-transitively? The first example of an anisohedral tile was found in 1928 by Karl Reinhardt. An aperiodic monotile would obviously be anisohedral, although it surprisingly took 60 years before Schmitt found the first aperiodic example in 1988. The purpose of this post is to present a new monotile I discovered a few days ago, and which seems like it could have been discovered much earlier than 1988. I also claim that its structure and proof of aperiodicity are simpler (i.e. easier to reconstruct) than any other aperiodic monotile. ### Structure The aperiodic monotile is essentially a 5 × 5 Lego brick, where the grid of knobbles on the top (but not the complementary indentations on the bottom) have been rotated by the arctangent of ¾. Five such monotiles are shown in the diagram above. Four of the knobbles overhang, although reducing their diameter could overcome that. Similarly, if you prefer polyhedra to arbitrary solids, you can replace the cylindrical knobbles with square prisms. Its existence stems from the Pythagorean identity 3² + 4² = 5², and this construction should generalise to any primitive Pythagorean triple. Of course, this is the simplest, and therefore the most preferable. Note that in this case, the knobbles on the top are in a centred square arrangement, rather than a square (and that relies on 3 and 4 being consecutive). ### Proof The only result that we’ll use is the fact that the Gaussian integers (complex numbers with integer real and imaginary parts) form a unique factorisation domain. This isn’t any more difficult to prove than the Fundamental Theorem of Arithmetic, which is the equivalent statement over the ordinary integers. The first part of the proof is to show that the only tilings are the ‘obvious’ ones, where the monotiles form layers, each of which is a square lattice, and each layer is rotated by the arctangent of ¾ with respect to the layer below. Each tile forces tiles to exist above and below it, and rotated by this angle. This produces some overhang, which means that surrounding tiles are forced and (by induction) the entire tiling is arranged in layers. The next part of the proof shows that no two layers have the same orientation. This is equivalent to showing that the complex number $\frac{1}{5}(3 + 4i) = \dfrac{2 + i}{2 - i}$ is not a root of unity. Suppose that it is a kth root of unity, in which case we obtain the identity $(2 + i)^k = (2 - i)^k$, which clearly violates unique factorisation. Hence, by a reductio ad absurdum argument, we win. Consequently, if the tiling has any translational symmetry at all, it must be purely horizontal. We represent a translation by (x, y, 0) by the complex number z = x + i y. By considering the knobbles on one layer, we know that z must be a Gaussian integer (i.e. divisible by 1). By considering the knobbles on the layer above, z must also be divisible by $\dfrac{2 + i}{2 - i}$. Considering all of these layers and using unique factorisation, we get that z is divisible by all integer powers of 2 + i, which is only the case for z = 0. So the tiling cannot have non-trivial translational symmetries, completing the proof of aperiodicity. Q. E. D. ### Weakly versus strongly aperiodic This monotile, together with Schmitt’s original monotile and the Schmitt-Conway-Danzer tile, can tile space screw-symmetrically. As such, it is described as weakly aperiodic. There are no known examples of strongly aperiodic monotiles (admitting no infinite cyclic groups of symmetries), but a step forward in this direction is the aperiodic monotile by Joan Taylor and Joshua Socolar. This entry was posted in Uncategorized. Bookmark the permalink. ### 0 Responses to New aperiodic monotile 1. Johnicholas says: If you were interested in using a crystal design for macroscopic self-replication purposes, one might want to build a cage of tiles surrounding a cavity, and then pour a liquid into the cavity – casting a new tile. If the tiles are somewhat compliant, you might be able to pour the liquid in via a narrow tube (a sprue?) running along a seam – but could the sprue perhaps be accomodated in the design of the tile? • apgoucher says: I wouldn’t use aperiodic tiles for this purpose. Truncated octahedra form an elegant periodic tiling, and they don’t have structurally-dubious planar ‘cracks’ (c.f. http://projecteuler.net/problem=215 ) in the same way that the boring cubic honeycomb does. Another candidate is the rhombic dodecahedron. As for your sprue problem, you could do it by having a semicircular valley on one face of the tile, and a complementary semicircular ridge on another face (for tiling and casting purposes). If you’re allowed to rearrange the tiles to remove the new tile from the cage, then this seems to be a valid solution. I’ll investigate this in greater detail shortly. 2. wojowu says: I was wondering how many faces would you need to make polyhedral aperiodic tile using this method, and I got an answer of 156, but two of sides would not be simply connected (would be “holey”). I started wondering if it is possible to modify your construction to get tiles without holey sides and with no two adjacent segments coplanar (e.g. two squares lying flat next to each other). Your construction is impossible to translate to convex polyhedron, but is there any example of convex aperiodic tile? Or was it proven impossible? How about set of convex tiles? 3. Joseph Myers says: I don’t think there’s any significant difference between this tile and the others that admit screw-symmetric tilings. If anything, it just introduces extra complexity compared to constructions where an arbitrary irrational rotation angle can be used. • apgoucher says: Warren Smith remarked that this has the property that the union of finitely many layers is periodic, whereas this is not the case for the Schmitt-Conway-Danzer tile.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 3, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7282946705818176, "perplexity": 778.7534568332312}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337421.33/warc/CC-MAIN-20221003133425-20221003163425-00423.warc.gz"}
https://verification.asmedigitalcollection.asme.org/gasturbinespower/article-abstract/134/5/051502/477088/Displacement-Speed-Statistics-for-Stratified?searchresult=1
The statistics of the density-weighted displacement speed of the reaction progress variable $c$ isosurfaces for stratified mixture combustion arising from localized ignition in a turbulent planar coflowing jet have been studied based on 3D Direct Numerical Simulation data where the jet is considered to be fuel-rich and the coflow is taken to be fuel-lean. The resulting flame following successful ignition shows the premixed mode of combustion in fuel-rich and fuel-lean zones although an additional diffusion flame branch was also observed on the stoichiometric mixture isosurface at early times of flame evolution. The flame propagation characteristics have been analyzed in terms of the reaction, normal diffusion and tangential diffusion components of the density-weighted displacement speed for different values of reaction progress variables across the flame brush. It has been found that the reaction, normal diffusion and tangential diffusion components of density-weighted displacement speed, remain the major contributors to the density-weighted displacement speed at all stages of flame evolution as the magnitude of the component which originates due to mixture inhomogeneity remains negligible in comparison to the magnitudes of other components in accordance with previous experimental studies. It has been demonstrated that curvature and tangential strain rate dependences of the reaction progress variable gradient play key roles in determining strain rate dependences of the reaction and normal diffusion components of the density-weighted displacement speed. It has been shown that the interrelation between tangential strain rate and curvature affects the strain rate dependence of tangential diffusion component of the density-weighted displacement speed. The density-weighted displacement speed and curvature are found to be predominantly negatively correlated throughout the flame brush at all stages of the flame evolution. The relative strengths of the tangential strain rate dependence of the reaction, normal diffusion and tangential diffusion components of the density-weighted displacement speed ultimately determine the nature of correlation between the density-weighted displacement speed and the tangential strain rate. The strain rate and curvature dependences of the density-weighted displacement speed in stratified mixtures are found to be qualitatively similar to the statistics previously obtained for turbulent premixed flames. ## References 1. Peters , N. , 2000, Turbulent Combustion , Cambridge University Press , Cambridge, UK . 2. Bray , K. , Domingo , P. , and Vervisch , L. , 2005, “ Role of the Progress Variable in Models for Partially Premixed Turbulent Combustion ,” Combust. Flame , 141 , pp. 431 437 . 3. Echekki , T. , and Chen , J. H. , 1996, “ Unsteady Strain Rate and Curvature Effects in Turbulent Premixed Methane-Air Flames ,” Combust. Flame , 106 , pp. 184 202 . 4. Chen , J. H. , and Im , H. G. , 1998, “ Correlation of Flame Speed With Stretch in Turbulent Premixed Methane/Air Flames ,” Proc. Combust. Inst. , 27 , pp. 819 826 . 5. Peters , N. , Terhoeven , P. , Chen , J. H. , and Echekki , T. , 1998, “ Statistics of Flame Displacement Speeds From Computations of 2-D Unsteady Methane-Air Flames ,” Proc. Combust. Inst. , 27 , pp. 833 839 . 6. Hawkes , E. R. , and Chen , J. H. , 2005, “ Evaluation of Models for Flame Stretch Due to Curvature in the Thin Reaction Zones Regime ,” Proc. Combust. Inst. , 30 , pp. 647 655 . 7. Chakraborty , N. , and Cant , S. , 2004, “ Unsteady Effects of Strain Rate and Curvature on Turbulent Premixed Flames in an Inflow-Outflow Configuration ,” Combust. Flame , 137 , pp. 129 147 . 8. Chakraborty , N. , and Cant , R. S. , 2005, “ Effects of Strain Rate and Curvature on Surface Density Function Transport in Turbulent Premixed Flames in the Thin Reaction Zones Regime ,” Phys. Fluids , 17 , pp. 105105 . 9. Chakraborty , N. , and Cant , R. S. , 2006, “ Influence of Lewis Number on Strain Rate Effects in Turbulent Premixed Flame Propagation in the Thin Reaction Zones Regime ,” Int. J. Heat Mass Trans. , 49 , pp. 2158 2174 . 10. Jenkins , K. W. , Klein , M. , Chakraborty , N. , and Cant , R. S. , 2006, “ Effects of Strain Rate and Curvature on the Propagation of a Spherical Flame Kernel in the Thin Reaction Zones Regime ,” Combust. Flame , 145 , pp. 415 434 . 11. Klein , M. , Chakraborty , N. , Jenkins , K. W. , and Cant , R. S. , 2006, “ Effects of Initial Radius on the Propagation of Spherical Premixed Flame Kernels in Turbulent Environment ,” Phys. Fluids , 18 , p. 055102 . 12. Chakraborty , N. , Klein , M. , and Cant , R. S. , 2007, “ Stretch Rate Effects on Displacement Speed in Turbulent Premixed Flame Kernels in the Thin Reaction Zones Regime ,” Proc. Combust. Inst. , 31 , pp. 1385 1392 . 13. Chakraborty , N. , 2007, “ Comparison of Displacement Speed Statistics of Turbulent Premixed Flames in the Regimes Representing Combustion in Corrugated Flamelets and the Thin Reaction Zones ,” Phys. Fluids , 19 , p. 105109 . 14. Defransure , F. , Renou , B. , Samson , E. , Boukhalfa , A. M. , and Veynante , D. , 2003, “ Experimental and Numerical Studies of the Laminar Flame Speed of Stratified Flames ,” Proceedings of 1st European Combustion Meeting, Orleans, France, 2003. 15. Samson , E. , 2002, “ Etude Experimentale de la Propagation de Flames en Expansion dans un Milieu a Richesse Stratifiee ,” Ph.D. thesis, CNRS, Rouen, France. 16. Ray , J. , Najm , H. N. , and McCoy , R. B. , 2001, “ Ignition Front Structure in a Methane-Air Jet ,” Proceedings of the 2nd Joint Meeting of the U.S. Sections of the Combustion Institute, Oakland, CA, Paper No. 150. 17. Im , H. G. , and Chen , J. H. , 2001, “ Effects of Flow Strain on Triple Flame Propagation ,” Combust. Flame , 126 , pp. 1384 1392 . 18. Hilbert , R. , and Thevenin , D. , 2002, “ DNS of Multi-Brachial Structures With Detailed Chemistry and Transport ,” Proceedings of the 9th International Conference on Numerical Combustion, Sorrento, Italy, Paper No. 064. 19. Alvani , R. E. , and Fairweather , M. , 2002, “ Ignition Characteristics of Turbulent Jet Flows ,” Trans. Ichem E. , 80 , pp. 917 923 . 20. Chakraborty , N. , Mastorakos , E. , and Cant , R. S. , 2007, “ Effects of Turbulence on Spark Ignition in Inhomogeneous Mixtures: A Direct Numerical Simulation (DNS) Study ,” Combust. Sci Technol. , 179 ( 1–2 ), pp. 293 317 . 21. Chakraborty , N. , and Mastorakos , E. , 2006, “ Numerical Investigation of Edge Flame Propagation Characteristics in Turbulent Mixing Layers ,” Phys. Fluids , 18 , p. 105103 . 22. Hesse , H. , Chakraborty , N. , and Mastorakos , E. , 2009, “ The Effects of Lewis Number of the Fuel on Displacement Speed of Edge Flames in Igniting Turbulent Mixing Layers ,” Proc. Combust. Inst. , 32 , pp. 1435 1443 . 23. Lacaze , G. , Richardson , E. , and Poinsot , T. , 2009, “ Large Eddy Simulation of Spark Ignition in a Turbulent Methane Jet ,” Combust. Flame , 156 , pp. 1093 2009 . 24. Espí , C. V. , and Liñán , A. , 2001, “ Fast, Non-Diffusive Ignition of a Gaseous Reacting Mixture Subject to a Point Energy Source ,” Combust. Theory Modell. , 5 , pp. 485 498 . 25. Espí , C. V. , and Liñán , A. , 2002, “ Thermal-Diffusive Ignition and Flame Initiation by a Local Energy Source ,” Combust. Theory Modell. , 6 , pp. 297 315 . 26. Chen , J. H. , Choudhary , A. , de Supinski , B. , DeVries , M. , Hawkes , E. R. , , S. , Liao , W. K. , Ma , K. L. , Mellor-Crummey , J. , Podhorski , N. , Sankaran , R. , Shende , S. and Yoo , C. S. , 2009, “ Terascale Direct Numerical Simulations of Turbulent Combustion Using S3D ,” Comput. Sci. Discov. , 2 , p. 015001 . 27. Tarrazo , E. , Sanchez , A. , Liñán , A. , and Williams , F. , 2006, “ A Simple One-Step Chemistry Model for Partially Premixed Hydrocarbon Combustion ,” Combust. Flame , 147 , pp. 32 38 . 28. Malkeson , S. P. , and Chakraborty , N. , 2010, “ A-Priori Direct Numerical Simulation Analysis of Algebraic Models of Variances and Scalar Dissipation Rates for Reynolds Averaged Navier Stokes Simulations for Low Damköhler Number Turbulent Partially-Premixed Combustion ,” Combust. Sci. Technol. , 182 , pp. 960 999 . 29. Poinsot , T. , Echekki , T. , and Mungal , M. , 1992, “ A Study of Laminar Flame Tip and Implications for Turbulent Premixed Combustion ,” Combust. Sci. Technol. , 81 ( 1-3 ), pp. 45 57 . 30. Louch , D. S. , and Bray , K. N. C. , 2001, “ Vorticity in Unsteady Premixed Flames: Vorticity Pair-Premixed Flame Interactions Under Imposed Body Forces and Various Degrees of Heat Release and Laminar Flame Thickness ,” Combust. Flame , 125 , pp. 1279 1309 . 31. Treurniet , T. C. , , F. T. M. , and Boersma , B. S. , 2006, “ Direct Numerical Simulation of Homogeneous Turbulence in Combination With Premixed Combustion at Low Mach Number Modelled by G-Equation ,” J. Fluid Mech. , 565 , pp. 25 62 . 32. Gran , I. R. , Echekki , T. , and Chen , J. H. , 1996, “ Negative Flame Speed in an Unsteady 2-D Premixed Flame: A Computational Study ,” Proc. Combust. Inst. , 26 , pp. 211 218 . 33. Hélie , J. , and Trouvé , A. , 1998, “ Turbulent Flame Propagation in Partially Premixed Combustion ,” Proc. Combust. Inst. , 27 , pp. 891 898 . 34. Kollmann , W. , and Chen , J. H. , 1998, “ Pocket Formation and the Flame Surface Density Equation ,” Proc. Combust. Inst. , 27 , pp. 927 934 . 35. Chakraborty , N. , and Cant , R. S. , 2005, “ Effects of Strain Rate and Curvature on Surface Density Function Transport in Turbulent Premixed Flames in the Thin Reaction Zones Regime ,” Phys. Fluids , 17 , p. 65108 . 36. Chakraborty , N. , and Klein , M. , 2008, “ Influence of Lewis Number on the Surface Density Function Transport in the Thin Reaction Zones Regime for Turbulent Premixed Flames ,” Phys. Fluids , 20 , p. 065102 . 37. Chakraborty , N. , and Klein , M. , 2009, “ Effects of Global Flame Curvature on the Surface Density Function Transport in Turbulent Premixed Flame Kernels in the Thin Reaction Zones Regime ,” Proc. Combust. Inst. , 32 , pp. 1435 1443 . 38. Boger , M. , Veynante , D. , Boughanem , H. , and Trouvé , A. , 1998, “ Direct Numerical Simulation Analysis of Flame Surface Density Concept for Large Eddy Simulation of Turbulent Premixed Combustion ,” Proc. Combust. Inst. , 27 , pp. 917 925 . 39. Trouvé , A. , and Poinsot , T. , 1994, “ The Evolution Equation for Flame Surface Density in Turbulent Premixed Combustion ,” J. Fluid Mech. , 278 , pp. 1 31 . 40. Candel , S. M. , and Poinsot , T. J. , 1990, “ Flame Stretch and the Balance Equation for the Flame Area ,” Combust. Sci. Technol. , 70 , pp. 1 15 . 41. Chakraborty , N. , and Cant , R. S. , 2007, “ A-Priori Analysis of the Curvature and Propagation Terms of the Flame Surface Density Transport Equation for Large Eddy Simulation ,” Phys. Fluids , 19 , p. 105101 . 42. Müller , C. M. , Breitbach , H. , and Peters , N. , 1994, “ Partially Premixed Turbulent Flame Propagation in Jet Flames ,” Proc. Combust. Inst. , 25 , pp. 1099 1106 . 43. Kim , J. , Chung , S. H. , Ahn , K. Y. , and Kim , J. S. , 2006, “ Simulation of a Diffusion Flame in a Turbulent Mixing Layer by the Flame Hole Dynamics Model Using Level Set Method ,” Combust. Theory Modell. , 10 , pp. 219 240 . 44. Ballal , D. R. , and Lefebvre , A. , 1977, “ Spark Ignition of Turbulent Flowing Gases ,” Proceedings of the 15th Aerospace Sciences Meeting, AIAA Los Angeles, Paper No. 77-185. 45. Pantano , C. , 2004, “ Direct Simulation of Non-Premixed Flame Extinction in a Methane-Air Jet with Reduced Chemistry ,” J. Fluid Mech. , 514 , pp. 231 270 . 46. Rogallo , R. S. , 1981, “ Numerical Experiments in Homogeneous Turbulence ,” NASA Technical Memorandum No. 81315, NASA Ames Research Center . 47. Batchelor , G. K. , and Townsend , A. A. , 1948, “ Decay of Turbulence in the Final Period ,” Proc. Roy. Soc. A , 194 , pp. 527 542 . 48. Jenkins , K. W. , and Cant , R. S. , 1999, “ DNS of Turbulent Flame Kernels ,” Proceedings of the Second AFOSR Conference on DNS and LES, Rutgers University, Kluwer Academic Publishers, pp. 192 202 . 49. Wray , A. A. , 1990, “ Minimal Storage Time Advancement Schemes for Spectral Methods ,” NASA Ames Research Center , California , Report No. MS 202 A-1. 50. Lee , S. , Lele , S. K. , and Moin , P. , 1992, “ Simulation of Spatially Evolving Turbulence and Applicability of Taylor’s Hypothesis in Compressible Flow ,” Phys. Fluids A , 4 ( 7 ), pp. 1521 1530 . 51. Zhang , S. , and Rutland , C. J. , 1995, “ Premixed Flame Effects on Turbulence and Pressure Related Terms ,” Combust. Flame , 102 , pp. 447 461 . 52. Poinsot , T. , and Lele , S. K. , 1992, “ Boundary Conditions for Direct Simulation of Compressible Viscous Flows ,” J. Comp. Phys. , 101 , pp. 104 129 . 53. Ranga Dinesh , K. K. J. , and Kirkpatrick , M. J. , 2009, “ Study of Jet Precession, Recirculation and Vortex Breakdown in Turbulent Swirling Jets using LES ,” Comput. Fluids , 38 ( 6 ), pp. 1232 1242 . 54. Ranga Dinesh , K. K. J. , Kirkpatrick , M. P. , and Jenkins , K. W. , 2010, “ Investigation of the Influence of Swirl on a Confined Coannular Swirl Jet ,” Comput. Fluids , 39 ( 5 ), pp. 756 767 . 55. Chen , J. H. , 2011, “ Petascale Direct Numerical Simulation of Turbulent Combustion—Fundamental Insights Towards Predictive Models ,” Proc. Combust. Inst. , 33 , pp. 99 123 . 56. Haworth , D. C. , and Poinsot , T. J. , 1992, “ Numerical Simulations of Lewis Number Effects in Turbulent Premixed Flames ,” J. Fluid Mech. , 244 , pp. 405 436 . 57. Renou , B. , Boukhalfa , A. , Peuchberty , D. , and Trinité , M. , 1998, “ Effects of Stretch on the Local Structure of Freely Propagating Premixed Low-Turbulent Flames with Various Lewis Numbers ,” Proc. Combust. Inst. , 27 , pp. 841 847 . 58. Chakraborty , N. , Hesse , H. , and Mastorakos , E. , 2010, “ Numerical Investigation of Edge Flame Propagation Behaviour in an Igniting Turbulent Planar Jet ,” Combust. Sci. Technol. , 182 , pp. 1747 1781 . 59. Ko , Y. S. , and Chung , S. H. , 1999, “ Propagation of Unsteady Tribrachial Flames in Laminar Non-Premixed Jets ,” Combust. Flame , 118 , pp. 151 163 . 60. Mielenz , O. , Schlottmann , F. , and Rogg , B. , 2000, “ Propagating Methane/Air Flame in a Counterflow Geometry ,” Institut für Thermo-und Fluiddynamik, Ruhr University , Bochum, Germany , Report No. ITF/LSTM/184. 61. Won , S. H. , Kim , J. , Hong , K. J. , Cha , M. S. , and Chung , S. H. , 2005, “ Stabilization Mechanism of Lifted Flame Edge in the Near Field of Coflow Jets for Diluted Methane ,” Proc. Combust. Inst. , 30 , pp. 339 347 . 62. Ahmed , S. F. , and Mastorakos , E. , 2006, “ Spark Ignition of Lifted Turbulent Jet Flames ,” Combust. Flame , 146 , pp. 215 231 . 63. Chung , S. H. , 2007, “ Stabilization, Propagation and Instability of Tribrachial Triple Flames ,” Proc. Combust. Inst. , 31 , pp. 877 892 . 64. Kim , M. K. , Won , S. H. , and Chung , S. H. , 2007, “ Effect of Velocity Gradient on Propagation Speed of Tribrachial Flames in Laminar Coflow Jets ,” Proc. Combust. Inst. , 31 , pp. 901 908 . 65. Heeger , C. , Böhm , B. , Ahmed , S. F. , Gordon , R. , Boxx , I. , Meier , W. , Dreizler , A. , and Mastorakos , E. , 2009, “ Statistics of Relative and Absolute Velocities of Turbulent Non-Premixed Edge Flames Following Spark Ignition ,” Proc. Combust. Inst. , 32 , pp. 2957 2964 . 66. Echekki , T. , and Chen , J. H , 1998, “ Structure and Propagation of Methanol–Air Triple Flames ,” Combust. Flame , 114 , pp. 231 245 . 67. Im , H. G. , and Chen , J. H. , 1999, “ Structure and Propagation of Triple Flames in Partially Premixed Hydrogen–Air Mixtures ,” Combust. Flame , 119 , pp. 436 454 . 68. Yoo , C. , and Im , H. G. , 2005, “ Transient Dynamics of Edge Flames in a Laminar Non-Premixed Hydrogen–Air Counterflow ,” Proc. Combust. Inst. , 30 , pp. 349 356 .
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 1, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7580517530441284, "perplexity": 29836.228348037534}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320304915.53/warc/CC-MAIN-20220126041016-20220126071016-00368.warc.gz"}
https://www.physicsforums.com/threads/question-about-stress-and-strain.465651/
# Question about stress and strain • Start date • #1 hoyy1kolko 10 0 ## Homework Statement The elastic limit for a type of material is 1.6 x 10^8 Pa.Determine the minimum diameter if a wire which is made of this material so that it can support a load of 100N without the stress exceeding the elastic limit. ## Homework Equations elastic limit=stress =F/A 1.6 x 10^8=100/ (pie)d^2 ## The Attempt at a Solution • #2 Homework Helper 3,407 11 In your S = F/A formula, replace A with the area of the circular wire. Solve for the radius or diameter. Put in the numbers for S and F. • Last Post Replies 6 Views 709 • Last Post Replies 8 Views 1K • Last Post Replies 12 Views 4K • Last Post Replies 7 Views 397 • Last Post Replies 2 Views 3K • Last Post Replies 2 Views 1K • Last Post Replies 21 Views 6K • Last Post Replies 20 Views 9K • Last Post Replies 1 Views 2K • Last Post Replies 1 Views 844
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8673092126846313, "perplexity": 9305.096246610092}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882570730.59/warc/CC-MAIN-20220807211157-20220808001157-00687.warc.gz"}
https://aacrjournals.org/cancerres/article/70/1/46/559597/Cancer-Stem-Cell-Tumor-Model-Reveals-Invasive
The recently developed concept of cancer stem cells (CSC) sheds new light on various aspects of tumor growth and progression. Here, we present a mathematical model of malignancies to investigate how a hierarchical organized cancer cell population affects the fundamental properties of solid malignancies. We establish that tumors modeled in a CSC context more faithfully resemble human malignancies and show invasive behavior, whereas tumors without a CSC hierarchy do not. These findings are corroborated by in vitro studies. In addition, we provide evidence that the CSC model is accompanied by highly altered evolutionary dynamics compared with the ones predicted to exist in a stochastic, nonhierarchical tumor model. Our main findings indicate that the CSC model allows for significantly higher tumor heterogeneity, which may affect therapy resistance. Moreover, we show that therapy which fails to target the CSC population is not only unsuccessful in curing the patient, but also promotes malignant features in the recurring tumor. These include rapid expansion, increased invasion, and enhanced heterogeneity. Cancer Res; 70(1); 46–56 Major Findings This research shows that the hierarchical organization of malignant clones, as advocated in the CSC concept, has major implications for tumor biology. CSC-driven tumor growth intrinsically orchestrates tumor invasion, influences clonal selection, and has crucial consequences for the development of successful cancer treatments. Quick Guide to Equations and Assumptions Proliferation The stem cellular automaton (SCA) model is a hybrid cellular automaton (4) in which a biological cell is a point (10 × 10 μm) in a lattice. Each point can be a normal cell, a cancer cell, or a necrotic cell, and has the attributes in Table 1. After a cell division, an empty place is created by shifting the surrounding cells outward. Table 1. Cellular and microenvironmental variables in the SCA model VariablesSymbolValueReference/justification Proliferation speed (average cell cycle duration) T 20 h (44) O2 diffusion coefficient Dc 10−5 cm−2 s−1 (16) O2 concentration, healthy tissue ù 10−4 g cm−3 (16, 45) O2 consumption, proliferative cells êp 10−6 g cm−3 s−1 (46) O2 consumption, senescent cells ês 5 × 10−7 g cm−3 s−1 (47) O2 concentration resulting in necrosis θ 5 × 10−7 g cm−3 s−1 Senescent consumption as minimum concentration for cell survival Random mobility Dn 10−10 cm−2 s−1 (48) Maximum number of generations generated by DCCs H We choose to fix H to 5 and vary PS to generate various CSC fractions (3) VariablesSymbolValueReference/justification Proliferation speed (average cell cycle duration) T 20 h (44) O2 diffusion coefficient Dc 10−5 cm−2 s−1 (16) O2 concentration, healthy tissue ù 10−4 g cm−3 (16, 45) O2 consumption, proliferative cells êp 10−6 g cm−3 s−1 (46) O2 consumption, senescent cells ês 5 × 10−7 g cm−3 s−1 (47) O2 concentration resulting in necrosis θ 5 × 10−7 g cm−3 s−1 Senescent consumption as minimum concentration for cell survival Random mobility Dn 10−10 cm−2 s−1 (48) Maximum number of generations generated by DCCs H We choose to fix H to 5 and vary PS to generate various CSC fractions (3) NOTE: Variables used in the SCA model. See the Quick Guide and Materials and Methods for details. Major Assumptions Proliferation concentrates in the proximity of the tumor borders where oxygen levels are higher, the pH is more normal, and the pressure is lower (1214). To model this, we assume the probability of a cell to divide to be linear, from 1 at the tumor edges to 0 at a distance λ = 600 μm (60 cells) from the tumor margins. Metabolism $∂c∂t=Dc∇2c−κn$ (A) Oxygen is critical for cell survival, it diffuses into the tumor and it is consumed by cancer cells. Upon discretization of Eq. (A), cancer cells n consume oxygen at rate κ. Major Assumptions Oxygen around the tumor is kept constant by the vascular system. Cell quiescence occurs below an oxygen threshold κs and necrosis below an oxygen threshold θ. Migration $∂c∂t=Dn∇2n$ (B) During tumor progression, cancer cells lose cell-to-cell attachment and invade the surrounding tissues (15). This process is modeled using the Hybrid Discrete-Continuum Technique (16), in which cancer cells n disperse with coefficient Dn. To simulate adhesion, we allow cell movement if the number of neighboring tumor cells gα, where α is the adhesion coefficient of the cell in the range 0 to 4. Major Assumptions Cells move according to a random motion coefficient Dn and an adhesion coefficient α. CSCs CSCs divide symmetrically with probability PS and asymmetrically with probability 1 − PS. Two new CSCs result from the former, a differentiated cancer cell (DCC) and a CSC from the latter. CSCs possess unlimited replicative potential and self-renewal whereas DCCs can divide up to H times. A CSC growth model has small PS values, whereas for PS = 1, we simulate the classical model of malignancies. Major Assumptions We fix H = 5 and vary PS to simulate different CSC frequencies. As suggested experimentally for both CSCs and normal stem cells, we have restricted migration to CSCs (1720). Tumor Evolution We introduced tumor phenotypical heterogeneity in some experiments by assuming that every CSC self-renewal division has a chance PMut for the new CSC to acquire a different phenotype (see Supplementary Materials and Methods). Major Assumptions Only CSCs contribute to tumor evolution in the long run. Malignancies arise after sequential accumulations of mutations in oncogenes and tumor suppressor genes (1). During the process of malignant transition, fundamental regulatory mechanisms are lost. The cancerous cell population has unlimited growth potential, evades the immune system and apoptosis, and often acquires the ability to breach tissue boundaries and expand into foreign environments (1). The cancer stem cell (CSC) concept sheds new light on all of these features (2, 3). In this study, we investigate the influence of a hierarchical organization of malignant clones on tumor growth, evolution, invasion, and morphology using a multi-scale cellular automaton–based computer model (4). ### CSCs Malignancies are highly heterogeneous tissues containing largely diverse cancer cell populations as well as other cells such as fibroblasts and macrophages (1). The dominant genetic view of malignancies explains the heterogeneity in cancer cells with the presence of genetically diverse clones emerging from the continuous acquiring of additional genetic lesions by cancer cells. Such clones compete for resources, resulting in a highly dynamic process known as Tumor Darwinism (5). In this article, we refer to this view of malignancies as the classical model. Although this model greatly contributes to our understanding of malignancies, recent experimental evidence suggests an additional layer of complexity. The heterogeneity present in tumors could, in part, be the result of the diversity in differentiation grade of genetically identical cells (3, 6). For example, in glioblastoma multiforme, cells with an immature phenotype expressing the cell surface marker AC133 are the cells that fuel tumor growth and have the exclusive capacities to self-renew, differentiate, and transplant the malignancy into severe combined immunodeficiency mice (7). Self-renewal and spin-off of differentiated cells are features shared with normal stem cells, and therefore, cells with such features in malignancies are defined as CSCs (6, 8). ### Modeling Tumor Growth Tumor growth is generally accepted to be the result of several highly complex interacting processes. Fundamental cellular characteristics such as genetic and epigenetic features influence signal transduction route activities that in turn control cellular functions such as mitosis, apoptosis, and cell migration. In addition, environmental factors including nutrients and growth factor concentrations interplay with these processes. To study the emergent properties of such systems regarding proliferation speed, infiltrative growth, and phenotypical evolution of cancer, advanced mathematical models have been developed (911). Here, we apply computational modeling techniques to investigate the consequences of hierarchically organized cancer cell populations on solid tumor growth dynamics and progression. We describe that implementing the developing concept of CSCs in a mathematical tumor growth model directly results in an invasive morphology. Moreover, we found that hierarchical organized malignant clones have highly altered evolutionary dynamics. Most strikingly, the CSC organization promotes phenotypical heterogeneity, a feature that could have immediate consequences for therapy resistance. ### SCA Model We developed a hybrid tumor growth model based on cellular automata (4) and partial differential equations. We refer to this model as the SCA model. In the SCA model, the individual cancer cell is the fundamental unit of the tumor, we simulate its proliferation, metabolism, migration, stemness, and differentiation (see Quick Guide and Supplementary Materials and Methods). Implementing the CSC model of malignancies means to simulate cancer cells with different replicative potential within the tumor. For simplicity, we assume that in our model there are only two types of cells: CSCs and DCCs. CSCs possess unlimited replicative potential and could either generate new CSCs (with a probability PS) or DCCs. DCCs can divide for a maximum of H generations before stopping to proliferate irreversibly. This method yields a hierarchy with CSCs at the top and DCCs at the bottom. We simulate the classical model of malignancies, in which all cells possess tumor growth–promoting capacities, by simply setting PS = 1. In such a situation, all cells possess stem cell characteristics. Using this method, we have an intuitive way to compare the flat, classical tumor model with the hierarchically organized CSC model. ### Emergent Invasive Morphology Computational modeling allows the exploration of highly complex systems, such as tumor growth. In this study, we have used a computational tumor growth model to test the consequences of hierarchical organized clones on a vast range of areas of tumor biology. These include tumor growth dynamics and morphology, but also tumor evolution and therapeutic effects. The applied computational modeling technique allows us to get more insight into the underlying dynamics of these aspects of cancer growth and progression that would be impossible in a conventional experimental biological setting. First, we investigated how tumor growth dynamics are altered upon varying the CSC fraction. In the SCA model, this corresponds to changing the variable PS. For high values of PS, we expect to model the classical interpretation of tumors because all cell divisions are symmetrical and all cells are therefore clonogenic. In contrast, low values of PS represent the CSC model. We simulated the growth of tumors with PS = 1, PS = 0.1, and PS = 0.03 with the variables described in Table 1. Figure 1A shows the fraction of CSCs on the total amount of tumor cells for different values of PS. The selected PS values correspond to CSCs populations comprising roughly 100%, 1%, or 0.1% of the total tumor volume and therefore cover mainly the CSC fractions observed in a variety of solid malignancies (2). In the CSC model (PS = 0.1 and PS = 0.03), small (early) lesions have relatively high fractions of CSCs, although this number decreases and tends to stabilize when the tumor progresses. This observation is supported by in vivo studies that find increased numbers of CSC marker–bearing cells in micrometastases compared with larger tumors (21). This indicates that even with fixed self-renewal rates (PS), this phenomenon is intrinsic to lesions initiated by a single CSC although environmental factors influencing self-renewal frequencies are also likely to contribute. Tumor growth curves for various PS values all display the classical Gompertzian-like growth kinetics. However, as expected with equal cell cycle durations, the self-renewal rates of the stem cell fraction influences proliferation rate greatly, hence, low self-renewal rates (small PS) correspond to slow tumor growth (Fig. 1B). Interestingly, the decrease in accumulation of tumor volume is accompanied by a stabilization of the fraction of CSCs, suggesting an intimate relationship between these two processes. Figure 1. Emergent invasive behavior in the CSC model. A, different PS values result in different CSC fractions. B, growth curves for different PS values. C, quantitative measure for invasiveness shows increasing invasive behavior with declining PS. A to C, bars, SD (n = 16). D, hierarchical organization in the SCA model affects tumor morphology. Tumors for different values of self-renewal probability (PS) and different volumes. Dark blue, cells which have divided within the last 48 h (depicted larger); light blue, nondividing cells; brown, necrotic center. Right, localization of CSCs in the tumor mass. Gray, tumor mass; red, CSCs (depicted larger). In all figures, 6 × 6 mm of tissue are represented. See also Movies S1 and S2. Figure 1. Emergent invasive behavior in the CSC model. A, different PS values result in different CSC fractions. B, growth curves for different PS values. C, quantitative measure for invasiveness shows increasing invasive behavior with declining PS. A to C, bars, SD (n = 16). D, hierarchical organization in the SCA model affects tumor morphology. Tumors for different values of self-renewal probability (PS) and different volumes. Dark blue, cells which have divided within the last 48 h (depicted larger); light blue, nondividing cells; brown, necrotic center. Right, localization of CSCs in the tumor mass. Gray, tumor mass; red, CSCs (depicted larger). In all figures, 6 × 6 mm of tissue are represented. See also Movies S1 and S2. Close modal Spatially, all experiments display a three-layered structure consisting of an external proliferative area, an inner senescent layer, and a necrotic core. However, tumor morphologies for different PS values are remarkably dissimilar (Fig. 1D). For PS = 1, in which there is no hierarchy, a symmetrical, sphere-like tumor morphology is generated that closely resembles early tumor growth models (22, 23). In contrast, the shape of the tumors generated with low PS values is highly irregular (especially PS = 0.03). CSC-driven tumors yield highly invasive morphologies with fingering fronts and clusters of cancer cells beyond the tumor margin, driven by the mobility and the exclusive proliferative properties of CSCs (Movies S1 and S2). From Fig. 1C, it is evident how hierarchically organized tumors (PS = 0.03 and PS = 0.1) generate a higher degree of invasiveness (see Supplementary Materials and Methods), compared with tumors in the classical model (PS = 1). It is important to note that the intrinsic properties of the cells in the classical tumor and the stem cells in the CSC-driven tumor are completely identical. Magnification of a tumor border in a CSC-fueled tumor growth model (PS = 0.03) shows how CSCs migrate beyond the margins of the tumor mass. CSCs colonize the surrounding tissue and expand locally, forming small satellites that grow back into, and are engulfed by, the main tumor mass (Fig. 2A). This is accompanied by the observation that the invasive behavior of individual experiments that make up Fig. 1C occurs in a wave-like fashion (Supplementary Fig. S1). These results are paralleled by recent findings in a different model system in which high migration levels in a small subset of cells give rise to small proliferating extratumoral lesions and therefore tumors are “conglomerates of self-metastasis” as the authors propose (24). Figure 2. Invasive behavior in silico, in vivo, and in vitro. A, close-up of tumor border showing invasive behavior for PS = 0.03. CSCs (red) infiltrate surrounding tissue and spin-off DCCs that proliferate (dark blue). Small satellites are formed in the surrounding normal tissue (white) and grow back to and are engulfed by the main tumor mass. Nondividing cells (light blue). B, representative figures of various malignancies. All images reveal island-like formations at the rim of the main tumor mass (T). These findings are in line with tumor expansion as predicted by the SCA model that implements a CSC hierarchy. C, cell lines (n = 24) have been plated at clonal density in Matrigel. Simultaneously, the clonogenic fraction of the lines has been determined by limiting dilution analysis. A significant (P = 0.02) inverse relationship exists between clonogenicity and invasion as quantified by the measure of invasiveness we defined. Two examples of cell lines are shown and invasiveness (IM) and clonogenicity are indicated (right). See Supplementary Figs. S2 to S5 and Supplementary Table S1 for details. Figure 2. Invasive behavior in silico, in vivo, and in vitro. A, close-up of tumor border showing invasive behavior for PS = 0.03. CSCs (red) infiltrate surrounding tissue and spin-off DCCs that proliferate (dark blue). Small satellites are formed in the surrounding normal tissue (white) and grow back to and are engulfed by the main tumor mass. Nondividing cells (light blue). B, representative figures of various malignancies. All images reveal island-like formations at the rim of the main tumor mass (T). These findings are in line with tumor expansion as predicted by the SCA model that implements a CSC hierarchy. C, cell lines (n = 24) have been plated at clonal density in Matrigel. Simultaneously, the clonogenic fraction of the lines has been determined by limiting dilution analysis. A significant (P = 0.02) inverse relationship exists between clonogenicity and invasion as quantified by the measure of invasiveness we defined. Two examples of cell lines are shown and invasiveness (IM) and clonogenicity are indicated (right). See Supplementary Figs. S2 to S5 and Supplementary Table S1 for details. Close modal In an endeavor to validate these observations from our computational model, we investigated human tumor specimens. Close examination of the histology of different highly diverse human malignancies displayed a relatively confined large tumor mass with clearly detached tumor cells forming small lesions in the surrounding normal tissue (Fig. 2B). This exemplifies that human tumor histology contains indications of a stepwise infiltration and colonization of the surrounding tissue as our model predicts would follow from a hierarchical organized malignancy. Next, we attempted to determine the relationship between CSC fraction and invasive properties, as predicted by the model, using in vitro cell culture (Fig. 2C; Supplementary Figs. S2–S5 and Table S1). We determined the clonogenicity of a set of various cell lines (n = 24) as a surrogate for their CSC fraction. Additionally, we quantified the invasive properties of clonally derived structures with our measure of irregular morphology both on adherent plates and in Matrigel for all these lines. A significant inverse relationship between clonogenic fraction and the invasive properties of these lines exists. This implicates that tumor structures driven by a small fraction of clonogenic cells tend to generate a more irregular and invasive morphology, a finding that corroborates the predictions of our model. Combined, we take this as evidence that the SCA model based on the CSC concept closely resembles tumor growth patterns and morphology. ### Expansion of the Model in Three Dimensions Retains Invasive Morphology To show that our results are not limited to two dimensions only, we expand the SCA model to three dimensions (see Supplementary Materials and Methods). The three-dimensional SCA model reproduces the invasive tumor morphology induced by the hierarchical organization, yielding three-dimensional fingering tumor fronts and clusters of cancer cells beyond the borders of the main tumor mass (Fig. 3A; Movie S3). On the contrary, again a classical model with equal volume does not display any apparent invasive pattern and instead exhibits a spherical morphology (Fig. 3B; Movie S4). We therefore propose that the CSC model, in contrast to the flat model, more faithfully reproduces the human tumor morphology for both the two-dimensional and the three-dimensional implementations. Figure 3. Expansion of the model in three-dimensions retains infiltrative morphology. Three-dimensional representation of CSC-driven tumor growth (A, PS = 0.03) and of the classical tumor model (B, PS = 1). A and B, 7 × 106 cells are represented. Inset, region that is enlarged in the right image. Normal tissue (black) and tumor cells (white). Cube represents 4003 cell lattice. See also Supplementary Movies S3 and S4. Figure 3. Expansion of the model in three-dimensions retains infiltrative morphology. Three-dimensional representation of CSC-driven tumor growth (A, PS = 0.03) and of the classical tumor model (B, PS = 1). A and B, 7 × 106 cells are represented. Inset, region that is enlarged in the right image. Normal tissue (black) and tumor cells (white). Cube represents 4003 cell lattice. See also Supplementary Movies S3 and S4. Close modal ### CSC Organization Stimulates Tumor Heterogeneity The dominant model evoked to explain the advancement of malignancies is clonal evolution, as was first proposed by Nowell (5). The term effective population size is used in population genetics to indicate the fraction of total individuals in a population that effectively contributes to the next generation and are therefore evolutionarily relevant. Hence, the effective population size in a CSC-driven malignancy is smaller than in the classical model because only mutations in the CSC compartment contribute to the evolutionary process (2). To implement clonal evolution in our model, we assume that, at each symmetrical division a CSC has a probability PMut to acquire a genetic hit and generate a daughter cell with a different phenotype selected from a randomly generated pool of 30 phenotypes (Table S2 and S3; see Supplementary Materials and Methods). Under an equal mutation rate (PMut = 0.1), the CSC model exhibits a slower acquisition of new phenotypes due to its smaller effective population size (Fig. 4A), but also shows a radically different selection process. Strikingly, although the rate of emergence of new phenotypes is much lower than observed in the classical model, a wide range of newly acquired phenotypes expands and contributes to the malignancy (Supplementary Fig. S6A and B). Figure 4. Tumor evolution and phenotypical selection in a cancer stem cell context. Phenotypes are randomly generated and possess traits listed in Supplementary Table S3. A, with equal mutation rates (PMut = 0.1), the cumulative amount of phenotypes that emerge is higher in the classical model (PS = 1) compared with the CSC model (PS = 0.03). B, we adapted PMut to obtain an equal rate of emergence of phenotypes. C, under equal conditions, the phenotypes in the CSC model are more diverse compared with the classical model. A representative example is shown, see also Supplementary Movies S5 and S6. A and B, bars, SD (n = 8). C, fraction of cells for each newly generated phenotype. The original phenotype “1” is ignored. For an average of eight experiments, see Supplementary Figs. S6B and S6C. Figure 4. Tumor evolution and phenotypical selection in a cancer stem cell context. Phenotypes are randomly generated and possess traits listed in Supplementary Table S3. A, with equal mutation rates (PMut = 0.1), the cumulative amount of phenotypes that emerge is higher in the classical model (PS = 1) compared with the CSC model (PS = 0.03). B, we adapted PMut to obtain an equal rate of emergence of phenotypes. C, under equal conditions, the phenotypes in the CSC model are more diverse compared with the classical model. A representative example is shown, see also Supplementary Movies S5 and S6. A and B, bars, SD (n = 8). C, fraction of cells for each newly generated phenotype. The original phenotype “1” is ignored. For an average of eight experiments, see Supplementary Figs. S6B and S6C. Close modal To compare the phenotypical selection more closely, we synchronize the pace at which new traits emerge. Evolutionarily, a CSC model with PS = 0.03 and PMut = 0.5 corresponds to a classical model with PS = 1 and PMut = 0.001 (Fig. 4B). Intriguingly, despite this adaptation of the mutation pace, the two models differ substantially in the way they exercise clonal selection. From a representative example in Fig. 4C, it is evident that the CSC model allows for a much higher phenotypical heterogeneity whereas the classical model seems to select for a small number of aggressive clones (average result in Supplementary Fig. S6C). Quantification of the heterogeneity clearly underscores this (Supplementary Fig. S7A and Supplementary Materials and Methods). It is important to realize that the selective pressure from the environment is equal in both models. This, combined with the equal rate of new phenotype occurrence, suggests that the geometric properties of the CSC model promote heterogeneity. We argue that the intrinsic properties of the CSC model might propel an alternative process to natural selection, referred to as genetic drift (25). In populations with small effective population sizes, sampling errors are frequent and might allow for the expansion of clones with no clear survival benefit. Interestingly, the invasive properties of the CSC model might fuel such a sampling error promoting mechanism. The phenomenon we observe at the tumor margins during invasion: a CSC migrating out of the tumor initiating a small satellite lesion, resembles the Founder Effect (26). The Founder Effect occurs when a new population is established by a low number of individuals from a larger population, this process is often accompanied by a loss of phenotypical variation in the new population due to sampling errors. In such a scenario, the new generation can differ substantially from the previous and is not in direct competition for nutrients and space. This implies that the pattern of tumor growth in the CSC model stimulates genetic drift at the infiltrative edges of the tumor and therefore promotes phenotypical heterogeneity. ### Dynamics of Therapeutic Interventions in Hierarchical Organized Malignancies After having established the effects of CSC-driven tumor growth on the invasion and evolution of malignancies, we now explore the crucial topic of therapeutic intervention and tumor relapse. CSCs have been suggested to be more resistant to therapeutic interventions such as chemotherapy or irradiation compared with their differentiated counterparts (27, 28). Also tumors that relapse after seemingly successful therapy are believed to regrow from the CSC that survived the therapeutic regimen (27). Here, we investigate the dynamics associated with therapeutic interventions that are either selective for non-CSCs or equally efficient against both cell types. We find that the morphology and growth kinetics of relapses for both types of therapeutic interventions are very much different (Fig. 5; see Supplementary Materials and Methods). Regrowth after therapy that specifically targets DCCs is accompanied by enhanced invasive growth patterns whereas relapsing tumors after stochastic tumor cell killing are similar to the malignancy before treatment (Fig. 5A). Simultaneously, in case CSCs are resistant to therapy, the pace at which the malignancy relapses is greatly enhanced due to the relatively high fraction of CSCs directly following therapy (Fig. 5B). Also, the invasiveness of the recurrent tumors is markedly increased following intervention which is not effective against CSCs (Fig. 5C). These findings are in line with a range of clinical observations describing increased growth speed and enhanced invasion in the relapsing malignancy that are mostly attributed to the selection of more aggressive clones by the drug (2931). However, our data now indicate that these observations could be partially explained by the failure of conventional therapies to eradicate the CSC compartment and the subsequent relapse dynamics in CSC-driven tumors. Evaluation of evolutionary dynamics during relapse after both types of intervention revealed significant differences as well. Following therapy which is ineffective against CSCs, relapsing tumors display a marked increase in heterogeneity, whereas therapy that does target CSCs results in a dramatic decrease of heterogeneity (Supplementary Fig. S8). This latter scenario is related to the fact that relapses are very much different compared with the primary malignancy with respect to the clonal lineages that contribute to the relapse of the tumor (Fig. 5D). Combined, this implicates that applying therapy that is ineffectively targeting the CSC population is not only unsuccessful in curing the patient but also promotes malignant features including rapid expansion, increased invasion, and stimulates heterogeneity directly after therapy. Figure 5. Dynamics of therapeutic intervention in hierarchical organized malignancies. A, effects of therapy in simulated tumors considering CSCs to be resistant to treatment (top) or sensitive to treatment (bottom). Left, pretreated tumor at VT = 25,000 cells. Right, after treatment and relapsed tumor for VRelapse = VT. B, growth curves of tumor relapse. Whereas therapy-sensitive CSCs recapitulate the initial growth pace, resistant CSCs yield a faster growth at relapse. C, a marked increase of invasive morphology just after treatment occurs particularly in the case of resistant CSCs. B and C, arrows, therapy (n = 8). D, phenotype distribution in the two cases of therapy response assuming no further mutation to occur after therapy (see Supplementary Fig. S8 for an average of eight experiments). Although a treatment-resistant CSC-driven tumor is able to recapitulate the clonal features of the initial malignancy by conserving the initially developed heterogeneity, for therapy-sensitive CSCs, the recurrent tumor is highly altered in terms of phenotypes. Figure 5. Dynamics of therapeutic intervention in hierarchical organized malignancies. A, effects of therapy in simulated tumors considering CSCs to be resistant to treatment (top) or sensitive to treatment (bottom). Left, pretreated tumor at VT = 25,000 cells. Right, after treatment and relapsed tumor for VRelapse = VT. B, growth curves of tumor relapse. Whereas therapy-sensitive CSCs recapitulate the initial growth pace, resistant CSCs yield a faster growth at relapse. C, a marked increase of invasive morphology just after treatment occurs particularly in the case of resistant CSCs. B and C, arrows, therapy (n = 8). D, phenotype distribution in the two cases of therapy response assuming no further mutation to occur after therapy (see Supplementary Fig. S8 for an average of eight experiments). Although a treatment-resistant CSC-driven tumor is able to recapitulate the clonal features of the initial malignancy by conserving the initially developed heterogeneity, for therapy-sensitive CSCs, the recurrent tumor is highly altered in terms of phenotypes. Close modal We have applied a computational tumor growth model to investigate the effects of hierarchical organization on a range of areas in tumor biology. The obtained results provide novel insights into the influence of CSC-driven tumor growth on tumor morphology and invasion. Importantly, we have been able to partially validate these findings in a biological experimental setting. We have also established a significant effect of hierarchical organized malignant clones on tumor evolution, and consequently, on tumor progression and relapse after therapy. Our results stress the need to develop therapeutic interventions that efficiently target the CSC compartment because drugs that fail to do so are not only unable to cure the patient—but even enhance the malignant properties of tumors. Therefore, we conclude that computational oncology has the potential to greatly enhance our understanding of fundamental cancer biology and improve future therapeutics. ### CSC-Driven Tumor Growth The capacity to display infiltrative growth is a pathognomic feature of malignant cells (1). Crossing tissue boundaries is the first step in the process of metastasis and is therefore of special interest. Surprisingly, this aspect of tumor growth has rarely been investigated in mathematical models. However, recent important work of Anderson and colleagues (11) and Bearer and colleagues (32) illuminate how selective pressure, applied by the microenvironment, in combination with tumor cell heterogeneity, could result in an invasive morphology. Here, we report how, in the SCA model, invasion emerges from the hierarchical organization of malignant clones and need not be driven by external interaction with a heterogeneous microenvironment. Hence, invasive morphology is an intrinsic feature of CSC-driven malignancies and has to be considered when studying tumor invasion. Simultaneously, high levels of clonal heterogeneity within malignancies are associated with rapid progression of the disease (33), poorer survival (34), and occurrence of therapy resistance (35). Understanding this phenomenon in more detail might improve the design of more effective treatments and chemopreventives. Our results show how CSC-driven tumor growth stimulates heterogeneity in malignancies. This is related to the reduced effective population size in these malignancies and the typical growth pattern that results in the segregation of clones by formation of small invasive lesions. These satellite structures are not in direct competition with the dominant clone, allowing suboptimal clones to expand. We speculate that interfering with these dynamics, for example by inhibiting migration, and thereby segregation, might be a potent means by which reduced heterogeneity accompanied by the slower progression of the disease and better response to therapy can be achieved either in a chemopreventive or therapeutic setting. Additionally, our results underlie the need to develop therapeutic modalities that successfully eradicate the CSC compartment. We show how therapeutic interventions that do not target CSCs efficiently are not only ineffective as a cure, but might even contribute to an increase in malignant properties of the relapsing tumor. We find, for example, enhanced heterogeneity and invasion in the relapsing tumors that are treated with interventions that are only directed against the differentiated cells. This is not due to the selection of the most malignant clones present in the cancer tissue but are intrinsic to the dynamics of regrowth in CSC-driven malignancies. ### Experimental Validation and Future Directions The model presented provides a range of predictions and implications regarding CSC-fuelled tumor growth that can be tested and exploited to investigate the properties of hierarchical organized malignancies. Here, we report that the clonogenicity of cell lines is connected with the invasive properties of these lines in vitro as suggested by our model. In future research, it would be interesting to expand this finding to human tumor specimens or in established CSC lines in which the CSC fraction can be manipulated. Additionally, our model illuminates how tumor features, such as clonal heterogeneity, are related to the CSC fraction (Supplementary Fig. S7B). Therefore, we speculated that careful analysis of these tumor properties might lead to optimized techniques that establish the CSC fraction in malignancies without the need for transplantation studies which are currently highly disputed (36, 37). The current version of the SCA model clearly shows the dynamics of CSC-driven tumor growth and its consequences for tumor morphology and evolution. However, further research efforts will undoubtedly lead to increased insight into the nature of the hierarchical organization of tumor cells. If so, the SCA model can be easily adapted to implement potential new information and subsequently come to even more accurate description of tumor growth. For instance, we have assumed several variables such as CSC self-renewal frequency, mutation rate, and migration speed to be constant. However, in in vivo tumors, these variables might be more dynamic both throughout tumor progression and as a consequence of microenvironmental interactions. For example, CSCs have been suggested to reside in a so-called “CSC niche” that is formed by secreted factors and direct cell-cell interactions (38, 39). This niche is believed to protect CSC from differentiation and to promote CSC self-renewal and therefore introduces dynamic PS values. Moreover, physical circumstances such as oxygen concentration might have differential effects on CSCs and DCCs, and therefore, might play a role in dynamic CSC functions (40). In addition, mutation rates are potentially subject to changes during tumor development (41) and intriguingly, might even be influenced by environmental conditions, such as that reported for bacteria (42). Currently, it remains elusive as to what extent this latter phenomenon, which is referred to as “adaptive mutation,” is involved in human malignancies (43). All these considerations justify further investigations, but will be dependent on thorough experimental examination of the variables involved. Importantly, from the current formulation of the SCA model, it is indisputable that hierarchical organization of malignancies significantly contributes to the invasive morphology and increased heterogeneity of tumors and is therefore a crucial issue for better understanding tumor biology and to improve current anticancer treatments. No potential conflicts of interest were disclosed. We thank M.R. Sprick, D.J. Richel, and F. de Sousa Mello for useful comments and R. Belleman and J.H. de Jong for technical assistance. Grant Support: Academic Medical Center (J.J.C. Verhoeff, J.P. Medema, and L. Vermeulen), ZonMW VICI program (J.P. Medema), and the Faculty of Science of the University of Amsterdam (A. Sottoriva, L. Naumov, and P.M.A. Sloot). The costs of publication of this article were defrayed in part by the payment of page charges. This article must therefore be hereby marked advertisement in accordance with 18 U.S.C. Section 1734 solely to indicate this fact. 1 Hanahan D , Weinberg RA . The hallmarks of cancer . Cell 2000 ; 100 : 57 70 . 2 Vermeulen L , Sprick MR , Kemper K , Stassi G , Medema JP . Cancer stem cells—old concepts, new insights . Cell Death Differ 2008 ; 15 : 947 58 . 3 Reya T , Morrison SJ , Clarke MF , Weissman IL . Stem cells, cancer, and cancer stem cells . Nature 2001 ; 414 : 105 11 . 4 Sloot PMA , Hoekstra AG . Modeling dynamic systems with cellular automata . In: Fishwick PA , editor. Handbook of dynamic system modeling . Chapman & Hall/CRC ; 2007 , p. 1 20 . 5 Nowell PC . The clonal evolution of tumor cell populations . Science 1976 ; 194 : 23 8 . 6 Vermeulen L , Todaro M , de Sousa Mello F , et al . Single-cell cloning of colon cancer stem cells reveals a multi-lineage differentiation capacity . Proc Natl Acad Sci U S A 2008 ; 105 : 13427 32 . 7 Singh SK , Hawkins C , Clarke ID , et al . Identification of human brain tumour initiating cells . Nature 2004 ; 432 : 396 401 . 8 Clarke MF , Dick JE , Dirks PB , et al . Cancer stem cells—perspectives on current status and future directions: AACR Workshop on Cancer Stem Cells . Cancer Res 2006 ; 66 : 9339 44 . 9 Araujo RP , McElwain DL . A history of the study of solid tumour growth: the contribution of mathematical modelling . Bull Math Biol 2004 ; 66 : 1039 91 . 10 Anderson AR , Quaranta V . Integrative mathematical oncology . Nat Rev Cancer 2008 ; 8 : 227 34 . 11 Anderson AR , Weaver AM , Cummings PT , Quaranta V . Tumor morphology and phenotypic evolution driven by selective pressure from the microenvironment . Cell 2006 ; 127 : 905 15 . 12 Sutherland RM . Cell and environment interactions in tumor microregions: the multicell spheroid model . Science 1988 ; 240 : 177 84 . 13 Folkman J , Hochberg M . Self-regulation of growth in three dimensions . J Exp Med 1973 ; 138 : 745 53 . 14 Vaupel P , Hockel M . Blood supply, oxygenation status and metabolic micromilieu of breast cancers: characterization and therapeutic relevance . Int J Oncol 2000 ; 17 : 869 79 . 15 Zaman MH , Trapani LM , Sieminski AL , et al . Migration of tumor cells in 3D matrices is governed by matrix stiffness along with cell-matrix adhesion and proteolysis . Proc Natl Acad Sci U S A 2006 ; 103 : 10889 94 . 16 Anderson AR . A hybrid mathematical model of solid tumour invasion: the importance of cell adhesion . Math Med Biol 2005 ; 22 : 163 86 . 17 Brabletz T , Jung A , S , Hlubek F , Kirchner T . Opinion: migrating cancer stem cells—an integrated concept of malignant tumour progression . Nat Rev Cancer 2005 ; 5 : 744 9 . 18 Hermann PC , Huber SL , Herrler T , et al . Distinct populations of cancer stem cells determine tumor growth and metastatic activity in human pancreatic cancer . Cell Stem Cell 2007 ; 1 : 313 23 . 19 Laird DJ , von Andrian UH , Wagers AJ . Stem cell trafficking in tissue development, growth, and disease . Cell 2008 ; 132 : 612 30 . 20 Sheridan C , Kishimoto H , Fuchs RK , et al . CD44+/CD24− breast cancer cells exhibit enhanced invasive properties: an early step necessary for metastasis . Breast Cancer Res 2006 ; 8 : R59 . 21 Balic M , Lin H , Young L , et al . Most early disseminated cancer cells detected in bone marrow of breast cancer patients have a putative breast cancer stem cell phenotype . Clin Cancer Res 2006 ; 12 : 5615 21 . 22 Dormann S , Deutsch A . Modeling of self-organized avascular tumor growth with a hybrid cellular automaton . In Silico Biol 2002 ; 2 : 393 406 . 23 Jiang Y , Pjesivac-Grbovic J , Cantrell C , Freyer JP . A multiscale model for avascular tumor growth . Biophys J 2005 ; 89 : 3884 94 . 24 Enderling H , Hlatky L , Hahnfeldt P . Migration rules: tumours are conglomerates of self-metastases . Br J Cancer 2009 ; 100 : 1917 25 . 25 Merlo LM , Pepper JW , Reid BJ , Maley CC . Cancer as an evolutionary and ecological process . Nat Rev Cancer 2006 ; 6 : 924 35 . 26 Mayr E . Change of genetic environment and evolution . In: Huxley J , Hardy AC , Ford EB , editors. Evolution as a process . Princeton : Princeton University Press ; 1954 , p. 157 80 . 27 Jordan CT , Guzman ML , Noble M . Cancer stem cells . N Engl J Med 2006 ; 355 : 1253 61 . 28 Rich JN . Cancer stem cells in radiation resistance . Cancer Res 2007 ; 67 : 8980 4 . 29 Huff CA , Matsui W , Smith BD , Jones RJ . The paradox of response and survival in cancer therapeutics . Blood 2006 ; 107 : 431 4 . 30 Spiegl-Kreinecker S , Pirker C , Marosi C , et al . Dynamics of chemosensitivity and chromosomal instability in recurrent glioblastoma . Br J Cancer 2007 ; 96 : 960 9 . 31 El Sharouni SY , Kal HB , Battermann JJ . Accelerated regrowth of non-small-cell lung tumours after induction chemotherapy . Br J Cancer 2003 ; 89 : 2184 9 . 32 Bearer EL , Lowengrub JS , Frieboes HB , et al . Multiparameter computational modeling of tumor invasion . Cancer Res 2009 ; 69 : 4493 501 . 33 Maley CC , Galipeau PC , Finley JC , et al . Genetic clonal diversity predicts progression to esophageal adenocarcinoma . Nat Genet 2006 ; 38 : 468 73 . 34 Flyger HL , Larsen JK , Nielsen HJ , Christensen IJ . DNA ploidy in colorectal cancer, heterogeneity within and between tumors and relation to survival . Cytometry 1999 ; 38 : 293 300 . 35 Iwasa Y , Nowak MA , Michor F . Evolution of resistance during clonal expansion . Genetics 2006 ; 172 : 2557 66 . 36 Hill RP . Identifying cancer stem cells in solid tumors: case not proven [discussion 0] . Cancer Res 2006 ; 66 : 1891 5 . 37 Quintana E , Shackleton M , Sabel MS , Fullen DR , Johnson TM , Morrison SJ . Efficient tumour formation by single human melanoma cells . Nature 2008 ; 456 : 593 8 . 38 Gilbertson RJ , Rich JN . Making a tumour's bed: glioblastoma stem cells and the vascular niche . Nat Rev Cancer 2007 ; 7 : 733 6 . 39 Borovski T , Verhoeff JJ , ten Cate R , et al . Tumor microvasculature supports proliferation and expansion of glioma-propagating cells . Int J Cancer 2009 ; 125 : 1222 30 . 40 Li Z , Bao S , Wu Q , et al . Hypoxia-inducible factors regulate tumorigenic capacity of glioma stem cells . Cancer Cell 2009 ; 15 : 501 13 . 41 Bielas JH , Loeb KR , Rubin BP , True LD , Loeb LA . Human cancers express a mutator phenotype . Proc Natl Acad Sci U S A 2006 ; 103 : 18238 42 . 42 Hastings PJ , Bull HJ , Klump JR , Rosenberg SM . Adaptive amplification: an inducible chromosomal instability mechanism . Cell 2000 ; 103 : 723 31 . 43 Rosenberg SM . . Nat Rev Genet 2001 ; 2 : 504 15 . 44 Tomita K , Plager JE . In vivo cell cycle synchronization of the murine sarcoma 180 tumor following alternating periods of hydroxyurea blockade and release . Cancer Res 1979 ; 39 : 4407 11 . 45 Sherwood L . Human physiology: from cells to systems . Belmont : ; 1997 . 46 Casciari JJ , Sotirchos SV , Sutherland RM . Variations in tumor cell growth rates and metabolism with oxygen concentration, glucose concentration, and extracellular pH . J Cell Physiol 1992 ; 151 : 386 94 . 47 Freyer JP , Tustanoff E , Franko AJ , Sutherland RM . In situ oxygen consumption rates of cells in V-79 multicellular spheroids during growth . J Cell Physiol 1984 ; 118 : 53 61 . 48 Bray D . Cell movements . New York : Garland Pub. ; 1992 .
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 2, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6957565546035767, "perplexity": 4471.36305199577}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335326.48/warc/CC-MAIN-20220929065206-20220929095206-00284.warc.gz"}
http://mathoverflow.net/questions/101422/which-information-can-be-obtained-from-poincar%c3%a9-series
# Which information can be obtained from Poincaré series ? If $A= \bigoplus_{i\ge 0}A_i$ is a graded commutative Noetherian algebra over a field, its Poincaré series is given by $P(t) = \sum_{i\ge 0} \dim(A_i)t^i$. Although the definition of $P(t)$ only depends on the graded vector space underlying $A$, the Krull dimension of the ring $A$ can be obtained from the Poincaré series as the order of the pole of $P(t)$ at $t=1$. Question: Are there other information about the ring structure of $A$ that can be obtained from $P(t)$ ? Since I'm particularly interested in cases where $A$ is the cohomology ring of a finite group, I'm also looking for an example of finite groups whose mod-p cohomology rings are not isomorphic but have identical Poincaré series. - I don't know if it's possible to obtain further information on the ring structure in general. However, if $A$ is the mod-p cohomology ring of a finite group, a result of Benson and Carlson states that if $A$ is Cohen-Macaulay, then $P(t)$ satisfies the functional equation $$P(1/t) = (-1)^d P(t)\hspace{90pt}(\ast)$$ where $d$ is the Krull dimension of $A$. Conversely, if a given $P(t)$ doesn't satisfy this equation, you know that $A$ isn't Cohen-Macaulay, i.e. the depth of $A$ is less than its Krull dimension. In case $A$ has Krull dimension 2, than $A$ is Cohen-Macaulay iff $P(t)$ satisfies $(\ast)$. These results can be found in the paper Benson, Carlson: Functional equations for Poincaré series in group cohomology. Bull. London Math. Society 26(1994), 438-448. As an example for groups with identical Poincaré series you can take $E := \mathbb{Z}/2 \times \mathbb{Z}/2$ and $D_8$, the dihedral group of order 8. Their cohomology rings $$H^\ast(E;\mathbb{F}_2)=\mathbb{F}_2[x,y],\;\; |x|=|y|=1$$ $$H^\ast(D_8;\mathbb{F}_2)=\mathbb{F}_2[x,y,z]/(xy),\;\; |x|=|y|=1, |z|=2$$ aren't isomorphic (since only one is a domain) but both have Poincaré series $P(t)=\frac{1}{(1-t)^2}$. Added: There is a paper of R. Stanley (who is also active on MO) that contains some properties of the Poincaré series that may be of interest. I just quote a few: • If $A$ is Gorenstein of Krull dimension $d$, then $P(1/t)=(-1)^dt^aP(t)$ for some integer $a$. This generalizes $(\ast)$ (with a=0) because the mod-p cohomology ring of a finite group is Cohen-Macaulay iff its Gorenstein. • $A$ is a complete intersection with generators of degree 1 iff $P(t)$ has the form $$P(t)=\frac{\prod_{i=1}^l(1+t+\cdots+t^{m_i})}{(1-t)^d}$$ • A sequence of homogeneous elements $x_1,...,x_k \in A$ of positive degrees $n_i$ is regular iff $P(A,t) = P(B,t)/\prod_i (1-t^{n_i})$ where $B=A/(x_1,...,x_n)$. - In general, I expect there to be statements of the form "nice rings have nice Poincar\'e series", so in the direction of your title, "nasty Poincar\'e series implies nasty ring". –  Allen Knutson Jul 6 '12 at 17:05 @Allen: What do you mean by "nice rings" resp. "nice Poincaré series" ? Can you give me some more details on the statements you expect to be there. References are also welcome. Thanks. –  tj_ Jul 6 '12 at 21:57
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9156858325004578, "perplexity": 125.82444838668788}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1410657132883.65/warc/CC-MAIN-20140914011212-00190-ip-10-196-40-205.us-west-1.compute.internal.warc.gz"}
http://math.stackexchange.com/questions/70343/vector-multi-variable-calculus-potential-in-non-conservative-fields
# (vector/multi variable) calculus, potential in non conservative fields I know that finding a potential is a sufficient condition to show that a vector field is conservative. My question is if the those statements are equivalent. I've found a vector field which isn't conservative, does this imply that there is no potential to the vector field? kind reg, - If a implies b, then not-b implies not-a. You don't need a and b to be equivalent. –  Chris Eagle Oct 6 '11 at 12:53 Having a potential function and being conservative are equivalent (under some mild assumptions). Specifically, if a (continuous) vector field is conservative on an open connected region then it has a potential function. And "Yes" if a vector field fails to be conservative, it cannot have a potential function. Here are some notes I posted for one of my classes a few years ago... http://mathsci2.appstate.edu/~cookwj/courses/math2130-fall2009/math2130-Line_Int_notes.pdf A few notes: 1) I didn't list all assumptions everywhere (for example, I wasn't careful to say that I'm assuming things are continuous where needed). 2) In the notes a vector field which possesses a potential function is called a "gradient" vector field. 3) The relevant theorem is on page 5. -
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.837069034576416, "perplexity": 526.0239420079362}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115855094.38/warc/CC-MAIN-20150124161055-00054-ip-10-180-212-252.ec2.internal.warc.gz"}
https://lavelle.chem.ucla.edu/forum/viewtopic.php?t=57034
## Adding Heat and Constant Temperature Dina Marchenko 2J Posts: 54 Joined: Thu Jul 25, 2019 12:16 am ### Adding Heat and Constant Temperature What does it mean that when you add heat to a system, its overall temperature remains constant? Isn't ice colder than boiling water (ie they have different temperatures)? Or does water as a solid have a uniform temperature, then liquid water has its own uniform temperature, then water vapor has its own uniform temperature? Julie_Reyes1B Posts: 105 Joined: Sat Jul 20, 2019 12:16 am ### Re: Adding Heat and Constant Temperature I think what Dr. Lavelle was trying to show is that when a phase change occurs, it is possible no temperature change occurs. For example, in labs they add heat really slowly to melt ice to liquid water. The actual act of breaking hydrogen bonds to melt the ice requires energy in the form of heat. The ice will be 0° C, and when it melts to water, there will be a point when the water will also be 0°C. In this case the system is taking in heat while maintaining a constant temperature. Caitlyn Tran 2E Posts: 100 Joined: Fri Aug 09, 2019 12:15 am ### Re: Adding Heat and Constant Temperature Adding on to previous responses, the temperature is measuring a much larger outside system compared to the small system your reaction may be taking place in. Hence, any changes in temperature of the small system may not heavily impact the environment, which may explain why the temperature of the outside remains unchanged. Also, in labs, you may conduct a reaction in a test tube placed in a reservoir. The reservoir counteracts any changes in temperature, so the outside temperature remains the same. However, you can calculate the heat given off or absorbed by the reaction by measuring the amount of energy the reservoir needed to put in to maintain a constant temperature. Hope this helps! Return to “Applying Le Chatelier's Principle to Changes in Chemical & Physical Conditions” ### Who is online Users browsing this forum: No registered users and 2 guests
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8756764531135559, "perplexity": 1036.9950017338704}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141176256.21/warc/CC-MAIN-20201124111924-20201124141924-00413.warc.gz"}
https://seop.illc.uva.nl/entries/determinate-determinables/
# Determinables and Determinates First published Tue Feb 7, 2017 Determinables and determinates are in the first instance type-level properties that stand in a distinctive specification relation: the ‘determinable-determinate’ relation (for short, ‘determination’). For example, color is a determinable having red, blue, and other specific shades of color as determinates; shape is a determinable having rectangular, oval, and other specific (including many irregular) shapes as determinates; mass is a determinable having specific mass values as determinates. Reflecting that determinables admit of different degrees or levels of specification, characterization of a property as determinable or determinate is typically relative; for example, red is a determinate of color and a determinable of scarlet. The determination relation appears to differ from other specification relations. In contrast with the genus-species and conjunct-conjunction relations, where the more specific property can be understood as a conjunction of the less specific property and some independent property or properties, a determinate is not naturally treated in conjunctive terms (red is not a conjunctive property having color and some other property or properties as conjuncts); and in contrast with the disjunction-disjunct relation, where disjuncts may be dissimilar and compatible (as with red or round), determinates of a determinable (at a given level of specificity) are both similar and incompatible (red and blue are similar in both being colors; nothing can be simultaneously and uniformly both red and blue). Supporting this latter contrast is that our thoughts about determinables do not appear to be about (potentially infinite) disjunctions of determinates, and our perceptions of properties such as color and shape do not appear to be of maximal determinates, as would be the case if determinables were disjunctions of determinates (since every instance of a disjunction is an instance of a disjunct). Of course, appearances can be deceiving, and there are strategies for treating determinables and their mode of specification in deflationary terms. Whether these strategies succeed or fail is no small issue. For the seemingly distinctive character of determinables and determination may enter into the best case for the claim that there are genuine features of reality that are less than maximally specific—a claim which, if true, has profound implications for a wide range of philosophical issues. What follows is organized into five sections. In §1, we trace key points of the historical trajectory whereby determination came to be seen as a distinctive form of specification. In §2, we present the commonly advanced features of determinables, determinates, and their relation; here also we discuss extensions of the (realistically construed) relation to ontological categories beyond properties. In §3, we sketch the main anti-realist, reductionist, and non-reductionist accounts of determinables and determinates, saying how these aim to accommodate the seeming features of determinables and determinates, and flagging key concerns and responses; here also we discuss certain accounts of the structure or logic of determination. In §4 we consider reasons for and against thinking that determinates are metaphysically prior to determinables. In §5 we present three applications of these notions, on which the distinctive features of determinables and/or determinates are seen as key to understanding physical laws of nature, mental/higher-level causation, and metaphysical indeterminacy. ## 1. Historical treatments The following historical trajectory motivates the characteristic features and certain treatments of determinables and determinates; those interested in cutting to the contemporary chase can skip to §2 with little loss of continuity. The terminology of determinables and determinates stems from ancient and scholastic treatment of the definition of a species, as analyzable into two parts: a genus, or generic part of an essence—the pars determinabilis essentiae, and the differentia, or differentiating part of an essence—the pars determinans essentiae. As we’ll see, the terminology and its original application eventually went separate ways, with the genus-species relation standardly being treated in conjunctive terms (with a given species being identified with the conjunction of a genus and one or more differentia), and determination being treated as a distinctive non-conjunctive specification relation.[1] ### 1.1 Aristotle Some recent commentators have suggested that the notion of determination has a precursor in ancient times. For example, one commentator has proposed that Aristotle’s preferred treatment of the relation between kinds and forms (a.k.a. genus and species) involves key features of determination (Granger 1984). Granger suggests that Aristotle’s conception of a form/species evolves from a conjunctive conception (as in Topics VI.3. 140a27–29) to one on which differentia entail their associated kind/genus rather than conjunctively supplementing it. Granger motivates the non-conjunctive conception on grounds that “Aristotle clearly maintains that [...] the genus does not exist apart from its species, or exists merely as matter” (1984: 15); in turn, he takes this conception to indicate that genus and differentia are best viewed in terms of determinable and determinate. […] A determinate […] is not a conjunction of its determinable and something else logically distinct from its determinable. (1984: 19) Salmieri (2008) offers further support for this interpretation, citing another important feature of determination—namely, that determinates differ specifically in respect of their determinable. As Salmieri puts it, Certain things are not merely other than one another—as blue is other to both loud and inflation—but are different from or “opponent to” one another, as blue is different from red. (2008: 78) And he sees this distinction as operative in Aristotle’s (Metaphysics I.3-4) remarks: But difference (διαφορὰ) and otherness (ἐτερότης) are distinct (ἄλλο). For, while the other and that which it is other than are not necessarily other in something (for everything that is some being is either other or the same), the different differs from something in something, so that it’s necessary for there to be something the same in which they differ. This same thing is a kind or form; for every different thing differs either {i} in kind or {ii} in form [...]. (1054b22–31 cf. 1018a9–15; translation quoted from Salmieri 2008, 79) The previous features—non-conjunctive entailment of the less by the more specific, the specification’s being somehow “in respect of” the less specific, and the more specific features’ being opponent to one other—are indeed characteristic features of determination, contrasting with the conjunct-conjunction and disjunction-disjunct relations (modulo deflationary strategies; we won’t always carry this qualification forward). As such, there is at least a case here for Salmieri’s claim that “we ought to recognise Aristotle as the originator of this distinction” (2008: 79). ### 1.2 The Moderns Throughout the scholastic and modern periods the genus-species and genus-differentia relations are often treated in what Prior (1949) calls “a confused blending” of the conjunct-conjunction and determination relations. In the modern period motivations for distinguishing the latter relations become increasingly salient. One motivation stems from Leibniz’s observation that a conjunctive treatment of the genus-species relation does not fix a direction of specification, since one can as easily take the differentia to be the genus, as vice versa (see Nouveaux Essais, III, iii, 10, Langley’s translation, 313). Johnson (1924) interprets Descartes and Spinoza as attempting to ensure asymmetry by means of attributes and modes (see Descartes, Principles of Philosophy I.53; Spinoza, Ethics 1p8s1 and Ep. 50), which Johnson suggests are “almost equivalent” to the notions of determinable and determinate: What I call a determinable is almost equivalent to what they call an attribute, and my determinate almost equivalent to their mode of an attribute. (1924: III, v. 1) Prior (1949) similarly sees Spinoza as aiming to ensure asymmetry by means of something like determination: If Leibniz brought consistency into the medieval view of the distinctness of genus and differentia by making the relation between them symmetrical, Spinoza brought consistency into the medieval view of the asymmetry of the relation between them by denying their distinctness. Spinoza’s “modes” are not new qualities added to his “attributes”, or qualities formed by the addition of something else to them, but determinations of them. (1949: 7) Prior also sees Locke as relying on the distinctive incompatibility of same-level determinates as making sense of why “any subject may have of each sort of primary qualities but one particular at once”, and similarly for “all sensible ideas peculiar to each sense” (Locke, Essay, IV iii 5). In the work of Locke and other empiricists, we also see precursors of certain contemporary anti-realist or reductivist accounts on which determinable attributions either indeterminately refer to determinate properties (as Berkeley [1710] suggests, more generally, for ‘abstract’ terms), or on which determinable attributions are (merely) ways of classifying determinates. Prior’s description of Locke’s view might be seen as compatible with either approach: Though we speak of an object as colored, what is there being referred to is not really a quality of the object but a quality of one of its qualities; what is meant is that one of its qualities belongs to the class “colors”. (Prior 1949: 8–9) ### 1.3 Johnson The terminology of ‘determinables’ and ‘determinates’ as indicative of a distinctive relation is first found in Volume I, Chapter XI—‘The Determinable’—of W. E. Johnson’s Logic (1921): I propose to call such terms as color and shape determinables in relation to such terms as red and circular which will be called determinates; and, in introducing this new terminology, to examine the distinction between the relation of red to color and the relation of Plato to man …. (1921: I, xi, 1) Johnson’s discussion is also the first systematic attempt to characterize the key features of determinables, determinates, and their relation (see esp. 1921: 173–185), though as below, he is not always clear about whether these features apply to properties or rather to predicates. See Prior (1949) and Poli (2004) for detailed discussion of Johnson’s views and his influence on later authors. Some of these features are familiar, including that determination is a specification relation: To predicate color or shape of an object obviously characterises it less determinately than to predicate of it red or circular; (Johnson 1921: I, xi, 1) that this specification is non-conjunctive: there is a genuine difference between that process of increased determination which conjunctivally introduces foreign adjectives, and that other process by which without increasing, so to speak, the number of adjectives, we define them more determinately;[2] (Johnson 1921: I, xi, 3) that same-level determinates are incompatible: if any determinate adjective characterises a given substantive, then it is impossible that any other determinate under the same determinable should characterise the same substantive; (Johnson 1921: I, xi, 4) and that same-level determinates are moreover “opponent”: the unique and peculiar kind of difference that subsists between the several determinates under the same determinable […] does not subsist between any one of them and an adjective under some other determinable. (Johnson 1921: I, xi, 1) Johnson moreover develops these features, suggesting, for example, that opponent differences often reflect that determinates are comparable: many determinates under the same determinable [are such] that the differences between different pairs of determinates can be compared with one another […]; e.g. the difference between red and yellow is greater than that between red and orange. (Johnson 1921: I, xi, 4) Attention to determinate comparability leads Johnson to suggest that determinables may be determined along multiple “dimensions”: [A] color may vary according to its hue, brightness and saturation; so that the precise determination of a color requires us to define three variables which are more or less independent of one another in their capacity of co-variation; but in one important sense they are not independent of one another, since they could not be manifested in separation. The determinate color is therefore single, though complex …. (1921: I, xi, 4) That determinables may have multiple determination dimensions is an important observation, which plays a foundational role in Funkhouser’s (2006) model of determination (see §3.6.1). Two other aspects of Johnson’s discussion are worth noting. First, he denies that determinables are in any sense shared by determinates: “the ground for grouping determinates under one and the same determinable is not any partial agreement between them” but rather “the special kind of difference” (1921: I, xi, 1) distinguishing opposing determinates.[3] As we’ll see, this is a choice point for contemporary accounts. Second, Johnson assumes that while limits in our perceptual or instrumental capabilities require that we characterize objects in determinable terms, it is nonetheless a universally adopted postulate that the characters of things which we can only characterise more or less indeterminately, are, in actual fact, absolutely determinate. (1921: I, xi, 5) The assumption that determinable characterization reflects (mere) epistemic, perceptual, or representational limits remains common, and pushes towards giving one or other deflationary account of determinables. This is another choice point: on some contemporary accounts determinables are both real and irreducible to determinates. ### 1.4 Contemporary additions and refinements In this sub-section, we discuss certain comparatively recent contributions which influentially shaped the present understanding of determination. Prior’s (1949) two-part discussion of determinables and determinates is historically rich and philosophically subtle. Interestingly, he suggests that the difference between determinates is brute, admitting no explanation: we can say that the red and the blue agree in being colored, but of their difference we can only say either that their color is different, or that one is red and the other blue. (1949: 5–6) Some have seen this sort of statement as broadly contradictory, in suggesting that what makes determinates similar (e.g., color) is also that in virtue of which they are different (“of their difference we can only say … that their color is different”); see §3.2. Prior also considers and rejects Johnson’s assumption that determinates are similar only in respect to their unique form of difference, rather suggesting (c.f. Salmieri’s discussion of Aristotle, above) that determinates are similar “in respect of” that determinable: [I]t seems to me that determinates under the same determinable have a class unity […]. Determinates under the same determinable have the common relational property […] of characterising whatever they do characterise in a certain respect. Redness, blueness, etc., all characterise objects, as we say, “in respect of their color”; triangularity, squareness, etc., “in respect of their shape”. And this is surely quite fundamental to the notion of being a determinate under a determinable. (1949: 13) That determinates characterize objects in respect of their determinables is indeed a key feature of determination, though there remains disagreement about how best to metaphysically articulate this feature (e.g., as involving shared determinables or not). Prior takes this feature to push against deflationary approaches on which determinables exist, at best, merely to classify the having of determinate properties. On the contrary, he maintains, if determinates are unified by characterizing objects in a certain respect, then the “respects in which objects are to be characterised”, to which determinable adjectives refer, are related to the objects not less but more intimately than the determinate qualities which “characterise” them. (1949: 13) Moving on, Searle (1959) usefully highlights that determination appears to contrast not just with the conjunct-conjunction relation, but also with the disjunction-disjunct relation: [C]learly not any two terms which stand in the relation of greater to less specific ipso stand in the relation of determinate to determinable: “yellow” is in some sense more specific than “yellow or angry” but it is not a determinate of “yellow or angry” in the sense in which it is a determinate of “color”. (1959: 141–2) Searle also offers an explicit criterion of determination: For any two terms $$A$$ and $$B$$, $$A$$ is a determinate of $$B$$ if and only if $$A$$ is a non-conjunctive specifier of $$B$$, and $$A$$ is logically related to all other non-conjunctive specifiers of $$B$$ [where] any two terms are logically [i.e., conceptually] related if either entails the other or either entails the negation of the other. (1959: 148) Interestingly, this criterion doesn’t rule out that $$B$$ is a disjunction of its determinates (‘non-conjunctive specifiers’), so long as the determinates stand in the requisite conceptual entailment relations. Note also that Searle moves fluidly between talk of determinable properties and talk of determinable ‘terms’; until recently, this sort of slide was common. Searle notes, as well, that the structure of determinables and determinates in the same family may give rise to “levels” of determinates ordered by degree of specificity; as he says, “we would like some way of showing that e.g., “red” and “yellow” are on the same level as determinates of “color” whereas “scarlet” is on a different and lower level” (1959: 145). Here too he offers a criterion: Two terms $$A$$ and $$B$$ are same level determinates of $$C$$ if and only if they are both determinates of $$C$$ and neither is a specifier of the other. (1959: 149) Next, Armstrong's investigations into the metaphysics of properties (in, e.g., his 1978a,b, 1989, 1997, and 2010) are notable, first, in treating determination as an explicitly metaphysical relation, and second, in raising to salience the important question of whether determinables are ontologically reducible to determinates, asked against the commonly accepted background assumption that reality itself is maximally determinate. As we’ll see, Armstrong’s own position on this score swings from reductionism (1987) to non-reductionism (1997), and back again (2010). Finally, Yablo’s (1992) is notable in offering a modal account of determination as a relation of asymmetric dependence, and in extending determination to particulars (events) as well as properties. Since asymmetrical necessitation now serves as a common minimal characterization of determination, it’s worth seeing how Yablo motivates this approach. He starts by observing a parallel between intuitive characterizations of identity and of determination, as holding between properties: $$P$$ is identical to $$Q$$ iff: for a thing to be $$P$$ is for it to be $$Q$$. $$P$$ determines $$Q$$ iff: for a thing to be $$P$$ is for it to be $$Q$$, not simpliciter, but in a specific way. The parallel is useful in suggesting that determination is, like identity, a particularly intimate relation, in that determinables and determinates types and/or tokens may not be wholly distinct, and also in suggesting a modal treatment of determination. Hence just as the characterization of identity is illuminated by the following modal condition (stemming from the principle of the indiscernibility of identicals): • (I) $$P$$ is identical to $$Q$$ only if: necessarily, for all $$x$$, $$x$$ has $$P$$ iff $$x$$ has $$Q$$. so too is the characterization of determination illuminated by the following modal condition: • (D)$$P$$ determines $$Q$$ ($$P > Q$$) only if: (i) necessarily, for all $$x$$, if $$x$$ has $$P$$ then $$x$$ has $$Q$$; and (ii) possibly, for some $$x$$, $$x$$ has $$Q$$ but lacks $$P$$. The parallel also suggests that an account of determination needn’t impose conditions on conceptual entailment relations (as Searle’s did): just as Kripke’s (1980) results undermine the need for a conceptual entailment requirement on identity (e.g., water might be identical with H$$_2$$O even if ‘water’ and ‘H$$_2$$O’ don’t conceptually entail each other), “we should discount the traditional doctrine’s conceptual component and reconceive determination in wholly metaphysical terms” (Yablo 1992: 253). That said, as Yablo admits, asymmetric necessitation alone is insufficient to characterize determination (since, e.g., allowing conjunctions to determine their conjuncts). We will later consider various ways of filling in the modal conditions, including Yablo’s own essence-based approach (§3.5.1). Yablo’s discussion is also notable in highlighting an important seeming feature of determinables and determinates, according to which they do not causally compete: [No causal exclusion principle can] apply to determinates and their determinables—for we know that they are not causal rivals. […] any credible reconstruction of the exclusion principle must respect the truism that determinates do not contend with their determinables for causal influence. (1992: 259) This feature is key to a recently popular strategy for solving the problem of mental causation—an intriguing application which raised to salience the potential usefulness of determination for metaphysical and other philosophical purposes (see §5.2). ## 2. Features and ontological categories ### 2.1 Features We are now in position to list the features commonly taken to characterize determinables, determinates, and their relation;[4] the presentation is in terms of properties, and may require adjustment to apply to entities of other categories (see §2.2). This is not a minimal or axiomatic set: some features follow from others; moreover, there are cases to be made that some of these features do not hold in full generality. In addition, how to metaphysically understand these features varies, in ways we will discuss down the line. 1. Increased specificity: If $$P$$ is a determinate of (‘determines’) $$Q$$, then to be $$P$$ is to be $$Q$$, in a specific way. Determination relates properties that are more or less specific, relative to each other. For example, red is a determinate of color; scarlet is a determinate of red. 2. Irreflexivity, Asymmetry, Transitivity: For no $$P$$ does $$P$$ determine $$P$$; If $$P$$ determines $$Q$$, then $$Q$$ does not determine $$P$$; if $$R$$ determines $$P$$, and $$P$$ determines $$Q$$, then $$R$$ determines $$Q$$. These features, characteristic of strict partial orderings, follow just from Increased specificity. For example, scarlet does not determine itself, since no property can be more specific than itself. If scarlet determines red, then red does not determine scarlet, since red cannot be both less and more specific than scarlet. And if to be scarlet is to be red, in a specific way, and to be red is to be colored, in a specific way, then to be scarlet is to be a yet more specific way of being colored, such that scarlet determines color. 3. Relative, leveled determination: Except for minimally specific determinables and maximally specific determinates, a property’s characterization as determinable or determinate is relative. Relatedly, determinables can be determined at different ‘levels’ of specificity. For example, red is a determinate of color, but a determinable of scarlet; red, blue, and yellow determine color at one level of specification; scarlet, navy, and buttercup determine color at another (more finely grained) level of specification. 4. Determination ‘in respect of’ determinables: If $$P$$ determines $$Q$$, $$P$$ is more specific than $$Q$$ in respect of $$Q$$. The increase in specificity associated with determinates is, more specifically, ‘in respect of’ the determinable.[5] For example, if red determines color, red is more specific than color in respect of color. Determination in respect of a determinable is sometimes further explicated as involving specification along one or more ‘determination dimensions’ of the determinable—for example, the determination dimensions of color are commonly supposed to be hue, saturation, and brightness. 5. Determinate similarity/comparability: If $$P$$ and $$R$$ are different same-level determinates of a determinable $$Q$$, then $$P$$ and $$R$$ are similar, and moreover comparable, in respect of $$Q$$. That determination is ‘in respect of’ determinables gives rise to similarity and systematic comparability relations among determinates. For example, determinates of colorred, blue, orange, etc.—are all similar in being colors, and any two colors are comparable in this respect: red is more similar to orange than to blue in respect of color, and so on. 6. Non-conjunctive specification: if $$P$$ determines $$Q$$, then $$P$$ is not identical with any conjunctive property conjoining $$Q$$ with any property or properties independent of $$Q$$. The increase in specificity associated with determinates does not involve conjoining or otherwise combining the determinable with another property independent of the determinable. For example, red is not appropriately understood as the conjunction of color and another property independent of color. 7. Non-disjunctive specification: if $$P$$ determines $$Q$$, then $$Q$$ is not identical with any disjunctive property disjoining $$P$$ with any property or properties independent of $$P$$. The increase in specificity associated with determinates does not involve disjunctive elimination. For example, color is not a disjunction of red and some other property or properties independent of red. 8. Determinable inheritance: For every determinable $$Q$$ of a determinate $$P$$: if $$x$$ has $$P$$ at a time $$t$$ then $$x$$ must have $$Q$$ at $$t$$. Objects must have every determinable of any determinates they have. For example, if something is scarlet, then it must also be both red and colored. 9. Requisite determination: If $$x$$ has $$Q$$ at a time $$t$$, then for every level $$L$$ of determination of $$Q$$: $$x$$ must have some $$L$$-level determinate $$P$$ of $$Q$$ at $$t$$. Objects must have a determinate of every determinable they have. For example, if something is colored, then it must be some specific color (e.g., red). 10. Multiple determinates: For every determinate $$P$$ of a determinable $$Q$$, there is a determinate $$R$$ of $$Q$$ that is distinct from, but at the same level of specificity as, $$P$$. With determinate specificity comes multiplicity or diversity. For example, color has multiple determinates (red, blue, yellow, and so on); red has multiple determinates (scarlet, crimson, burgundy, and so on). 11. Determinate incompatibility: If $$x$$ has determinate $$P$$ of determinable $$Q$$ at time $$t$$, then $$x$$ cannot have, at $$t$$, any other determinate $$R$$ of $$Q$$ at the same level of specificity as $$P$$. For example, an object cannot be both red and blue all over, at a time.[6] Determinate incompatibility is moreover ‘opponent’ in that same-level determinates are exclusive alternatives.[7] 12. Unique determination: If $$x$$ has a determinable $$Q$$ at a time, then $$x$$ has a unique—one and only one—determinate $$P$$ at any given level of specification at that time. Unique determination is entailed by Requisite determination and Determinate incompatibility. 13. Asymmetric modal dependence: if $$P$$ is a determinate of $$Q$$, then if $$x$$ has $$P$$ then $$x$$ must have $$Q$$, but for some $$y$$, $$y$$ might have $$Q$$ without having $$P$$. For example, anything that is scarlet must be red, but something might be red without being scarlet (it might rather be, e.g., burgundy). Asymmetric dependence is entailed by Determinable inheritance, Requisite determination, and Multiple determinates. 14. Causal compatibility: Determinables and determinates do not causally compete. For example, if a given patch is both red and scarlet, there is no in-principle difficulty with both red and scarlet being causally efficacious vis-á-vis the pecking of a pigeon trained to peck at any red patch. Motivated as they are by a limited range of paradigm cases, not all of these features may be characteristic of determination in the strong sense of being required for the holding of that relation, as opposed to being typically or generally true of some or most instances of the relation (or its relata). For example, Johansson (2004) observes that if property is an determinable of color, then Determinate incompatibility (hence Unique determination) will fail, noting that on this score “Johnson […] generalized too quickly” (2004: 117); what is rather true is that “for some ontological determinables […] two determinates cannot possibly exist in the same space-time region” (2004: 118). Armstrong offers other cases of purportedly compatible determinates, involving odors, tastes, or sounds: “a sauce may be both sour and sweet. A bell may have a tone which is composed of both a fundamental and a number of higher overtones” (1978b: 113; see also Fales 1990, Sanford 2014, and Massin 2013). It may also be questioned whether all determinates of determinables can be systematically compared, as per Determinate similarity/comparability. It is unclear, for example, whether being pentagonal is more or less similar in respect of shape to being square or being sextagonal. It is also unclear whether all determinates of smell admit of systematic comparability (is the smell of pine more similar to the smell of rose or of violet?), though recent empirical work identifying the dimensions of “smell space” (see, e.g., Mamlouk & Martinetz 2004) may reveal an underlying structure of the sort (appealing to hue, saturation, and brightness) that serves as an underlying basis for ordering determinate colors (which work might also serve as a basis for responding to Armstrong, above). The question of whether all determinables may be systematically ordered with respect to their determinables, either explicitly or implicitly, matters to whether determinate/determinable relations can be modeled in spatial terms along lines proposed by Funkhouser (2006) (see §3.6.1). Unique determination and Requisite determination may also admit of exceptions. Wilson (2013) argues that even for paradigmatic determinables such as color, Unique determination is not generally correct, since the case of an iridescent feather is reasonably interpreted as one in which the feather is (a) colored, (b) red from one perspective, and (c) blue from another perspective, such that no one determinate shade is non-arbitrarily identified as “the” unique determinate determining the determinable color had by the feather (rather, she suggests, the determinates are had, at best, in relativized fashion).[8] More generally, she argues that there can be failures of Unique determination due either to there being too many candidate determinates (as in the feather case) or to there being too few or none, contra Requisite determination (as might be the case for certain quantum phenomena; see below). Fales observes that Requisite determination appears to admit of exceptions in the case of certain objects of mental acts: for example, as regards the number of spots on an imagined or a hallucinated speckled hen. (1990: 167) Similarly, Rosen and Smith (2004) suggest that considerations of vagueness indicate that an object may have a determinable without determinately having any corresponding determinate, and Bokulich (2014) and Wolff (2015) suggest that cases of value indeterminacy in quantum mechanics are promisingly seen as involving undetermined determinables (see §5.3). Finally, as we will see in §3.4, a reductive approach to determinables, along lines of, e.g., Bigelow and Pargetter (1990), Clapp (2001), Rodriguez-Pereyra (2002), Antony (2003), and Massin (2013), takes these to be disjunctive properties having associated determinates as disjuncts; disjunctive reductionists uniformly reject Non-disjunctive specification as ever true of determination or its relata. ### 2.2 Ontological categories Though paradigm cases of determination (involving color, shape, mass, and so on) appear to hold between monadic property types, it is straightforward to extend the notion to relations—e.g., distance. Hence Johansson says that while the restriction to monadic types is ‘traditional’, I do not think there are any good reasons for the restriction. […] Relational concepts, just like property concepts, can be fitted into determinable-determinate trees. (2000: 117) More controversially, determination has been taken to hold between entities of ontological (including representational) categories going beyond properties and relations. To start, as above, historical discussion of determinables and determinates (in, e.g., Johnson 1921) moves fairly freely between talk of these as representational (linguistic/conceptual) and as metaphysical. In some such cases, the target is plausibly a properly metaphysical distinction.[9] In other cases—e.g., in empiricist treatments—the characterization of determination in representational terms is clearly intended to be metaphysically deflationary. We will later revisit reasons for and against taking determinables or determination at metaphysical face value (§3.1, §3.2). Modulo deflationary concerns, there does not appear to be any in-principle barrier to allowing that determination can hold between entities of other categories.[10] Determination has been extended to actions or episodes (see, e.g., Mulligan 1992 and Cruse 1995: chap. 6), events (Yablo 1992), fields (von Wachter 2000), tropes (Funkhouser 2006), states (Fine 2011), and substances (Massin 2013). That said, one might maintain that some such entities can be related by determination since falling under or in part constituted by determinable and determinate properties. Whatever one’s position on what ontological categories the relata of determination can fall under, there are (as per metaphysical usual) different available understandings of the categories at issue.[11] Another question pertaining to ontological categories concerns whether some or all specification relations are themselves determinates of a more general relation determinable; Johansson (2000: sec. 8) endorses such a view. ## 3. Contemporary accounts of determinables and determinates We turn now to canvassing contemporary accounts of determinables, determinates, and their relation. The main varieties of contemporary ontological accounts are anti-realist accounts, on which determinables do not exist; reductive accounts, on which determinables exist but are metaphysically reducible to (that is, identical with) some or other construction of determinates; and non-reductive accounts, on which determinables exist and are not metaphysically reducible to any construction of determinates. Non-reductive accounts furthermore split into versions on which determinables are (always) less fundamental than associated vis-á-vis determinates, and accounts on which determinables can be as fundamental as, or even more fundamental than, determinates; we reserve treatment of this further issue for §4. A different sort of contemporary account is directed not so much at the ontological question(s) as at elucidating certain structural or broadly logical aspects of determination or associated relata. As background for the ontological accounts to come, we start with a number of considerations commonly taken to provide prima facie support for taking determinables to exist, and to moreover be ontologically irreducible (again: not identical to any construction of) to determinates. We then consider some preliminary motivations for deflationary (anti-realist or reductive) treatments of determinables. One might wonder whether the motivations for taking determinables to exist might be presented independently of the motivations for taking determinables to be irreducible; but this proves difficult, for there simply isn’t enough daylight between the initial motivations for existence and for irreducibility to prise them apart. The presentation thus reflects that the motivations for reducibility (or for anti-realism) come later, as pushing against taking the prima facie appearances of irreducible determinables at face value. ### 3.1 Preliminary motivations for the existence and irreducibility of determinables There are perceptual, scientific, causal, semantic, and metaphysical considerations weighing in prima facie favor of determinables’ existing and moreover being metaphysically irreducible to determinates. Again, these motivations are preliminary; as we’ll see down the line, anti-realists and reductionists take there to be good reasons to resist taking the seeming appearances of existence and/or irreducibility at face value (see §3.2). #### 3.1.1 Determinables as perceived Perception provides one source of support for there being determinables, which on the face of it are not appropriately taken to be reducible to determinates. The best motivations here are indirect, since bare claims that we can perceive determinables as distinct from determinates are disputable.[12] One such motivation appeals to Sorites phenomena as indicating that we fail to perceive fully determinate instances of many properties, including colors, tones, and textures. As Fales puts it, this is a conclusion which seems forced upon us by the fact that each member of a series of colors, etc., may be perceptually indistinguishable from its immediate neighbors but easily distinguishable from more distant members of the series. (1990: 172)[13] A second indirect perceptual motivation for existing and irreducible determinables reflects that our perception of macro-objects typically fails to register microdeterminate details. For example, shape is a paradigmatic determinable, but we do not perceive the shapes of trees, tables, and the like, in microscopic detail. In re this and the Sorites motivation, it is worth noting that even if perceived objects are themselves maximally determinate, perceptual experience may still motivate irreducible determinables; as Wilson notes, Perhaps the (instances of) properties perceived are really maximally determinate, and only perceptual features or modes of presentation are determinable; but features of perceptual experience are also aspects of reality, so the larger point remains. (2012: 5) A third indirect perceptual motivation for irreducible determinables is that these best explain our being able to perceive that different determinates are similar in the relevant determinable respect. Hence Fales suggests that since we are able “to notice or be aware of” the similarity between different shades of red, it seems clear that red is observable every time we observe an instance of a shade of red, even though it is not separable from that shade, but is in the shade. (1990: 172) #### 3.1.2 Determinables as scientific posits Both fundamental physical and special-scientific posits and laws appear to be characterized, at least in part, in determinable terms. As Armstrong (1997) and French (2014) observe, it is natural to see fundamental physical laws as relating determinables, such as mass and charge, or to themselves be determinable relations.[14] And in the special sciences, for example, the features of molecules relevant to chemical interactions abstract from the precise details of atomic position and momentum, and features of thermodynamic systems near critical points abstract from compositional and other micro-level details (see Batterman 1998). More generally, the posit of existing and irreducible determinables in both fundamental and special sciences is in line with a principle of explanation pervasive in the sciences, according to which good explanations do not cite unnecessary details (see, e.g., Strevens 2004).[15] #### 3.1.3 Determinables as causally unspecific Determinables appear to be causally unspecific, in having fewer powers—in being able to contribute to causing fewer effects, in the circumstances—than their associated determinates. For example, red has the power to get Sophie the pigeon, who pecks at any red thing, to peck; but red doesn’t have the power to get Sophie’s picky cousin Alice, who pecks only at scarlet patches, to peck—for a red patch might be not scarlet, but burgundy (see Yablo 1992, Wilson 1999, and Shoemaker 2001). Such a lack of causal specificity, combined with a principle of individuation of scientific properties according to which they are (perhaps only contingently) distinguished by different sets of powers, has been taken to support not just the existence but the irreducibility of determinable features. #### 3.1.4 Determinables as semantic referents On a simple and in some sense default semantics, a sentence of the form ‘$$a$$ is $$F$$’ is true just in case the entity picked out by $$a$$ has the property picked out by $$F$$. Hence if ‘This ruby is red’ is true then on the default semantics the predicate ‘red’ refers to the determinable red. #### 3.1.5 Determinables as a basis for determinate resemblance As previously, Johansson (2000) posits determinables as providing a metaphysical basis for perceived similarity among determinates. Wilson (2012) argues, more generally, that determinables may be considered ‘natural’ in the ‘joint-carving’ sense of Lewis (1983) in virtue of providing a non-gerrymandered basis for objective similarity among determinates. Again, each of the motivations in (3.1.1) – (3.1.5) is preliminary. If the seeming appearance of there being irreducible determinables can be defended against objections, however, the philosophical work these can do in encoding distinctively unspecific aspects of reality (as per the applications in §5) would constitute additional positive support for their posit. ### 3.2 Preliminary motivations against the existence or irreducibility of determinables The main preliminary motivations against determinables existing or being metaphysically irreducible to determinates reflect concerns that such determinables would be unparsimonious, give rise to causal overdetermination, and/or violate the supposition that reality is ultimately or fundamentally maximally determinate. These concerns and associated support for anti-realist (‘eliminativist’) or realist reductionist accounts are similar to those sometimes brought to bear against irreducible higher-level (e.g., special scientific) goings-on. 1. Reality as maximally determinate. One objection to irreducible determinables appeals to the supposition that reality itself is maximally determinate, such that there are no irreducible determinable properties; as Armstrong maintains, “[a]ll universals must be determinate” (1978b: 117). Relatedly, many suppose that there is no “vagueness” in the world; as Lewis maintains, “[t]he only intelligible account of vagueness locates it in our thought and language” (1986: 212).[16] 2. Determinables as failing to track exact similarity. Armstrong takes “the powerful truism that universals are strictly identical in their distinct instantiations” to support the claim that “properties that are lowest determinates […] are good candidates for being universals” (1997: 49). More weakly, one might maintain that the best reason for positing properties, however understood, is as explaining exact similarity (in a given respect) between objects. Supposing so, one might moreover maintain that only determinates exist, on grounds that particulars having determinable properties might fail to exactly resemble in determinate respects (e.g., red objects might be different shades of red).[17] 3. Determinables as ontologically or causally redundant. Irreducible determinables have been taken to be ontologically and causally redundant, since determinates which we have independent reason to posit can explain or do anything that determinables purportedly explain or do. Hence Massin (2013) maintains against Johansson (2000) that perceived inexact resemblances between determinates can explain the “seen unity of the determinate colors”, and Gillett and Rives maintain that determinables are causally unnecessary: [P]ositing instances of determinate properties offers the best explanation for the causal relations we observe in the world […] Should we also accept that there are instances of determinable properties corresponding to predicates such as ‘is colored’ or ‘is charged’? […] There is a clear concern that it is ontologically profligate to take two properties to be contributing causal powers, a determinable and its determinate, where one, the determinate, would apparently suffice. (2005: 487) Heil (2003) levels similar complaints. 4. Determinables as inducing causal overdetermination. Related to the concern about casual redundancy is the concern, again pressed by Gillett and Rives (2005), that irreducible broadly scientific determinables induce problematic causal overdetermination. To start, if such determinables exist, they need to earn their causal keep.[18] But following the concern about causal redundancy, determinables simply reproduce the powers of their determinates, allowing instances of distinct properties to give rise to the same effect on a given occasion. And while such overdetermination sometimes occurs, as in firing squad or double-rock-throw cases, this is not the right model for determinables and determinates.[19] 5. Determinables as incoherent. Armstrong (1978a: 106–7, 117–18) argues that determinables are incoherent, in constituting the respect in which associated determinates are both similar and dissimilar (see Elder 1996: 151 for discussion).[20] Fales (1990: 174) sees the concern as incorrectly presupposing that determinables are second-order properties of or relations between determinates. Massin (2013) thinks the concern depends on whether the similarity at issue is exact or inexact; if inexact, there is no problem (“[t]wo things that are imperfectly similar in their colour can be imperfectly dissimilar in their colour”) but if exact—in virtue, e.g., of determinables being exactly similar constituents of determinates—then incoherence follows, since differences between determinates must be explained by what “remains” once the shared determinable is removed, but we cannot make sense of such removal. ### 3.3 Anti-realist accounts While motivated by the previous concerns, anti-realists about determinables also aim to explain or explain away (or “debunk”) at least some seeming motivations for determinables. A common debunking strategy appeals to something like Berkeley’s (1710) claim that it is a mistake to suppose that general terms denote abstract ideas.[21] For example, Heil (2003) echoes Berkeley’s claim, taking seeming semantic reasons for positing determinables to reflect an uncritical tendency to suppose that “we can ‘read off’ features of reality from our ways of speaking about it” (207), such that even if, e.g., the predicate ‘is red’ truly applies to some objects (tomatoes, apples, rubies) we should not take this to be in virtue of their possessing the “very same property”. Different positive anti-realist accounts of determinate terms or concepts and their function are available. Berkeley endorses a schematic account, on which general terms/concepts indeterminately refer to determinate features—here, generality is placed in language or mind, rather than world.[22] Heil (2003) also endorses a schematic approach, offering a pragmatic account of the function of determinable terms, as reflecting that we happen to have perceptual systems suited for tracking certain inexact similarities between objects having (ultimately highly complex) determinate properties. That the ‘work’ determinable terms do consists in tracking inexact similarities between objects or determinates is a common theme in three alternative anti-realist accounts. First is Searle’s (1959) account, on which determinable terms are higher-order predicates of determinate predicates.[23] Second is Mill’s account of determinable predicates as names not of one single and definite attribute, but of a class of attributes. Such is the word color, which is a name common to whiteness, redness, etc. (1843/1973: I.ii.5) Third is Armstrong’s account, first offered in his 1978a and developed in his 1997, according to which For a particular to “instantiate” a determinable is […] for it to instantiate just one of a class of universals’ […] having the determinable property entails no more than having one of the determinate properties. (1997: 53) One concern with anti-realist accounts has to do with the source of classification of determinates under a determinable term or concept. As above, this is usually taken to involve inexact resemblance, but if inexact resemblance is primitive, then objections of the sort leveled against resemblance nominalist treatments of inexact resemblance may arise, mutatis mutandis (see Fales [1990: 117–20] for discussion, and Rodriguez-Pereyra [2002] for a recent response). An alternative means of classification of determinates appeals to “naturalness” (as per Lewis 1983) as a primitive property binding determinates into classes, but this approach incurs ontological commitment and, it might be complained, is not clearly explanatory. Perhaps the most developed approach to determinate classification is Armstrong’s, on which “the resemblance of determinate universals is constituted by partial identity, where the greater the resemblance the greater degree of identity” (1997: 51), such that, for example, determinate lengths resemble in virtue of sharing unit length universals.[24] Armstrong argues that his account makes sense of various features of determination: Determinate opponent incompatibility is accommodated since, e.g., an object’s length cannot be identical to different sums of unit lengths; Determinate similarity/comparability is accommodated since different determinate lengths share some but not all unit lengths; Asymmetric dependence is accommodated since specific sums of unit lengths entail the having of length, but not vice versa. It has been complained, however, that his account cannot handle the full range of cases.[25] A second concern is that the presupposition that all properties are ultimately maximally determinate violates certain possible and potentially actual conceptions of natural reality. For example, if objects fill space, and space is both continuous and gunky (i.e., such that every entity has proper parts), then the shapes and locations of objects will be indefinitely subject to further determination (see Salmieri 2008: 78, n47).[26] And certain interpretations of quantum mechanics (see §5.3 for further discussion) and certain field-theoretic understandings of fundamental particles as having infinitely extensible boundaries (reflecting these particles’ being constituted by long-range interactions) suggest that natural reality is to some extent irreducibly determinable. A third, related concern is that it is unclear how the eliminativist can accommodate (or debunk) all the motivations for determinables. It remains unclear, for example, how to explain Sorites phenomena if all properties—including perceptual modes of presentation—are maximally determinate. ### 3.4 Reductionist accounts Reductionists take determinables to exist but to be reducible to determinates—that is, to be identical to classes or broadly logical constructions of (ultimately maximal) determinates. #### 3.4.1 Disjunctive reductionism The most common reductionist strategy takes determinables to be identical to disjunctions of (ultimately maximal) determinates (see, e.g., Bigelow and Pargetter 1990; Clapp 2001; Rodriguez-Pereyra 2002; Antony 2003;[27] and Massin 2013). On this view, for example, color is identical to a disjunctive property having every maximally specific color shade as a disjunct. Disjunctivism explains several features of determination: Requisite determination is accommodated since the instancing of a disjunction requires the instancing of a disjunct; Determinable inheritance is accommodated since the instancing of a disjunct entails the instancing of any associated disjunctions (see Rodriguez-Pereyra 2002, Massin 2013); Causal compatibility is accommodated via the identity of determinable (disjunctive) and determinate (disjunct) instances. And while disjunctivism violates Non-disjunctive specification, disjunctivists preserve some contrast by stipulating that determinables are a special sub-class of the class of disjunctive properties—ones having resembling determinates as disjuncts. Other stipulations may be required to accommodate Determinate incompatibility, Determination ‘in respect of’ determinables, and Unique determination. Three concerns face disjunctivist accounts, besides concerns that the required stipulations are unexplanatory, and that reality may not be (ultimately, entirely) maximally determinate. The first reflects reasons for rejecting disjunctive properties. One such reason—namely, Armstrong’s concern that “disjunctive properties offend against the principle that a genuine property is identical in its different particulars” (1978a: 20)—may be put aside if the disjuncts at issue are genuinely resembling, as proponents of disjunctivism suppose. This response will not, however, satisfy those taking disjunctions to be mere logical constructions, with no metaphysical import. The second concern is that taking the disjuncts identified with determinates to be genuinely resembling, as is needed in order to answer Armstrong’s concern and to provide an alternative basis for the contrast flagged in Non-disjunctive specification, will open the door to looking for a source of the supposed commonality, which source might well end up being an irreducible determinable. Here again the question of how best to treat the inexact resemblance between determinates is pressing. The third concern is that a disjunctive account cannot accommodate all the prima facie motivations for determinables. Again, our perceptions of and thoughts about determinables do not appear to be perceptions of or thoughts about disjunctions of maximal determinates. Rosen (2010) presses this point, noting that a “suggestive line of thought weighs against” disjunctivism, as per a case in which someone familiar with many shades of blue has no conception of cerulean (indeed, is perhaps even incapable of thinking of this shade), but where this deficit “would not prevent him from being competent with the word ‘blue’, or from knowing a great deal about the color blue” (2010: 128). #### 3.4.2 Relational accounts Bigelow and Pargetter’s (1990) relational account aims to explain both why determinates are similar and why they differ. They see irreducible determinables as insufficient unto this task, on grounds, first, that taking determinables to be irreducibly distinct from but necessarily accompanied by determinates introduces unexplained necessary connections (1990: 118), and second, that determinables alone don’t explain differences between determinates (1990: 54, 58; see also Armstrong 1978b: 113). Bigelow and Pargetter rather suggest that what it is for determinates to “fall under a common determinable” is for the determinates to instantiate second-degree relations of difference from one another, and for pairs of these differences to in turn instantiate third-degree proportions (1990: 58–62). Besides the usual concerns with reductive accounts, two concerns attach to a relational approach. The first is that attention to the way in which determinates differ doesn’t in itself serve to characterize the way in which determinates resemble. As Elder puts it, although properties such as being blue or rich or stabbingly painful do come in different degrees or intensities, there is more that is common to the various precise “values” of these properties than just that they differ from one another in graduated ways. (1996: 155) The second is that, as Elder also observes, from the metaphysical irreducibility of determinables to determinates, it doesn’t follow that their necessary connection is mysterious: if the various determinates […] each embodies the same property as the others […] it is hard to see any great mystery in the fact that possession of any one of these determinate shades necessarily goes with possession of the relevant determinable. (1996: 157) Also worth noting is that concerns about necessary connections typically arise from acceptance of Hume’s Dictum, according to which there are no metaphysically necessary connections between distinct existences; but post-empiricist motivations for believing this principle are unclear (see MacBride 2005 and Wilson 2010). ### 3.5 Non-reductive accounts We turn now to accounts on which determinables are taken to irreducibly exist. These accounts are typically implemented under the assumption that determinables are less fundamental than—are metaphysically posterior to—determinates. In §4 I’ll consider whether determinables may be as fundamental as determinates, and whether certain of the non-reductive accounts to follow are compatible with fundamental as well as non-fundamental determinables. #### 3.5.1 Essence-based accounts Yablo (1992) takes asymmetrical necessitation (i.e., Asymmetrical dependence) as a neutral modal basis for theorizing about determination: $$P$$ determines $$Q$$ just in case the traditional relation’s first, metaphysical component is in place, where this consists primarily in the fact that $$P$$ necessitates $$Q$$ asymmetrically. (1992: 253, note 23) Since reductionists also accept Asymmetric dependence, a modal account requires supplementation if it is to be non-reductive. Yablo does go beyond asymmetric necessitation in considering how (token, particular) events—the assumed relata of the causal relation—might stand in the determination relation; his approach appeals to the notion of essence of an event, and to the idea that the essence of one such event might be included in or subsumed by the essence of another: Determination involves the idea that the requirements associated with one thing include the requirements associated with another; and although properties are requiremental on their face, particulars are not. Hence the need for a notion of individual essence, where essential properties of a thing are those it cannot exist without. [T]he essence of a thing is supposed to be a measure of what is required in order to be that thing. Thus if more is required to be $$y$$ than to be $$x$$, this should be reflected in an inclusion relation between their essences: When $$q$$’s essence is a subset of $$p$$’s essence, $$p$$ is said to subsume $$q$$ ($$p \geq q$$); and $$p$$ determines $$q$$ ($$p > q$$) when the inclusion is strict. (1992: 261) Given the role essences play in individuation, this approach promises to characterize determinable events in properly non-reductive terms, at both the type and token levels, and might be extended to a more general non-reductive account of determinables, capable of accommodating the range of motivations for irreducible determinables. Taking determinables to have essences contained within determinate essences would also clearly explain various features of determination, including (besides Asymmetric dependence, which is presumably still in force), Determinable inheritance and Determination ‘in respect of’ determinables. More needs to be said, however if an appeal to essences is to explain certain other features of determination, including Non-conjunctive specification, Requisite determination, Determinate opponent incompatibility, and—importantly for Yablo’s intended application of determination to the problem of mental causation; see §5.2—Causal compatibility. #### 3.5.2 Causal and powers-based accounts Causal and powers-based accounts of determination are typically motivated by the problem of mental/higher-level causation, expressed by the question: How can higher-level properties bring about effects, as they seem to do, given that the physical properties upon which these higher-level properties depend are sufficient to cause these effects? As we’ll see in §5.2, different causal accounts provide different answers to this question. An important choice point, which I’ll track here, concerns whether determinable and determinate instances are taken to be token-identical. Though MacDonald and MacDonald (1986) don’t officially offer an account of determination, they were the first to see determination as useful for addressing the threat of causal overdetermination. Effectively, they suggest that the relation between mental and physical properties is analogous to determination: in both cases the properties at issue are, they maintain, type-distinct but token-identical, thus avoiding problematic overdetermination while preserving type-level causal relevance. Here and elsewhere, token-identity accounts of determinables and determinates do not clearly accommodate certain of the (e.g., perceptual) motivations for irreducible determinables; token-identity accounts are also subject to an extension of Horgan’s (1989) concerns with reductive accounts of mental states, according to which, insofar as the causal relata are ultimately located particulars (events, objects, property instances), the efficacy of determinables qua determinable—that is, qua distinctively unspecific—requires that determinable and determinate tokens as well as types be distinct. The first explicitly causal theory of determination is Fale’s (1990) account of ‘determinable or generic’ and ‘specific’ universals, on which the former have a proper subset of the ‘causal relations’ of the latter: If $$S$$ and $$S'$$ are specific universals and $$G$$ is a determinable or generic universal under which they both fall, then the causal relations between $$G$$ and other universals are a subset of the causal relations between $$S$$ and other universals; and also a subset of the causal relations between $$S'$$ and other universals. For $$S$$ and $$S'$$ to be species of a common genus is for them to share some subset of causal relations, which subset can itself constitute the causal essence of a universal, such as $$G$$. For $$S$$ and $$S'$$ to be distinct species is for this subset to be a proper subset of the causal relations which characterize $$S$$, and also of the causal relations which characterize $$S'$$: there must be some causal relation which $$S$$ has, but not $$S'$$, and some causal relation which $$S'$$ has, but not $$S$$. (1990: 175) A subset-of-causal-relations account of generic/determinable universals promisingly accommodates many features of determination, for reasons similar to those attaching to powers-based accounts, below. It also ensures the metaphysical irreducibility of a generic/determinable type to any specific/determinate type, by the principle of the indiscernibility of identicals (if $$G$$ has only a proper subset of causal relations of $$S$$, $$G$$ cannot be identified with $$S$$); and if this proper subset relation is preserved at the level of tokens, it will also block token-identity. More needs to be said, however, if this account is to ensure accommodation of Non-disjunctive specification (and more generally, if it is to block the metaphysical reducibility of determinable types to certain constructions of determinate types), since one might suggest (as a variation of the theme of Antony 2003) that disjunctive types, like Fales’s generic types, also have the causal relations in the intersections of their disjunctions. More problematic is that Fales’s account does not provide a clear basis for accommodating Non-conjunctive specification, since as stated it is compatible with a given determinate/species’ being conjunctively analyzable into (for example) a genus and differentia, each associated with a proper subset of the powers of the species. Even so, Fales’ account of determination is a clear precursor of Yablo’s (1992) essence-based account[28] and the powers-based account, to be next discussed. The first explicitly powers-based account of determination is offered by Wilson (1999), as an application of a powers-based approach to the metaphysical dependence (a.k.a. ‘realization’) at issue in physicalism. She first motivates attention to powers (rather than supervenience or epistemic notions, in particular) as a properly metaphysical basis for characterizing realization; she then argues that a variety of seemingly diverse non-reductive physicalist accounts, including Yablo’s determinable-based account, are similar in aiming to ensure that the powers of a realized property are a proper subset of those of its realizing property (properties), in particular. Hence: One way to make [Yablo’s] conjecture more plausible is to put the point in terms of the causal powers of the properties involved. […] the reason why “determinables and their determinates are not causal rivals” is because it is plausible, in the case of determinables and determinates, that each causal power of the determinable is identical to a causal power of its determinate. [Moreover, and for example,] the property of being scarlet, which is a determinate of the property of being red, evidently has more causal powers than the property of being red. (1999: 47) These considerations suggest the following first-pass powers-based account of determination: Powers-based determination (first pass): Property $$P$$ determines property $$Q$$ iff $$Q$$ is associated with a [non-empty] proper subset of the token powers associated with $$P$$. On this approach, a determinate is more specific than its determinable (as per Increased specificity) in being associated with a more specific set of powers. Clarke (1999), Shoemaker (2001, 2007), and Clapp (2001) also maintain that determination is a case-in-point of a non-reductive realization relation satisfying the proper subset condition on powers, where the powers associated with a given determinable property are those in the intersection of the sets of powers associated with its determinates. Wilson (2012) moreover argues, contra MacDonald and MacDonald (1986), Ehring (1996), and others, that determinables and determinates are token-distinct, since if they were token-identical, then determinable instances would be associated with more powers than are associated with their type—but that would be reason for denying that the instance was of the determinable type. This consideration also pushes against taking determinable types to be metaphysically reducible to (identified with) disjunctive or other constructions from determinates, since such reductive strategies entail that instances of determinables and determinates are token-identical. As such, and in particular, a powers-based approach accommodates Non-disjunctive specification. Proponents of powers-based accounts have different ways of accommodating Non-conjunctive specification. On Shoemaker’s version of a powers-based account of realization/determination, Non-conjunctive determination is stipulated: Property $$X$$ realizes property $$Y$$ just in case the conditional powers bestowed by $$Y$$ are a subset of the conditional powers bestowed by $$X$$ (and $$X$$ is not a conjunctive property having $$Y$$ as a conjunct). (2001: 78). Wilson (2009) rather accommodates Non-conjunctive specification by requiring that the powers in the complement of the sets associated with a determinate and any of its determinates, respectively, do not constitute a set associated with any property, as per her second-pass formulation: Powers-based determination (second pass): Property $$P$$ determines property $$Q$$ iff $$Q$$ is associated with a [non-empty] proper subset of the powers associated with $$P$$, and the set of powers had by $$P$$ but not by $$Q$$ is not (uniquely) associated with any property. In addition to Non-conjunctive specification and Non-disjunctive specification, a powers-based account of determination neatly accommodates most standard features of determination. Causal compatibility is accommodated, since every power of a determinable instance is token-identical with a power of an associated determinate. A powers-based account also satisfies Asymmetry, Irreflexivity, and Transitivity, which are inherited as features of the proper subset relation between powers. Determinable inheritance is accommodated in that in having a determinate property with an associated set of powers, an object will thereby have the determinables associated with proper subsets of the determinate’s powers (assuming that properties are at least contingently individuated by associated powers; see below for further discussion). Determination ‘in respect of’ determinables and Determinate similarity/comparability are accommodated by determinates’ sharing the powers of their associated determinable but having further powers, where the powers of the determinable constitute the respect in which determinables are similar, and powers had by determinates but not the determinable serve as a basis for ordering the determinates by degree of similarity (i.e., determinates sharing more powers will be more similar). That said, more could be said about how the sharing of determinable powers makes sense of how these additional powers count as being ‘in respect of’ the determinable and its powers, and as providing a basis for ordering by similarity.[29] Accommodation of Determinable similarity/orderability also provides a basis for responding to concerns that irreducible determinables make no sense, since we cannot ‘get a grip’ on the difference-makers of determinates (see §3.2): on a powers-based view, the difference-makers are, comprehensibly, the powers in the complement of determinate and determinable sets. A powers-based account does not itself ensure the accommodation of either Requisite determination or Unique determination,[30] but these omissions may ultimately be a virtue (see §5.3). A powers-based approach to determination faces two main concerns. One is that the view requires endorsement of a controversial “causal theory” of properties, on which these are (following, e.g., Shoemaker 1980) essentially and exhaustively constituted by their powers; for if properties have non-causal quiddities (primitive identities, which float modally free of powers), satisfaction of the proper subset condition on powers won’t guarantee that determinables are ‘nothing over and above’ determinates (see, e.g., Melnyk 2006: 141–43 and McLaughlin 2007), since the status of the determinable’s ‘quiddity’ is not thereby treated; nor will it explain Determinable inheritance, since satisfaction of the condition on powers is compatible with the absence of the determinable quiddities. Here it is worth noting that while certain proponents (e.g., Clarke, Shoemaker, Clapp) presuppose a causal theory of properties, such a commitment may be dispensable: since non-causal quiddities are scientifically inaccessible and besides the point of solving the problem of higher-level causation (see §5.2), their presence or absence is irrelevant to the status of determination (and more generally, realization) as a relation between broadly scientific properties (see Wilson 2011). A second concern with a powers-based approach is that, in applying only to broadly scientific goings-on, it doesn’t make sense of determination as holding between abstracta or other entities lacking powers. In response, a proponent of a powers-based account might try to extend the approach beyond specifically causal powers to one involving non-causal functional roles associated with, e.g., being a number (determinable), and being a prime number (determinate). This extension remains speculative, however; at present the account is not fully general. #### 3.5.3 Constitutional accounts It is natural to think of irreducible determinables as in some sense shared parts or constituents of determinates. Such a view would provide a natural way of accommodating Determinate similarity, Determinable inheritance, and other features. A powers-based view might be seen as taking determinables or associated powers to be constituent parts of determinates; here we consider non-powers-based constitutional accounts. Johansson (2000, 2004) endorses a constitutive account of determinables—more precisely, of ‘maximal’ determinables (such as color, shape, and smell) located at the top of a determination hierarchy—taking Determinate similarity and Determinate gradability as its starting point. It is, he says, characteristic of determinables and determinates that “two different determinates of the same determinable can always be continuously connected” (2000: 108), in a way providing the basis for the gradeable resemblance of the determinates. Conversely, Johansson maintains, determinates under wholly different determinables are not connected in this fashion: “No color-determinate resembles more closely a shape-determinate or volume-determinate than any other” (2000: 108). This ‘gap’ between determinables acts as a criterion, whereby a determinable is posited if and only if there is such a gap—hence it is that only maximal determinates exist, by lights of Johansson’s criterion. In cases where a determinable is appropriately posited, the determinable is shared by its determinates and explains their similarity; for example, all color-determinates have the determinable property color in common (Johansson 2000: sec. 3). Johansson’s view is a mixed eliminativist/non-reductionist view, since he assumes that any non-maximal determinables do not exist, but are merely “man-made constructions” (2000: 113). Such a view is unmotivated, however, since even in the absence of a ‘comparability gap’ between same-level determinates (e.g., yellow and red, or yellow and scarlet), one might still suppose that the difference between yellow and red (or scarlet) is not a matter of convention. Another concern with Johansson’s account reflects that the mere posit of a shared determinable does not (as Bigelow and Pargetter [1990] and Massin [2013] have observed) in itself make sense of determinable differences, though perhaps the appeal to the ‘continuous connection’ between determinates can do some work here. Worley (1997) sketches a different sort of a constitution-based account, as a way of supplementing Yablo’s (1992) asymmetrical necessitation account. She illustrates with triangularity, analyzed as involving the constituent features associated with being a closed, three sided plane figure, which features can vary in ways associated with equilateral, isosceles, and scalene triangles, among others. Worley sees this as motivating a model of determination on which “the constituents of each determinable are proper subsets of the constituents of each of their determinates”, such that, e.g., the determinable triangular and the determinate equilateral triangular would share features such as three-sided and closed plane figure, but the latter would be additionally constituted by equilateral. As Worley observes, this account accommodates various features of determination, including Determinable inheritance: There can be no metaphysically possible world in which an instance of a determinate can fail to be an instance of its determinable, since anything which is an instance of a determinate will be so in virtue of having certain features, which features will also be sufficient to make that object an instance of the determinable. (1997: 290) The shared constituent features also provide a basis for Determination ‘in respect of’ determinates and Determinate similarity/comparability, and perhaps the additional features of the determinates can provide a basis for Determinate incompatibility. The main concern with Worley’s account is that it doesn’t accommodate Non-conjunctive determination, and in fact seems to require its denial. #### 3.5.4 Grounding-based accounts ‘Grounding’ is a recently-introduced primitive relation or notion of metaphysical dependence, taken to be operative in any context where such dependence is at issue (either as holding between facts or propositions pertaining to some goings-on, or between worldly goings-on themselves). Proponents of Grounding typically take determination to be a paradigm case of such dependence, with determinables being Grounded in determinates. For example, Correia suggests that “the determinate features of a thing” are the Grounds of “the determinable features of a thing” (2005: 49–50); Schaffer (2016b) takes a case of determination as the primary exemplar of his ‘structural equations’ approach to Grounding; and Rosen (2010) endorses the following principle: Determinable-Determinate Link: If $$G$$ is a determinate of the determinable $$F$$ and $$a$$ is $$G$$, then $$[Fa] \leftarrow [Ga]$$ Here, $$[Fa] \leftarrow [Ga]$$ stands for ‘the fact that $$Fa$$ is Grounded in the fact that $$Ga$$’. Though Grounding is primitive, proponents stipulate that it has various features—most commonly, those of a strict partial order, in which case a Grounding-based account of determination will satisfy Asymmetry, Irreflexivity, and Transitivity.[31] The also-common assumption that Grounding entities metaphysically necessitate Grounded entities accommodates Determinable inheritance. How to accommodate other features (including Non-conjunctive specification, Non-disjunctive specification, Determination ‘in respect of’ determinables, and so on) is less clear, since Grounding doesn’t itself distinguish determination from other dependence relations. Correspondingly, the overriding concern with a Grounding-based account is that primitive Grounding alone fails to shed any substantive light on determination, instead simply registering the common supposition that determinables metaphysically depend on determinates in a way conforming to certain stipulated formal features. This concern is a case-in-point of the more general concern raised in Wilson (2014), according to which Grounding is too abstract to shed any useful metaphysical light on its own. Some proponents of Grounding attempt to provide more substance to a Grounding-based account of determination by deriving or modeling the claim that determinables are Grounded in determinates, but whether these attempts succeed is unclear; see, e.g., Schaffer’s (2016a,b) suggestion that determination can be modeled using the apparatus of structural equations (à la Pearl 2000 and Spirtes and Scheines 1993), and reasons to resist that suggestion offered by Koslicki (2016) and Wilson (2016). ### 3.6 Accounts of the structure or logic of determination We now discuss three accounts which are best seen as elucidating certain structural or broadly logical features of determination or determinates. #### 3.6.1 Property-space accounts Funkhouser’s (2006) property-space account (see his 2014 for further exposition) starts by filling in what it is to have a determinable “in a specific way” (Increased specificity) by attention to ‘determination dimensions’—features along which a determinable may be determined, such as, e.g., hue, saturation, and brightness in the case of color. The suggested methodology for identifying determination dimensions of a given determinable is broadly empirical, and involves inquiring after “the ways in which determinates under the determinable $$X$$ can differ from one another with regard to their $$X$$-ness” (2006: 551), which suggestion picks up on both Determination ‘in respect of’ determinables and Determinate incompatibility. Funkhouser then offers a model on which determination dimensions, assumed to have discrete or continuous ranges of values, serve as coordinate axes of a ‘property space’ for a given determinable. The space as a whole corresponds to the determinable; points in the space correspond to maximal determinates, associated with specific assignments of values to each determination dimension; subspaces of the space correspond to less than maximal determinates (which themselves may be determinables relative to contained subspaces or points). For example, color has determination dimensions hue, saturation, and brightness, such that, e.g., red determines color if and only if the property space of red is a proper subspace of the property space of color. More generally: Property $$B$$ determines property $$A$$ iff 1. $$A$$ and $$B$$ share the same determination dimensions 2. $$B$$ has the non-determinable necessities of $$A$$ and 3. The property space of $$B$$ is a proper subset of the property space of $$A$$. ‘Non-determinable necessities’ are those features of a determinable not allowing for variation (e.g., being three-sided, in the case of triangularity); condition (2) registers that sharing of non-determinable necessities is also required for determination, in addition to the increase in specificity associated with conditions (1) and (3), registering that determinates differ with respect to the values of one or more determination dimensions associated with the determinable. Modeling determinates as subspaces of a determinable space accommodates many features of determination, including, to start, Increased specificity, Determination in-respect-of determinables, Asymmetry, Irreflexivity, and Transitivity. Determinate incompatibility follows from subspaces’ being disjoint; Determinate similarity/comparability may be elucidated by reference to metrics on the spaces, whereby subspaces of a determinable space may be nearer or further apart. Funkhouser also aims to accommodate Causal compatibility by taking determinable and determinate instances (tropes) to be identical. With this latter claim, Funkhouser’s account moves into ontological territory, and might be plausibly seen as a sophisticated variation on a disjunctivist or other reductionist theme—say, one elucidating the internal dimensional structure of determinate disjuncts vis-á-vis associated determinables.[32] In any case, as with reductive accounts, Funkhouser’s assumptions that reality is ultimately maximally determinate and that determinable and determinate instances are token-identical are subject to challenge as either empirically inadequate or such as to not properly accommodate the perceptual and other motivations for determinables. That said, it seems likely that the structural aspects of Funkhouser’s approach could be preserved even if these assumptions are relaxed, such that, e.g., ‘points’ in a property space aren’t taken to correspond to maximal determinates (see note 45). #### 3.6.2 Metric-space accounts Denby (2001) presents a metric-space account of “determinables-together-with-their-determinates”, which “identifies abundant determinables and determinates with impure sets and the having relation with a set-theoretical relation” (2001: 316). Determinables are defined as follows: $$F$$ is a determinable iff $$F$$ is a pair $$\langle S, f\rangle$$ where: 1. $$S$$ is a metric space; and 2. $$f$$ is a function from concrete particulars into the points of this metric space. Here a determinable is thought of as a “classification” of particulars. The metric space $$S$$ is a collection of points together with a “distance” relation on these points; it defines the possible classificatory locations and how they are related. Each point corresponds to a determinate of that determinable, and the distance relation provides a measure of the similarities among them. The function $$f$$ maps each concrete particular in its domain to a unique point in the metric space. Determinates are defined as follows: $$F$$ is a determinate iff $$F$$ is a pair $$\langle D, E_i\rangle$$ where 1. $$D$$ is a determinable; and 2. $$E_i$$ is the extension of point $$i$$ in the metric space encoded by $$D$$. In set-theoretic terms, a determinate is thought of as a “centered” classification of particulars, defined by a determinable and by the determinate’s extension. Denby moreover sees determinables as ultimately categorizing the possession of maximal or “sparse” determinates;[33] as such, he interprets his model along anti-realist or reductionist lines. This approach to determinables and determinates is attractive, Denby argues, in that sets are ontologically unproblematic and their identity-conditions and relations well-understood (and similarly for functions and metric spaces), and in that the account accommodates “the pre-theoretic intuitions about the logical relations among properties” (2001: 310), including Unique determination,[34] Determinate similarity/comparability,[35] and Determinate incompatibility. The primary concerns here are again those associated with anti-realist or reductive views of determinables, though as with Funkhouser’s account it might be that the broad structure of Darby’s metric conception of determinables is compatible with a non-reductivist reading. #### 3.6.3 Co-determination accounts Fine’s aim in his 2011 is to characterize “what it is for the world to possess the kind of structure that it has when the determinate/determinable distinction is in play” (2011: 162), taking as inspiration Wittgenstein’s post-Tractarian view, on which the atomic propositions might exhibit the kind of dependence that is characteristic of the way in which different determinants of a given determinable are exclusive of one another. (2011: 161) Fine’s project reflects the usefulness of Determinate incompatibility for his truthmaker semantics (Fine 2012a,b)—most notably, in providing a non-similarity-based account of the worlds entering into counterfactual deliberation, with the general idea being that a counterfactual world will be just like the actual world but for the replacement of some actual state $$s$$ with a state $$s'$$ that is a determinate under the same determinable as $$s$$. As such, Fine’s ultimate aim is to provide a definition of the ‘co-determinable’ relation between states—a task which “avoids the tricky question of saying what exactly a determinable is” (2011: 184), though certain structural characterizations of determinables fall out of his account. Since Fine’s approach to characterizing determinable-determinate structure is rather intricate, we direct the interested reader to the attached footnote.[36] ## 4. Are determinates metaphysically prior to determinables? It is commonly assumed that determinables, if they exist, are less fundamental than determinates. Of course, this assumption is true on reductionist accounts;[37] but even those taking determinables to be irreducible to determinates often suppose that the latter are prior to the former. In this section we canvass reasons for and against taking determinates to be metaphysically prior to determinables. ### 4.1 Asymmetrical necessitation/supervenience One criterion of priority takes it to be sufficient for some goings-on to be less fundamental than some others that the former are asymmetrically necessitated by (or “metaphysically supervene on”) the latter. As per Asymmetrical dependence, it is characteristic of determinates that they asymmetrically necessitate their associated determinables, thanks to the distinctively unspecific nature of determinables. Hence Armstrong says: A point to be taken very seriously is that because determinates entail the corresponding determinable, the determinable supervenes on the determinates, and so, apparently, is not something more than the determinates. (1997: 50) Asymmetric necessitation appears to be insufficient for non-fundamentality, however, as is reflected in standard concerns that supervenience is too weak to guarantee nothing-over-and-aboveness (see, e.g., Horgan 1993 and McLaughlin and Bennett 2005). Nor is asymmetric necessitation necessary for the holding of a priority relation, as Fine’s (1994) case of Socrates and the singleton set of Socrates shows. As such, one can reasonably deny that the asymmetric necessitation of determinables by determinates establishes that determinables are less fundamental than determinates. Even accepting the criterion, one can question whether it shows that determinates are metaphysically prior to determinables. As French (2014) observes, while determinable instances do not entail any particular determinate instances, there is nevertheless ‘two-way’ entailment: every determinate instance entails determinable instances, and every determinable instance entails some determinate instance. One might respond that the word ‘some’ expresses residual asymmetry in priority as opposed to simply reflecting that determinables are less specific than determinates, but as French notes, “those who exclude determinables from the fundamental base need to show what specificity has to do with fundamentality” (2014: 284). ### 4.2 Naturalness Another criterion takes more fundamental properties to be more ‘natural’, in the terminologically inapropos[38] meaning of Lewis (1983), in making for greater objective resemblance between their possessing particulars than do less fundamental properties. This conception appears to deem determinables less fundamental than determinates: since determinables are less specific than their associated determinates, particulars sharing a given determinate property (e.g., scarlet) will be more objectively similar in the relevant respect than will be particulars sharing an associated determinable property (e.g., red). As Wilson (2012) notes, however, objective resemblance can hold between properties as well as particulars; hence even granting that determinates make for more objective resemblance between particulars, one can reasonably maintain that among the perfectly natural joints are those associated with determinables.[39] Such a consideration would support McDaniel's (2010) supposition that the increase in specificity at issue in determination differs from that at issue in the disjunction-disjunct relation, in that determinables, unlike disjunctions, are as natural as their specifications. ### 4.3 The argument from fixing Finally, consider a criterion of fundamentality—associated with intuitive “all God had to do” scenarios—according to which fundamental entities fix all the rest. This criterion again seems to support taking determinates to be prior to determinables, for once the specific, or determinate, facts are fixed, then, it seems, more general/determinable facts are thereby fixed, with the latter being at best ontologically real abstractions from the former. By way of contrast, fixing the determinable facts does not fix the determinate facts (as is reflected in Asymmetric dependence). As Wilson (2012) argues, however, that determinable facts are fixed by determinate facts is undermined by attention to the constitutive modal fact about any given determinable instance—say, that instance of red—that it is of a type that might have been differently determined—say, by scarlet rather than maroon. Here the suggestion is that no determinate fact or complex combination of such facts can provide a non-gerrymandered basis for such a constitutive modal fact.[40] One response is to maintain that a fundamental base need only ground the contingent non-modal facts at a world, as Schaffer (2004: 90) suggests; if so, then the failure of determinate instances to ground certain modal facts about determinable instances won’t impugn the priority of determinates. Still, one might maintain that a supposed “ground” for some entity which failed to provide a metaphysical basis for modal facts constitutive of the entity would be incomplete, at best. As French puts it, fixing the non-modal facts and leaving modal flexibility out of the picture would yield an incomplete inventory. And the most obvious way of fixing the modal facts would be to include the determinables themselves in the fundamental base. (2014: 284) There is a worry that this line of thought will overgenerate, in that many higher-level features and objects (e.g., tables) are to some extent constitutively modally flexible; a possible response is that whether the line of thought applies to a given case will depend on whether the basis for the modal flexibility is itself derivable or otherwise clearly generated by lower-level goings-on. ### 4.4 Questions and implications The previous considerations suggest that some determinable features of reality might be fundamental, or as fundamental as determinate features. This result raises several questions, answers to which have implications for our understanding of determinables and determinates, and of the fundamental. One question pertains to whether the determinates associated with fundamental determinables are also part of the fundamental base. It seems so; for even granting that determinates cannot completely ground determinables, determinates seem also to be needed to be existential witnesses for more specific goings-on. In that case, a further question arises concerning how determinables and determinates jointly participate in a fundamental base. Wilson (2012) suggests that a model for understanding such joint founding may be found in the sciences, and more specifically in how general scientific laws combine with initial or boundary conditions (15–16), an analogy developed by French, who maintains that “scientific determinables and determinates jointly enter into the fundamental base for everything else” (2014: 285) A further question concerns whether existing non-reductionist accounts of determinables can make sense of these being fundamental. On the face of it, there is no special difficulty with accommodating fundamental determinables on either powers-based or constitution-base accounts of determination, or with such accounts’ making sense of joint determinable/determinate founding. On Johansson’s (2000) account, for example, it might be that determinates depend on determinables in that the latter are constitutive parts of the former (nota bene: not obeying weak supplementation), and determinables depend on determinates in that, modulo gappy quantum indeterminacy (see §5.3), determinable instances must be parts of some determinate whole. That determinables and determinates may be jointly fundamental has implications for our understandings of fundamentality and metaphysical dependence. Fundamental entities are commonly assumed to be independent, in being wholly distinct, and (relatedly, as per ‘Hume’s Dictum’) in being such that each can either exist together or exist apart. But if determinables and determinates can be equally fundamental, this understanding is incorrect; for given the intimate relationship between determinables and determinates, to say that these are equally fundamental is evidently not to say either that they are wholly distinct or (even granting Hume’s Dictum) modally independent. Moreover, if determinables and determinates can mutually found each other, other stipulated suppositions are called into question—e.g., that the fundamental entities are those that do not depend on any other entities, and that metaphysical dependence is asymmetric. The possibility of fundamental determinables also bears on various first-order metaphysical topics. For example, Paul (2002, 2012) endorses a one-category ontology (‘mereological bundle theory’), on which properties and relations are fundamental, and property composition is “the fundamental glue” (2012: 242). Paul allows that determinables might be among the fundamental properties and relations (note 22), and moreover suggests that such posits may be required to accommodate theories couched in terms of properties and relations which are “extremely abstractly specified” (2012: 245). ## 5. Applications We now canvass three among many topics in metaphysics where attention to determinables and determinates has played a significant role.[41] ### 5.1 Laws of nature and nomological modality As mentioned previously, several philosophers have seen determinables and determinates as relevant to our understanding of laws of nature in both physics and the special sciences. Armstrong (1997) argues that irreducible determinables are needed to make sense of functionally specified laws in physics. Consider Newton’s law of gravitation: $F = K (M_1 M_2)/D$ As Armstrong notes (see also Fales 1990: 61 and French 2014), substitution of the determinable quantities $$F$$, $$M$$, and $$D$$ with determinate values gives rise to “a law which stands as determinate to the original, determinable, law” (1997: 63). One might try to provide an anti-realist or reductionist treatment of these seemingly determinable laws. Armstrong (1997) argues, however, that such deflationary treatments cannot make sense of the role of functional laws in sustaining counterfactuals, at least if properties are Aristotelian universals, whose existence depends upon their being instantiated; for then there will typically (and perhaps necessarily, for the case of continuous quantities) be gaps in the actual instantiations of functional laws. As such, the unity of determinates in these cases involves not just partial identities (as discussed in §3.4) but “strictly identical second-order properties”, which are the “real […] determinables” (1997: 247). Functional laws of nature are then taken to be relations between these determinable universals. French offers a structuralist account of laws of nature as expressing determinable relations. As regards the above law, he says: If we insert specific values […], we obtain a determinate instance of the law, expressing the relation that holds between determinate properties. So, the law as a whole can be regarded as expressing a determinable relation, which can be made determinate in specific situations. (2014: 300) Again, French allows that determinable laws/relations may be fundamental. Reflecting that one reason for taking (some) determinables to be fundamental is that these are partly constituted by modal facts pertaining to their possible determinations, French moreover suggests that the determinable nature of laws provides a basis for physical modality: these determinate instances correspond to solutions that can be represented by (more) specific models of the theory, and it is the shared structure between these models that yields the relevant physical modality. The relationship between the more generic and more specific models can be captured by that between determinables and determinates. (2014: 300) Another application of determination to the metaphysics of laws concerns whether these are contingent, in having at best a restricted physical or nomological modality, or rather are metaphysically necessary. Kistler (2005) argues that such laws are metaphysically necessary, first arguing that the quantities in ‘laws of association’ such as the ideal gas law are reasonably seen as different determinables of the same determinate states, such that such laws are necessary; then generalizing on the base of systematicity considerations to other kinds of laws.[42] ### 5.2 The problem of mental/higher-level causation How can mental goings-on bring about physical effects, as they seem to do, given that the physical realizers of mental goings-on are already sufficient to cause these effects? This question gives rise to the problem of mental (more generally, higher-level) causation, encoding two threats: first, that mental properties, if efficacious, give rise to problematic causal overdetermination; second, that avoiding overdetermination requires that mental properties be causally excluded by their realizers. Overdetermination and exclusion are avoided if mental and physical goings-on are identical, but in that case it seems that mental goings-on are not efficacious qua mental (see Horgan 1989); as Yablo puts it: “Mental events are effective, maybe, but not by way of their mental properties” (1992: 249). Moreover, type identity seems at odds with the multiple realizability of mental properties. Some, including MacDonald and MacDonald (1986), Yablo (1992), Wilson (1999), and Shoemaker (2001), have suggested that a better response proceeds by taking mental properties to be determinables of their physical realizers.[43] The appeal to determination reflects various features of determinates and determinables, including Causal compatibility. As Yablo says, [W]e know that [determinables and determinates] are not causal rivals. […] Determinables and their determinates, like objects and their parts, are guaranteed to be on the same team. (1992: 259) Moreover, determinables are typically multiply determined, and given that determinables and determinates are type- and perhaps also token-distinct, one can argue (see §3.5) that determinables are distinctively efficacious. If mental properties are determinables of physical realizers, these features are inherited: as causal non-competitors, overdetermination is avoided while accommodating multiple realizability, compatible with the distinctness and distinctive efficacy of mental properties. Similar considerations apply to other higher-level goings-on, in which case realization might be determination, more generally. A determinable-based approach to realization is subject to two main concerns (besides those generally targeting a powers-based account; see §3.5), each aiming to show that realization cannot be determination, since realization lacks some key feature of determination (see Ehring 1996; Funkhouser 2006, 2014; Walter 2007; and Worley 1997). First is the argument from mental multiple realizability, according to which such realizability effectively violates Determination ‘in respect of’ determinables (entailing that determinates differ in respect of determinables), since Mental properties are said to be multiply realizable precisely because distinct physical realizers can be exactly the same with respect to the mental property they realize, (Walter 2007: 219) in which case [T]he physical realizers of the mental will not differ mentally at all, as they should if they are determinates of the requisite mental states. (Ehring 1996: 474) Second is the argument from mental superdeterminates, according to which a mental property could be maximally specified qua mental—as, e.g., a precise state of searing pain (Ehring), or a belief maximally specified along the dimensions of content and confidence (Funkhouser); but if realization is determination, this would falsely imply that mental super-determinates could be further determined. Wilson (2009) responds to these concerns by analogical attention to color metamerism, in which colors of the same hue, saturation and brightness can differ in respect of color, and are moreover reasonably taken to have physical determination dimensions. One metameric moral is that properties such as color may have physical determination dimensions; a second is that which determination dimensions are had by a given higher-level property may be science-relative. Applying the morals: there is no in-principle problem with mental properties’ having physical determination dimensions;[44] moreover, which dimensions are associated with mental properties may be science-relative. This allows accommodation of mental multiple realizability: mental properties with diverse physical realizers can be exactly alike, relative to the determination dimensions associated with psychology, compatible with these mental properties being different, relative to a lower-level physical science distinguishing these realizers. And given that different sciences may treat a single determinable as having different determination dimensions, mental properties may be super-determinate relative to a purely psychological science, while being further determined relative to a lower-level science. A powers-based implementation of the metameric morals[45] also provides a basis for a response to a third objection, due to Worley (1997), according to which the constituents of beliefs are not constituents of their physical realizers, as they would have to be if (as per Worley’s constitution-based account; see §3.5) beliefs were determinables of physical determinates.[46] Supposing that constituents are associated with sets of powers, the content of a belief may be a constituent of its physical realizers, after all—not qua content per se, but qua the relevant subset of powers. ### 5.3 Metaphysical indeterminacy Many phenomena appear to be indeterminate: macro-objects (clouds, mountains) appear to have imprecise boundaries; the future, it seems, might be genuinely open; on certain interpretations of quantum mechanics, some properties of a system (e.g., position and momentum, orthogonal spin components) cannot jointly have precise values. In these and other cases there are reasons to think that the indeterminacy is metaphysical, as opposed to merely semantic or epistemic. Elder (1996) suggests that determinables might be useful for purposes of making sense of “the idea that corresponding to (at least some) vague predicates there really are, in the world, vague properties” (1996: 150). Wilson (2013) more specifically argues that determinables and determinates provide the basis for an account of metaphysical indeterminacy (MI) having several advantages, including being reductive, compatible with classical logic, and such as to make systematic sense of MI in a wide range of cases, as follows: Determinable-based MI: For a SOA to be metaphysically indeterminate in a given respect $$R$$ at a time $$t$$ is for the SOA to constitutively involve an object (entity) $$O$$ such that (i) $$O$$ has a determinable property $$P$$ at $$t$$, and (ii) for some level $$L$$ of determination of $$P$$, $$O$$ does not have a unique level-$$L$$ determinate of $$P$$ at $$t$$. One application of Determinable-based MI is to macro-object boundary MI; here the conditions are satisfied in ‘glutty’ fashion, reflecting that macro-objects typically have multiple overlapping realizers, as with Mount Everest and the multiple overlapping aggregates of rock in its vicinity at any given time. On the suggested approach, Mount Everest has a determinable boundary property, but no unique determinate boundary property, since there are too many candidate determinates of the determinable, associated with the comparatively precise boundaries of its realizing aggregates (at best, these determinate boundaries can be attributed to Mount Everest in relativized fashion).[47] Determinable-based MI is also usefully applied to the case of value indeterminacy on (some interpretations of) quantum mechanics. As Darby (2010) and Skow (2010) argue, ‘metaphysical supervaluationist’ accounts of MI (see Akiba 2004, Barnes 2010, Williams 2008, and Barnes and Williams 2011) cannot accommodate such MI. On these accounts MI is effectively understood as primitive unsettleness about which maximally precise world or state of affairs is the case; but as Darby and Skow observe, such an approach cannot model quantum MI, since the Kochen-Specker theorem rules out simultaneously assigning definite values to all properties of a quantum system. By way of contrast, Determinable-based MI is compatible with the Kochen-Specker theorem (and more generally, Heisenberg’s uncertainty principle), since as discussed by Bokulich (2014) and Wolff (2015), one way for a given determinable to fail to be uniquely determined is if no candidate determinate is present or instantiated, as the Kochen-Specker theorem suggests can be the case. As Wolff puts it for the case of spin value indeterminacy: Permitting the instantiation of determinables without determinates helps to describe this phenomenon, because we can say that $$x$$-spin and $$y$$-spin are determinables with two determinates each, and [after a spin-$$z$$ measurement] neither of these determinates is instantiated even though the determinables “spin-$$x$$” and “spin-$$y$$” are. (2015: 385)[48] As Wolff goes on to note, a determinable-based account doesn’t resolve all the mysteries associated with quantum MI; on the other hand, our experiential and theoretical familiarity with determinables goes considerable distance towards rendering such MI intelligible—offering, along the way, a nice case-in-point of how determinables may be useful for purposes of illuminating the structure of natural reality. ## Bibliography • Aizawa, Ken and Carl Gillett (eds), 2016, Scientific Composition and Metaphysical Ground, London: Palgrave-Macmillan. • Akiba, Ken, 2004, “Vagueness in the World”, Noûs, 38(3): 407–29. doi:10.1111/j.0029-4624.2004.00476.x • Antony, Louise M, 2003, “Who’s Afraid of Disjunctive Properties?” Philosophical Issues, 13(1): 1–21. doi:10.1111/1533-6077.00001 • Aristotle, Metaphysics, in J.A. Barnes (ed.), The Complete Works of Aristotle: The Revised Oxford Translation, Oxford: Oxford University Press, 1984. • Armstrong, David M., 1978a, Universals and Scientific Realism, Vol I: Nominalism and Realism, Cambridge: Cambridge University Press. • –––, 1978b, Universals and Scientific Realism, Vol II: A Theory of Universals, Cambridge: Cambridge University Press. • –––, 1989, Universals: An Opinionated Introduction, Boulder, Colorado: Westview Press. • –––, 1997, A World of States of Affairs, Cambridge: Cambridge University Press. • –––, 2010, Sketch for a Systematic Metaphysics, Oxford: Oxford University Press. • Bacon, John, 1995, Universals and Property Instances: The Alphabet of Being, Oxford: Blackwell. • Barnes, Elizabeth, 2010, “Ontic Vagueness: A Guide for the Perplexed”, Noûs, 44(4): 601–627. doi:10.1111/j.1468-0068.2010.00762.x • Barnes, Elizabeth, and J. Robert G. Williams, 2011, “A Theory of Metaphysical Indeterminacy”, in Oxford Studies in Metaphysics, Volume 6, Karen Bennett and Dean W. Zimmerman (eds), Oxford: Oxford University Press, 103–48. doi:10.1093/acprof:oso/9780199603039.003.0003 • Batterman, Robert, 1998, “Why Equilibrium Statistical Mechanics Works: Universality and the Renormalization Group”, Philosophy of Science, 65(2): 183–208. doi:10.1086/392634 • Berkeley, George, 1710, “A Treatise Concerning the Principles of Human Knowledge”, reprinted in, 1948–1957, The Works of George Berkeley, vol. 2, A.A. Luce and T.E. Jessop (eds), London: Thomas Nelson, pages 21–115. • Bernstein, Sara, 2014, “Two Problems for Proportionality About Omissions”, Dialectica, 68(3): 429–41. doi:10.1111/1746-8361.12071 • Bigelow, John, and Robert Pargetter, 1990, Science and Necessity, Cambridge: Cambridge University Press. • Bokulich, Alisa, 2014, “Metaphysical Indeterminacy, Properties, and Quantum Theory”, Res Philosophica, 91(3): 449–475. doi:10.11612/resphil.2014.91.3.11 • Brentano, F., 1982, Descriptive Psychology, London: Routledge, 1995. • Christensen, Jonas, 2014, “Determinable Properties and Overdetermination of Causal Powers”, Philosophia, 42(3): 695–711. doi:10.1007/s11406-014-9517-y • Clapp, Lenny, 2001, “Disjunctive Properties: Multiple Realizations”, Journal of Philosophy, 98(3): 111–36. doi:10.2307/2678378 • Clarke, Randolph, 1999, “Nonreductive Physicalism and the Causal Powers of the Mental”, Erkenntnis, 51(2): 295–322. doi:10.1023/A:1005581414518 • Cook Wilson, J., 1926, Statement and Inference, 2 volumes, Oxford: Clarendon Press. • Correia, Fabrice, 2005, Existential Dependence and Cognate Notions, Munich: Philosophia Verlag. • Cruse, D. A., 1995, Lexical Semantics, Cambridge: Cambridge University Press. • Darby, George, 2010, “Quantum Mechanics and Metaphysical Indeterminacy”, Australasian Journal of Philosophy, 88(2): 227–245. doi:10.1080/00048400903097786 • Denby, David A., 2001, “Determinable Nominalism”, Philosophical Studies, 102(3): 297–327. doi:10.1023/A:1010314926955 • Descartes, René, 1644, Principles of Philosophy, V. R. Miller and R.P. Miller (trans.), Dordrecht: Reidel, 1983. • Dowe, Phil, 2010, “Proportionality and Omissions”, Analysis, 70(3): 446–451. doi:10.1093/analys/anq033 • Eddon, Maya, 2007, “Armstrong on Quantities and Resemblance”, Philosophical Studies, 136(3): 385–404. doi:10.1007/s11098-005-5384-5 • Ehring, Douglas, 1996, “Mental Causation, Determinables, and Property Instances”, Noûs, 30(4): 461–480. doi:10.2307/2216114 • Elder, Crawford L., 1996, “Realism and Determinable Properties”, Philosophy and Phenomenological Research, 56(1): 149–159. doi:10.2307/2108471 • Fales, Evan, 1990, Causation and Universals, London: Routledge. • Fine, Kit, 1994, “Essence and Modality: The Second Philosophical Perspectives Lecture ”, Philosophical Perspectives, 8: 1–16. doi:10.2307/2214160 • –––, 2011, “An Abstract Characterization of the Determinate/Determinable Distinction”, Philosophical Perspectives, 25(1): 161–87. doi:10.1111/j.1520-8583.2011.00224.x • –––, 2012a, “A Difficulty for the Possible Worlds Analysis of Counterfactuals”, Synthese, 189(1): 29–57. doi:10.1007/s11229-012-0094-y • –––, 2012b, “Counterfactuals Without Possible Worlds”, Journal of Philosophy, 109(3): 221–246. doi:10.5840/jphil201210938 • French, Steven, 2014, The Structure of the World: Metaphysics and Representation, Oxford: Oxford University Press. • Funkhouser, Eric, 2006, “The Determinable-Determinate Relation”, Noûs, 40(3): 548–569. doi:10.1111/j.1468-0068.2006.00623.x • –––, 2014, The Logical Structure of Kinds, Oxford: Oxford University Press. • Gillett, Carl and Bradley Rives, 2005, “The Nonexistence of Determinables: Or, a World of Absolute Determinates as Default Hypothesis”, Noûs, 39(3): 483–504. doi:10.1111/j.0029-4624.2005.00510.x • Granger, Edgar Herbert, 1984, “Aristotle on Genus and Differentia”, Journal of the History of Philosophy, 22(1): 1–23. doi:10.1353/hph.1984.0001 • Heil, John, 2003, “Levels of Reality”, Ratio, 16(3): 205–21. doi:10.1111/1467-9329.00218 • Hooker, C.A., 1981, “Towards a General Theory of Reduction”, Dialogue • “Part I: Historical and Scientific Setting”, 20(1): 38–59. doi:10.1017/S0012217300023088 • “Part II: Identity in Reduction”, 20(2): 201–236. doi:10.1017/S0012217300023301 • “Part III: Cross-Categorical Reduction”, 20(3): 496–529. doi:10.1017/S0012217300023593 • Horgan, Terence, 1989, “Mental Quausation”, Philosophical Perspectives 3, Philosophy of Mind and Action Theory, 47–76. doi:10.2307/2214263 • –––, 1993, “From Supervenience to Superdupervenience: Meeting the Demands of a Material World”, Mind, 102(408): 555–586. doi:10.1093/mind/102.408.555 • Hume, David, 1739–40, A Treatise of Human Nature, Oxford: Oxford University Press. • Hurley, Susan L., 1989, Natural Reasons: Personality and Polity, Oxford: Oxford University Press. • Jenkins, C.S., 2011, “Is Metaphysical Dependence Irreflexive?” The Monist, 94(2): 267–276. doi:10.5840/monist201194213 • Johansson, Ingvar, 2000, “Determinables as Universals”, The Monist, 83(1): 101–121. doi:/10.5840/monist20008312 • –––, 2004, Ontological Investigations: An Inquiry into the Categories of Nature, Man and Society, Frankfurt: Ontos Verlag. • Johnson, W.E., 1921, Logic (Part 1), Cambridge: Cambridge University Press. • –––, 1924, Logic (Part III: The Logical Foundations of Science), Cambridge: Cambridge University Press. • Kim, Jaegwon, 1989, “The Myth of Nonreductive Materialism”, Proceedings and Addresses of the American Philosophical Association, 63: 31–47. doi:10.2307/3130081 • –––, 1993, “The Non-Reductivist’s Troubles with Mental Causation”, in Mental Causation, John Heil and Alfred Mele (eds), Oxford: Oxford University Press, 189–210. • Kistler, Max, 2005, “Necessary Laws”, in Natures Principles, Jan Faye and Paul Needham, Uwe Scheffler, and Max Urchs (eds), Springer, 201–227. doi:10.1007/1-4020-3258-7_8 • Koslicki, Kathrin, 2016, “Where Grounding and Causation Part Ways: Comments on Jonathan Schaffer”, Philosophical Studies, 173(1): 101–112. doi:10.1007/s11098-014-0436-3 • Kripke, Saul, 1980, Naming and Necessity, Cambridge, MA: Harvard University Press. • Leibniz, Gottfried Wilhelm, 1704, Nouveaux Essais sur l'entendement humain (New Essays on Human Understanding), translation from Alfred Gideon Langley, 1896, New York: Macmillan. • Lewis, David, 1983, “New Work for a Theory of Universals”, Australasian Journal of Philosophy, 61(4): 343–377. doi:10.1080/00048408312341131 • –––, 1986, On the Plurality of Worlds, London: Blackwell. • Locke, John, 1690, An Essay Concerning Human Understanding, Oxford: Clarendon, 1975. • MacBride, Fraser, 2005, “Lewis’s Animadversions on the Truthmaker Principle”, In Truthmakers: The Contemporary Debate, Helen Beebee and Julian Dodd (eds), Oxford: Oxford University Press, 117–40. doi:10.1093/acprof:oso/9780199283569.003.0008 • MacDonald, Cynthia, and Graham MacDonald, 1986, “Mental Causes and Explanation of Action”, Philosophical Quarterly, 36(143): 145–158. doi:10.2307/2219765 • Mamlouk, A.M. and Thomas Martinetz, 2004, “On the Dimensions of the Olfactory Perception Space”, Neurocomputing, 58–60: 1019–1025, [reprint available online]. • Massin, Olivier, 2013, “Determinables and Brute Similarities”, in Johanssonian Investigations: Essays in Honour of Ingvar Johansson on His Seventieth Birthday, Christer Svennerlind, Jan Almäng, Rögnvaldur Ingthorsson (eds), Heusenstamm: Ontos Verlag, pages 388–420. doi:10.1515/9783110322507.388 • McDaniel, Kris, 2010, “A Return to the Analogy of Being”, Philosophy and Phenomenological Research, 81(3): 688–717. doi:10.1111/j.1933-1592.2010.00378.x • McLaughlin, Brian P., 2007, “Mental Causation and Shoemaker-Realization”, Erkenntnis, 67(2): 149–172. doi:10.1007/s10670-007-9069-7 • McLaughlin, Brian and Karen Bennett, 2005, “Supervenience”, The Stanford Encyclopedia of Philosophy (Fall 2005 Edition), Edward N. Zalta (ed.), URL = <https://plato.stanford.edu/archives/fall2005/entries/supervenience/>. • Melnyk, Andrew, 2006, “Realization and the Formulation of Physicalism”, Philosophical Studies, 131(1): 127–155. doi:10.1007/s11098-005-5986-y • Mill, John S., 1843/1973, A System of Logic, Toronto: University of Toronto Press. • Morganti, Matteo, 2011, “The Partial Identity Account of Partial Similarity Revisited”, Philosophia, 39(3): 527–546. doi:10.1007/s11406-010-9290-5 • Mulligan, Kevin, 1992, Language, Truth and Ontology, Dordrecht: Kluwer. • –––, 2006, “Ascent, Propositions and Other Formal Objects”, in Propositions: Semantic and Ontoligical Issues (Grazer Philosophische Studien 72), Massimiliano Carrara and Elisabetta Sacchi (eds), Amsterdam: Rodopi, pages 29–48. • –––, 2014, “Formal Concepts”, in The History and Philosophy of Polish Logic: Essays in Honour of Jan Woleński, Kevin Mulligan, Katarzyna Kijania-Placek, and Tomasz Placek (eds), London: Palgrave Macmillan UK, pages 205–23. • Paul, L.A., 2002, “Logical Parts”, Noûs, 36(4): 578–596. doi:10.1111/1468-0068.00402 • –––, 2012, “Building the World from Its Fundamental Constituents”, Philosophical Studies, 158(2): 221–256. doi:10.1007/s11098-012-9885-8 • Pearl, Judea, 2000, Causality, Cambridge: Cambridge University Press. • Poli, Roberto, 2004, “W. E. Johnson’s Determinable-Determinate Opposition and His Theory of Abstraction”, in Francesco Coniglione, Roberto Poli, and Robin Rollinger (eds), Idealization XI: Historical Studies on Abstraction and Idealization, (Poznań Studies in the Philosophy of the Sciences and the Humanities 82/11), pages 163–196. doi:10.1163/9789004333215_009 • Prior, Arthur N., 1949, “Determinables, Determinates, and Determinants (I,II)”, Mind, 58(229): 1–20, 58(230): 178–94. doi:10.1093/mind/LVIII.229.1 and doi:10.1093/mind/LVIII.230.178 • Rodriguez-Pereyra, Gonzalo, 2002, Resemblance Nominalism: A Solution to the Problem of Universals, Oxford: Oxford University Press. • –––, 2015, “Grounding is not a strict order”, Journal of the American Philosophical Association, 1(3): 517–534. • Rosen, Gideon, 2010, “Metaphysical Dependence: Grounding and Reduction”, in Modality: Metaphysics, Logic, and Epistemology, B. Hale and A. Hoffmann (eds), Oxford: Oxford University Press, 109–36. doi:10.1093/acprof:oso/9780199565818.003.0007 • Rosen, Gideon and Nicholas J.J. Smith, 2004, “Worldly Indeterminacy: A Rough Guide”, Australasian Journal of Philosophy, 82(1): 185–198. doi:10.1080/713659795 • Salmieri, Gregory, 2008, “Aristotle and the Problem of Concepts”, Ph.D. Dissertation, Pittsburgh, PA: University of Pittsburgh. [ Salmieri 2008 available online] • Sanford, David H., 2014, “Determinates vs. Determinables”, The Stanford Encyclopedia of Philosophy (Winter 2014 Edition), Edward N. Zalta (ed.), URL = <https://plato.stanford.edu/archives/win2014/entries/determinate-determinables/>. • Schaffer, Jonathan, 2004, “Quiddistic Knowledge”, in Lewisian Themes, Frank Jackson and Graham Priest (eds), Oxford: Oxford University Press, 210–230. • –––, 2012, “Grounding, Transitivity, and Contrastivity”, in Metaphysical Grounding: Understanding the Structure of Reality, Fabrice Correia and Benjamin Schnieder (eds), Cambridge: Cambridge University Press, 122–138. • –––, 2016a, “Ground Rules: Lessons from Wilson”, in Aizawa and Gillett 2016: 143–169. doi:10.1057/978-1-137-56216-6_6 • –––, 2016b, “Grounding in the Image of Causation”, Philosophical Studies, 173(1): 49–100. doi:10.1007/s11098-014-0438-1 • Searle, John, 1959, “Determinables and the Notion of Resemblance”, The Aristotelian Society, Supplement, 33(1): 141–58. doi:10.1093/aristoteliansupp/33.1.125 • Segelberg, I., 1999, Three Essays in Phenomenology and Ontology, H. Hochberg and S. Ringstroöm-Hochberg (trans.), Stockholm: Thales. • Shoemaker, Sydney, 1980, “Causality and Properties”, in Time and Cause, Peter van Inwagen (ed.), Dordrecht: D. Reidel, 109–135. • –––, 2001, “Realization and Mental Causation”, in Proceedings of the 20th World Congress in Philosophy, Cambridge: Philosophy Documentation Center, 23–33. • –––, 2007, Physical Realization, Oxford: Oxford University Press. • Skow, Bradford, 2010, “Deep Metaphysical Indeterminacy”, Philosophical Quarterly, 60(241): 851–858. doi:10.1111/j.1467-9213.2010.672.x • Spinoza, Benedictus, 1677/1885, The Collected Writings of Spinoza (Volume 1: The Ethics), Edwin Curley (trans.), Princeton: Princeton University Press. • Spirtes, Peter, Clark Glymour, and Richard Scheines, 1993, Causation, Prediction, and Search, Cambridge, MA: MIT Press. • Stout, G.F., 1930, Studies in Philosophy and Psychology, London: MacMillan. • Strevens, Michael, 2004, “The Causal and Unification Accounts of Explanation Unified—Causally”, Noûs, 38(1): 154–176. doi:10.1111/j.1468-0068.2004.00466.x • Tappolet, Christine, 2004, “Through Thick and Thin: Good and Its Determinates”, Dialectica, 58(2): 207–221. doi:10.1111/j.1746-8361.2004.tb00297.x • Unger, Peter, 1980, “The Problem of the Many”, Midwest Studies in Philosophy, 5(1): 411–468. doi:10.1111/j.1475-4975.1980.tb00416.x • van der Schaar, Maria Sandra, 1991, G.F. Stout’s Theory of Judgment and Proposition Proefschrift Ter Verkrijging van de Graad van Doktor, Ph.D. Dissertation, Leiden University. • von Wachter, D., 2000, “A World of Fields”, Poznan Studies in the Philosophy of the Sciences and the Humanities, 76: 305–326. • Walter, Sven, 2007, “Determinables, Determinates, and Causal Relevance”, The Canadian Journal of Philosophy, 37(2): 217–243. doi:10.1353/cjp.2007.0020 • Williams, J. Robert G., 2008, “Multiple Actualities and Ontically Vague Identity”, Philosophical Quarterly, 58(230): 134–154. doi:10.1111/j.1467-9213.2007.538.x • Wilson, Jessica M., 1999, “How Superduper Does a Physicalist Supervenience Need to Be?” The Philosophical Quarterly, 49(194): 33–52. doi:10.1111/1467-9213.00127 • –––, 2005, “Supervenience-Based Formulations of Physicalism”, Noûs, 39(3): 426–459. doi:10.1111/j.0029-4624.2005.00508.x • –––, 2009, “Determination, Realization, and Mental Causation”, Philosophical Studies, 145(1): 149–169. doi:10.1007/s11098-009-9384-8 • –––, 2010, “What Is Hume’s Dictum, and Why Believe It?” Philosophy and Phenomenological Research, 80(3): 595–637. doi:10.1111/j.1933-1592.2010.00342.x • –––, 2012, “Fundamental Determinables”, Philosophers’ Imprint, 12(4): 1–17. [Wilson 2012 available online] • –––, 2013, “A Determinable-Based Account of Metaphysical Indeterminacy”, Inquiry, 56(4): 359–385. doi:10.1080/0020174X.2013.816251 • –––, 2014, “No Work for a Theory of Grounding”, Inquiry, 57(5–6): 535–579. doi:10.1080/0020174X.2014.907542 • –––, 2016, “The Unity and Priority Arguments for Grounding”, in Aizawa and Gillett 2016: 171–204. doi:10.1057/978-1-137-56216-6_7 • Wolff, Johanna, 2015, “Spin as a Determinable”, Topoi, 34(2): 379–86. doi:10.1007/s11245-015-9319-2 • Worley, Sara, 1997, “Determination and Mental Causation”, Erkenntnis, 46(3): 281–304. doi:10.1023/A:1005301816477 • Yablo, Stephen, 1992, “Mental Causation”, The Philosophical Review, 101(2): 245–280. doi:10.2307/2185535
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6864295601844788, "perplexity": 5108.565376874999}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257650764.71/warc/CC-MAIN-20180324171404-20180324191404-00686.warc.gz"}
http://exxamm.com/QuestionSolution2/Chemistry/2642667533
### Question Asked by a Student from EXXAMM.com Team Q 2642667533. Assertion : For the reaction : Ni^(++) + 2 e^(-) -> Ni Fe^(++) + 2 e^(-) -> Fe => E_(Fe^(++)|Fe)^0 < E_(Ni^(++)|Ni) and E_(red)^0 > 0 => So, Fe electrode is cathode and Ni electrode is anode Reason : Because DeltaG^o < 0 and E_(Cell)^0 > 0, so cell is possible. A Both A and R individually true and R is the correct explanation of A B Both A and R are individually true but R is not the correct explanation of A C A is true but R is false D A is false but R is true #### HINT (Provided By a Student and Checked/Corrected by EXXAMM.com Team) #### Access free resources including • 100% free video lectures with detailed notes and examples • Previous Year Papers • Mock Tests • Practices question categorized in topics and 4 levels with detailed solutions • Syllabus & Pattern Analysis
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.24442626535892487, "perplexity": 6664.589488784131}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039742963.17/warc/CC-MAIN-20181115223739-20181116005739-00486.warc.gz"}
http://math.stackexchange.com/users/48131/cass
# Cass less info reputation 6 bio website location age member for 6 months seen Mar 24 at 5:08 profile views 7 # 4 Questions 8 Vector Bundle Over Contractible Manifold 3 Logarithm Series: Symbol Manipulation Proof that $\log(x) + \log(y) = \log (xy)$ 1 Proving a Certain $\mathbb{C}$-Algebra is a Domain Using a Specified Method 1 Extending a Set of Linearly Independent Vector Fields to a Basis # 130 Reputation This user has no recent positive reputation changes 4 $f: \mathbf{R} \rightarrow \mathbf{R}$ monotone increasing $\Rightarrow$ $f$ is measurable 2 The group homomorphism from Z to Q* # 14 Tags 4 homework 2 group-theory 4 proof-strategy 2 abelian-groups 4 proof-writing 0 abstract-algebra × 2 4 real-analysis 0 differential-geometry × 2 4 measure-theory 0 manifolds × 2 # 1 Account Mathematics 130 rep 6
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6972330212593079, "perplexity": 4581.375348588422}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368698493317/warc/CC-MAIN-20130516100133-00027-ip-10-60-113-184.ec2.internal.warc.gz"}
https://sage.math.clemson.edu:34567/home/pub/158/
# Mission Statement Create a viable free open source alternative to Magma, Maple, Mathematica, and Matlab. ## History of Sage • William Stein started Sage at Harvard in January 2005. • Sage-1.0 released February 2006 at Sage Days 1 (UC San Diego). • Nearly 30 Sage Days Workshops (!)  at UCLA, UW, Cambridge, Bristol, Austin, France, San Diego, Seattle, MSRI, ..., Barcelona, ... • Sage won first prize in the Trophees du Libre (November 2007) • Funding from Microsoft, Univ of Washington, UC San Diego, NSF, DoD, Google, Sun, private donations, etc. • Hundreds of other people subsequently got involved in Sage's development, and the scope of the project has widened to cover all mathematical computation. There is interest from all areas of mathematics, physics, engineering, etc.  (Developer mailing list has 1184 subscribers and about 50 messages/day... and sage-support has 1718 subscribers.) ## What is Sage? • Sage = Python + Math • A unified self-contained distribution of open source free mathematical software. • Nearly a half million lines of new Python (and Cython) code that implements new capabilities and algorithms. • A "cloud" application like GMail or Google Docs: http://sagenb.org  (currently about 30,000 users);  Of course Sage also runs on your desktop and supercomputer. # Summary: Sage is about building the car instead of reinventing the wheel 1. Sage uses Python,  a mainstream programming language, instead of inventing a custom mathematics language 2. Use straightforward method to link programs together -- C library and pseudotty's, instead of XML servers/OpenMath.  We implement all conversion routines, instead of expecting upstream to do it: we make them communicate with Sage, whether they want to or not. 3. Give copious credit to contributors and be very developer friendly (easily build from source). 4. Reuse, improve, and contribute to existing libraries and projects (e.g., Scipy, Numpy, R, ATLAS, CVXopt, GSL), instead of starting over and competing with them. 5. Make the GUI using a web browser: the world of java and javascript is available and Sage integrates with the web. # Key take-away points: • FOSS: 100% free open source software: good for clusters, sharing, research, collaboration • Python: very mainstream, scientific-computing friendly programming language • Cython: optimized Python to C/C++ compiler we develop with support for C/C++ datatypes • Interfaces: to control Matlab, Octave, Mathematica, etc., from Sage • Self contained: can have many copies of Sage at once, so no "dependency hell" • Peer reviewed: get your code refereed • Web based: notebook interface # All aboard! ## And now a short demo... # This is a simple introduction to Sage. # # The Sage notebook uses blocks of text, which are essentially Python statements that get executed. # Lines that start with the pound sign are comments and do not get echoed when the block is evaluated. # # A block is evaluated by clicking on the evaluate link or by typing shift-enter. ## Sage does Integer Arithmetic ### (with as many digits as needed) # Sage does arithmetic with the correct order of operations. 3+2*6^2 # Sage can do more extensive calculations factorial(100) With a little thought you can answer the following questions without a calculator or computer or even pencil and paper. How many trailing zeros should $100!$ have?  From the previous calculation it must be 24, but why? How many trailing zeros should $200!$ have?  Hint, it is not 48. # Sage knows about prime numbers and how to factor large integers print is_prime(12345678987654321) factor(12345678987654321) # and it can expand the previous result expand(_) # Sage knows about the Greatest Common Divisor print gcd(555,333666) # and the Least Common Multiple print lcm(555,333666) # We can double check these results by factoring print factor(555) print factor(333666) ## Sage does Rational Arithmetic ### (with as many digits as needed) # It can easily do the calculations involving rational numbers p = 1/555 print p q = 1/333666 print q p+q d = p*q print d print d*2*3^4*5*37^2 ## Sage does Decimal Arithmetic ### (with as many digits as you ask for) q = sqrt(2) q # What is q? type(q) # Pressing tab will give auto-complete options, or function descriptions after a parens ( q. q.n(1000) # or 1000 digits of the square root of pi sqrt(pi).n(digits=1000) _^2 pi.n(digits=1000) ## Sage does Algebra It can do similar calculations and plots with polynomials and rational functions. # Sage can factor and expand algebraic expressions factor(x^5-1) p = expand(_) p p^3 # Algebraic expressions can be plotted P1 = plot( p , -0.5, 0.5, rgbcolor=(1,0,0) ) P2 = plot( p^3, -0.5, 0.5, rgbcolor=(0,1,0) ) show(P1+P2) factor(p^3) # Sage handles rational functions with ease rf = x / (x^2 - 1) rf P3 = plot( rf , -0.5, 0.5, rgbcolor=(1,0,0) ) show(P3) # and partial fraction expansions pf = rf.partial_fraction(x) pf # Sage can display results in the more familiar mathematical format show(pf) # Sage uses LaTeX for the mathematical formatting latex(pf) ## Sage can solve algebraic equations. # Sage can solve algebraic equations eq1 = x^2 - 2 == x show(eq1) solve([eq1],x) # Cubic polynomials are just as easy eq2 = x^3 + 2*x^2 - 5*x - 6 == 0 show(eq2) solve([eq2],x) # Sage can even find the ugly answers eq3 = x^3 - 2 == x show(eq3) solve([eq3],x) ## At this point we need to pause and look a little more carefully at how Sage works with Python. First, Sage makes extensive use of classes # Pressing tab will give auto-complete options, or function descriptions after a parens ( # A '?' after a command gives detailed information, e.g. 'plot?' tells you about plot # Sage can do simple calculations # What is pi? type(pi) # How do we tell Sage to compute pi to 4493 digits if we don't remember the command? # The tab key after "." of "(" generates a help display which can be used to # autocomplete the statement. pi.n(4493) # Sage can solve equations for one variable in terms of others. # var('x b c') show(solve([x^2 + b*x + c == 0],x)) ## ERROR?? This example illustrated the fact that Python variables by default are "programming variables" and must be initialized. To create a symbolic variable we must declare the variable with the var command. The variable x is a special case for convenience. # Sage can solve equations for one variable in terms of others. var('x b c') show(solve([x^2 + b*x + c == 0],x)) # Sage can even find the ugly answers eq3 = x^3 - 2 == x show(eq3) sol1 = solve([eq3],x) for k in range(3): show(sol1[k]) This example illustrated the for loop in Python. First, the for statement walking through a list of objects and it ends with the colon. Second, the block inside the loop is indented.  There are no begin ... end constructs in Python. Good programming style strongly suggests that you indent subordinate blocks of code. Python insists that you indent and rewards you with shorter programs. # Sage can solve linear systems of algebraic equations # Note the introduction of the symbolic variable y var('y') eq4 = x-1 == y+1 eq5 = x+1 == 2*(y-1) eq4;eq5 This example underscored the fact that the single "=" is assignment and the double "==" is equality. It also showed that multiple statements can be on the same line when separated by ";" sol2 = solve([eq4,eq5],x,y) sol2 In this example, the solution is returned in a list.  Lists are the principal data structure workhorse. # Sage can also return the solutions in a dictionary # This provides access by name sol2 = solve([eq4,eq5],x,y,solution_dict=True) sd = sol2[0] print sd (sd[x].n(digits=2),sd[y].n(digits=3)) In this example we used the dictionary, which is the most sophisticated built-in data structure in Python. ## Sage's graphics capabilities can provide helpful illustrations. A simple nonlinear example consisting of a straight  line intersecting a circle. eq6 = (x-1)^2 + (y-1)^2 == 4; show(eq6) eq7 = 4*y - 3*x == -1; show(eq7) sol3 = solve([eq6,eq7],x,y,solution_dict=True) show(sol3) [(s[x].n(digits=5),s[y].n(digits=5)) for s in sol3] We can easily create a nice illustration of this problem and its solution. x0 = sol3[0][x] y0 = sol3[0][y] s0 = '(%5.2f,%5.2f)' % (x0,y0) p0 = (x0-0.47,y0) x1 = sol3[1][x] y1 = sol3[1][y] s1 = '(%5.2f,%5.2f)' % (x1,y1) p1 = (x1-0.47,y1) g = Graphics() g += circle((1,1),2) g += line([(-1,-1),(3,2)]) g += points([(s[x],s[y]) for s in sol3],rgbcolor=(1,0,0)) g += text(s0,p0) g += text(s1,p1) show(g,aspect_ratio=1) The following example by Jason Grout uses Sage to solve a system of three non-linear equations with 4 unknowns, where one of the unknowns, $u$, is treated as a parameter and set to 1. # First, we solve the system symbolically: var('x y u v') eq8 = u+v==9; show(eq8) eq9 = v*y+u*x==-6; show(eq9) eq10 = v*y^2+u*x^2==24; show(eq10) show(solve([eq8,eq9,eq10,u==1],u,v,x,y)) # Second, we can solve it numerically solns = solve([eq8,eq9,eq10,u==1],u,v,x,y, solution_dict=True) [(s[u].n(digits=10), s[v].n(digits=10), s[x].n(digits=10), s[y].n(digits=10)) for s in solns] ## Sage can do Calculus p = (x^5-1) diff(p^3,x) integral(p^3,x) A slightly more complicated formula involving trigonometric functions with parameters. var('a b c') q = (x-a)*(cos(b*x)+sin(c*x)) q Differentiating this expression illustrates that Sage knows the product rule, the chain rule, and the derivatives of sin and cos. diff(q,x) Integrating this expression illustrates that Sage knows integration by parts. integral(q,x) Sage can define and manipulate functions. f(x) = sin(x)^2*cos(x)*exp(x) show(f) plot(f, 0, 3) x=4 print f(x) plot(f,0,3) show(f(x^3)) # Surface plot example var('x,y') plot3d(cos(x^2 + y^2), (x, -2, 2), (y, -2, 2))
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.16695918142795563, "perplexity": 5982.7443048165405}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964361169.72/warc/CC-MAIN-20211202054457-20211202084457-00382.warc.gz"}
http://villesalo.com/notes/CAoCAoCSSNotes.html
## Errata for "Computational Aspects of Cellular Automata on Countable Sofic Shifts" The proof of Lemma 1 is sloppy. For a stronger statement for SFTs with a more careful proof, see Lemma 2.1.10 in my PhD thesis. In my thesis instead of countable sofics I work directly with the more general class of subshifts with a bounded language in the sense of formal languages, and deal with non-uniqueness of representations only when needed. A structure theorem for general sofics, which constructs a nice SFT cover, is proved in [Pavlov-Schraudner], but it does not quite give unique presentations. We do not cite [Pavlov-Schraudner] in the paper. We did learn about it somewhere around 2012, but may not have understood how direct the connection between the papers is. The proof of Lemma 1 is not literally correct, because no tuples corresponding to periodic points are ever added to $$T$$ so periodic points have no representations. Even forgiving this, I do not believe what we say is sufficient proof that the representations are truly unique for all points, as it is easy to concoct "deterministic parsing processes" which lead to some points having multiple decompositions (given by not parsing according to the process outlined). The particular process may well work, but there are some technical details that would need clarification. The proof of Theorem 5 lacks some crucial details related to the handling of periodic points. I believe the idea and statement are correct, but that a full proof of would require a separate article. Presumably the statement could also be sharpened, as asymptotic sets need not be closed sets. ## References [Pavlov-Schraudner] Classification of sofic projective subdynamics of multidimensional shifts of finite type, Ronnie Pavlov and Michael Schraudner, Trans. Amer. Math. Soc. 367 (2015), 3371-3421. Last update: 9.8.2017
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5631687045097351, "perplexity": 548.0909931860226}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540584491.89/warc/CC-MAIN-20191214042241-20191214070241-00243.warc.gz"}
https://phantran.net/questions-answers-using-the-apsnd-table/
Now that we are acquainted with the numbers in Table 17.5, we may try to answer the questions we posed before and related ones. A standard normal curve implies normal distribution. This means the data we have on hand should conform, in terms of frequency distribution, reasonably closely to normal distribution. If the population is very large or infinite, a sample of suitable size is collected using an appropriate method (or combination of methods), as described in Chapter 16. Elements in the sample set, in such a case, should conform reasonably to normal distribution. To be able to read the probability values in terms of APSND as a way of answering questions, we need to have the numerical values of just two statistics: the mean (μ) and the standard deviation (σ) of the random variable. Computation of these is fairly easy, as even hand-held calculators are capable of this. Once the computation is done, it is not necessary to have aplot of the data for a nswering the various questions. Here lies the “ready reference” nature  of the standard normal curve, with the associated APSND table. For purposes of illustration, we assume that the data we have on hand are of original population and, further, that the mean and the standard deviation of the random variables are known. Example 1: (Question 1 stated in Section 17.6) Data: Mean (nominal length):                                 1.0 in. Standard deviation, assumed to be known:0.22 X1 = 0.97      X2 = 1.04 1. For Z1 = -1.36, read from table, APSND: 0.4131 For Z2 = 1.81, read from table, APSND: 0.4649 2. Area corresponding to the question: total area on both sides of z = 0, that is, (0.4131 + 0.4649) = 0.8780 Answer: The required probability is 87.8 percent. Example 2: (Question 2 stated in Section 17.6) The length of full-term newborn babies in United States is a matter of national statistics, but we will assume the following data: Following are the steps for the answer: 1. For z1 = -0.57, APSND: 0.2157 For z2 = 1.71, APSND: 0.4564 2. Interpreting the above numbers, the area corre­sponding to the question: the total area on both sides of z = 0, that is, (0.2157 + 0.4564) = 0.672 (approx.) 3. Answer: The required probability is 67.2 percent. We may close this chapter with a reminder that to be able to use this powerful tool, the standard normal probability function, we do not need to have a plotted normal distribution curve. To be able to answer the relevant questions, we only need to know the range of the variable and the corresponding numerical values of ft and O. However, unless these values have been derived from a population of a reasonably large number of elements or from a reasonably large sample when the population is very large, we cannot develop confidence in the answers. Besides, it is impera­tive in the function y = f(x) that the values ofy are such that the lines connecting these offer a reasonable approximation to nor­mal distribution. This needs to be confirmed. Any and every y = fx) relation will not necessarily offer a tendency to normal distri­bution (see Chapter 15). Source: Srinagesh K (2005), The Principles of Experimental Research, Butterworth-Heinemann; 1st edition.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9146698713302612, "perplexity": 662.8976998943076}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662517245.1/warc/CC-MAIN-20220517095022-20220517125022-00241.warc.gz"}
https://homework.cpm.org/category/CC/textbook/CCA2/chapter/Ch12/lesson/12.1.3/problem/12-59
### Home > CCA2 > Chapter Ch12 > Lesson 12.1.3 > Problem12-59 12-59. Expand. 1. $(a + b)^4$ Use the binomial theorem. $a^4+4a^3b+6a^2b^2+4ab^3+b^4$ 1. $(3m - 2)^4$ Use the results of part (a). Let $a = 3m$ and $b = -2$. Simplify. $81m^4 -216m^3 +216m^2 -96m+16$
{"extraction_info": {"found_math": true, "script_math_tex": 6, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9647220373153687, "perplexity": 16712.172743578205}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107877420.17/warc/CC-MAIN-20201021180646-20201021210646-00084.warc.gz"}
https://socratic.org/questions/how-do-i-find-f-theta-for-f-theta-e-sin2theta
Calculus Topics What is the derivative of f(theta)=e^(sin2theta) ? Aug 3, 2014 $f ' \left(\theta\right) = 2 \cos \left(2 \theta\right) {e}^{\sin} \left(2 \theta\right)$ Explanation, let's we have $u \left(\theta\right) = {e}^{g} \left(f \left(\theta\right)\right)$ then, using Chain Rule , $u ' \left(\theta\right) = {e}^{g} \left(f \left(\theta\right)\right) \left(g ' f \left(\theta\right)\right) f ' \left(\theta\right)$ Similarly differentiating the given function with respect to $\theta$, $f \left(\theta\right) = {e}^{\sin} \left(2 \theta\right)$ $f ' \left(\theta\right) = {e}^{\sin} \left(2 \theta\right) \left(\cos \left(2 \theta\right)\right) \left(2\right)$ $f ' \left(\theta\right) = 2 \cos \left(2 \theta\right) {e}^{\sin} \left(2 \theta\right)$ Impact of this question 3572 views around the world
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 7, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8138827085494995, "perplexity": 5112.85177123597}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178361723.15/warc/CC-MAIN-20210228175250-20210228205250-00078.warc.gz"}
http://matchupmedia.com/06hfa/symmetric-relation-in-reasoning-e429fd
Select Page The relation, thus, should be represented in a It is nonsymmetric . [better source needed]. Again < is the only asymmetric relation … Properties are “one-place” or“m… Symmetrical definition is - having, involving, or exhibiting symmetry. A typical $\succeq$ preference relation is the union of an asymmetric and symmetric part. A relation R on a set A is called a partial order relation if it satisfies the following three properties: Relation R is Reflexive, i.e. After 4 hr. Clearly a ~b implies b ~a. Relation prediction for knowledge graphs aims at predicting missing relationships between entities. But you need to understand how, relativelyspeaking, things got started. Notice the previous example illustrates that any function has a relation that is associated with it. Plausibly, our third example is symmetric: it depends a bit on how we read 'knows', but maybe if I know you then it follows that you know me as well, which would make the knowing relation symmetric. ∙ 10 ∙ share . A relation $$R$$ on $$A$$ is said to be symmetricif for all $$x,y \in A$$ $$x\,R\,y$$ if and only if $$y\,R\,x\text{. Formally, a binary relation R over a set X is symmetric if: 2.5 Reasoning Using Properties from Algebra. Relation R is transitive, i.e., aRb and bRc aRc. The symmetric difference of the sets A and B is commonly denoted by ⊖, or ⁡. this reasoning, symmetry is the same as the above (i.e., special relativity), but it appliedto the object different from the constancy of the ... the relation between the length of the three edges must be invariant under dilation. Name the property of equality the statement illustrates. Formally: a binary relation Rover a set Ais called symmetricif the following first-order statement is true about R: ∀a∈ A. Ex 2: Causal relation dictated by temporal relation. Now for a symmetric relation, if (a,b) is present in R, then (b,a) must be present in R. In Matrix form, if a 12 is present in relation, then a 21 is also present in relation and As we know reflexive relation is part of symmetric relation. If A is a symmetrix matrix then A-1 is also symmetric. To begin let’s distinguish between the “degree” or“adicity” or “arity” of relations (see, e.g.,Armstrong 1978b: 75). SYMMETRIC RELATION. A relation has ordered pairs (a,b). Relation prediction for knowledge graphs aims at predicting missing relationships between entities. A relation \(R$$ on $$A$$ is said to be transitiveif for all $$x,y,z \in A\text{,}$$ if $$x\,R\,y$$ and $$y\,R\,z\text{,}$$ then $$x\,R\,z$$ Definition3.2.4Antisymmetric. (aRb→ bRa) MCQ Questions for Class 12 Maths with Answers were prepared based on the latest exam pattern. c) Is the function f : R → R such that f(x) = x 3 bijective? AB =BA, then the product of A and B is symmetric. State of the art reasoning methods are dominated by neu-ral module networks [2]. These relations are called symmetric. Despite the importance of inductive relation prediction, most previous works are limited to a transductive setting and cannot process previously unseen entities. Despite the importance of inductive relation prediction, most previous works are limited to a transductive setting and cannot process previously unseen entities. An example is the relation "is equal to", because if a = b is true then b = a is also true. A car is being driven at a constant rate. Communicative Message Passing for Inductive Relation Reasoning. reasoning from relations both spatial relations (the borough is in New Jersey) and temporal relations (its last local elections were on ... And a relation such as nearest to is neither symmetric nor asym-metric. Examples: •If x= y, then y= x. Thesedistinctions aren’t to be taken for granted. How to use symmetrical in a sentence. a) Determine whether the relation R is reflexive, symmetric, antisymmetric or transitive. To understand the contemporary debate about relations we will need tohave some logical and philosophical distinctions in place. reasoning. of traveling, it had traveled 90 mi. The dense relations are then combined with language features for language guided reasoning. So from total n 2 pairs, only n(n+1)/2 pairs will be chosen for symmetric relation. DistMult [48] models the symmetric relation with a bilinear scoring function. Terminology: Vocabulary for graphs often different from that for relations. So from total n 2 pairs, only n(n+1)/2 pairs will be chosen for symmetric relation. For JEE Main other Engineering Entrance Exam Preparation, JEE Main Physics Relations and Functions Previous Year Questions with Solutions is given below. This post covers in detail understanding of allthese Play this game to review Life Skills. •If x≡ₖy, then y≡ₖx. Relation R is Antisymmetric, i.e., aRb and bRa a = b. b) Is the relation R a bijective function? ∀b∈ A. ComplEx [44] models the asymmetric relations by using a bilinear scoring function in complex space. Since e3:raged is temporally after e4:stifle, e4 should be the cause of e3. 2-congruence (n,r)-congruence. Suppose f: R !R is de ned by f(x) = bx=2c. Let R be a relation defined on the set A. Ans. With respect to symmetric preferences there are not really pure symmetric ones in consumer theory at least, the only symmetric relation is total indifference ~ over the whole choice set. A symmetric relation is a type of binary relation. Part of thedevelopment of the debate has consisted in the refinement of preciselythese distinctions. }\) Definition3.2.3Transitive. Early visual reasoning works [26] generate dense relations for each pair of pix-els. relation is thus e2 being before e1. of traveling, that same car had traveled 180 mi. a b c If there is a path from one vertex to another, there is an edge from the vertex to another. If matrix A is symmetric then A n is also symmetric, where n is an integer. If a relation is Reflexive symmetric and transitive then it is called equivalence relation. In neural module networks, language representation is parsed into a series of logical steps. For each point, state your reasoning in proper sentences. assuming a constant speed during both time periods, what is the speed of the car? Check the below NCERT MCQ Questions for Class 12 Maths Chapter 1 Relations and Functions with Answers Pdf free download. If JK = KL and KL = 12, then JK = 12. m W = m W. Symmetric… A relation can be neither symmetric nor antisymmetric. If m 6 = m 7, then m 7 = m 6. Rather, it is having a deep understanding and an appreciation of the concept and its relationships with a connected web of other concepts, operations and relations. Addition and difference of two symmetric matrices results in symmetric matrix. Reasoning about symmetry is not merely about noticing if a shape is symmetrical. We have provided Relations and Functions Class 12 Maths MCQs Questions with Answers to help students understand the concept very well. For example, TransE [3] defines each relation as a translation vector, which can effectively model the composition and inverse relation patterns. Transitive: A relation R on a set A is called transitive if whenever (a;b) 2R and (b;c) 2R, then (a;c) 2R, for all a;b;c 2A. JEE Main Mathematics Relations, Functions, and Reasoning Previous Year Papers Questions With Solutions. 12/16/2020 ∙ by Sijie Mai, et al. After 2 hr. If A and B are two symmetric matrices and they follow the commutative property, i.e. This is the verbal reasoning questions and answers section on "Blood Relation Test" with explanation for various interview, competitive examination and entrance test. In simple terms, a R b -----> b R a. aRa ∀ a∈A. Symmetry In some relations, the relative order of the objects doesn't matter. If R is symmetric relation, then. asymmetric if the relation is irreversible: ∀(x,y: Rxy) ¬Ryx. Solved examples with detailed answer description, explanation are given and it would be easy to understand. 1 This is the standard notation for citing articles in Peirce s (1931 1958) R = { (a, b), (b, a) / for all a, b ∈ A} That is, if "a" is related to "b", then "b" has to be related to "a" for all "a" and "b" belonging to A. Mir-Hossein Moussavi (e3:raged) after government’s efforts to (e4:stifle) protesters. Given relation R = {(x, y) ∈ Z × Z | y = x 3}. R such that f ( x ) = bx=2c: •If x= y then! If m 6 of pix-els relations are then combined with language features for language guided reasoning edge the... The vertex to another, there is a type of binary relation Rover set! Another, there is a path from one vertex to another, there is edge! Formally: a binary relation 7, then y= x b is symmetric then n... Has consisted in the refinement of preciselythese distinctions covers in detail understanding allthese... On the latest Exam pattern and b is commonly denoted by ⊖, or exhibiting symmetry = 7. Complex space matrix then A-1 is symmetric relation in reasoning symmetric Rover a set Ais called the... ) ¬Ryx b c if there is a symmetrix matrix then A-1 is also symmetric commutative property i.e! Things got started the debate has consisted in the symmetric relation in reasoning of preciselythese distinctions the art methods! What is the union of an asymmetric and symmetric part denoted by ⊖, or.! Irreversible: ∀ ( x ) = x 3 } pairs, only n n+1! Understand the concept very well symmetric and transitive then it is called relation! Transitive, i.e., aRb and bRa a = b Moussavi (:... Guided reasoning ( x, y: Rxy ) ¬Ryx binary relation previous... For graphs often different from that for relations cause of e3 of a symmetric relation in reasoning b is symmetric:... Arb and bRa a = b at a constant speed during both time periods what! Antisymmetric or transitive of inductive relation prediction for knowledge graphs aims at predicting missing relationships between entities then m =... And b are two symmetric matrices results in symmetric matrix, i.e binary relation Rover set., or exhibiting symmetry assuming a constant rate, language representation is parsed into a series of steps... Predicting missing relationships between entities Questions with Solutions during both time periods, what is union. Works are limited to a transductive setting and can not process previously unseen entities Answers were prepared based the., Functions, and reasoning previous Year Questions with Solutions that f ( x, y ∈... By ⊖, or exhibiting symmetry b c if there is a symmetrix matrix then A-1 is also.! Dominated by neu-ral module networks, language representation is parsed into a series of logical steps ( aRb→ )! To another, there is a type of binary relation R is transitive,,! Efforts to ( e4: stifle ) protesters are limited to a setting... Previously unseen entities are given and it would be easy to understand how, relativelyspeaking, things got started steps. State your reasoning in proper sentences combined with language features for language guided reasoning bRa. Set a that f ( x, y ) ∈ Z × |... The debate has consisted in the refinement of preciselythese distinctions and transitive then is! S efforts to ( e4: stifle ) protesters the concept very.! Art reasoning methods are dominated by neu-ral module networks symmetric relation in reasoning 2 ] whether relation. Commonly denoted by ⊖, or exhibiting symmetry Preparation, JEE Main other Engineering Entrance Preparation... In the refinement of preciselythese distinctions for knowledge graphs aims at predicting missing relationships between entities each pair of.. Of inductive relation prediction for knowledge graphs aims at predicting missing relationships between entities understand how,,! Knowledge graphs aims at predicting missing relationships between entities only n ( n+1 ) pairs... Previous example illustrates that any function has a relation is the union of asymmetric... Has ordered pairs ( a, b ) is the relation is Reflexive symmetric and transitive it... $\succeq$ preference relation is irreversible: ∀ ( x ) = x 3 bijective ( a, )! Examples with detailed answer description, explanation are given and it would be easy to understand,! Brc aRc or transitive car had traveled 180 mi setting and can not process unseen. 1 relations and Functions previous Year Questions with Solutions is given below symmetric relation in reasoning other Engineering Exam. Is not merely about noticing if a and b is commonly denoted by ⊖, or exhibiting symmetry 2. B is symmetric Causal relation dictated by temporal relation, i.e given below by! Point, state your reasoning in proper sentences a bijective function b is commonly denoted by,! Solutions is given below, symmetric, Antisymmetric or transitive statement is true about R ∀a∈... Car is being driven at a constant speed during both time periods, what the. And they follow the commutative property, i.e 48 ] models the symmetric difference two... Preciselythese distinctions and reasoning previous Year Papers Questions with Solutions it is called equivalence relation using a bilinear scoring.. Asymmetric if the relation is a path from one vertex to another, there is an integer the previous illustrates. Help students understand the concept very well dominated by neu-ral module networks [ 2 ] and symmetric.... Between entities stifle ) protesters: R! R is de ned by f x... In symmetric matrix covers in detail understanding of allthese Early visual reasoning works [ 26 ] generate relations... A bijective function ) protesters is temporally after e4: stifle, e4 should be the cause of e3 logical.: a binary relation Rover a set x is symmetric then a n is integer! 3 bijective constant speed during both time periods, what is the speed of the debate has in! I.E., aRb and bRa a = b networks, language representation is into... Often different from that for relations b c if there is a path from one vertex to another state the! Into a series of logical steps let R be a relation has ordered (... Previous Year Papers Questions with Solutions properties are “ one-place ” or “ m… relation prediction knowledge... N 2 pairs, only n ( n+1 ) /2 pairs will be chosen symmetric. Then y= x, Functions, and reasoning previous Year Papers Questions with Answers were prepared based on the a! Entrance Exam Preparation, JEE Main Mathematics relations, Functions, and reasoning previous Year Papers with... Involving, or exhibiting symmetry ( e3: raged is temporally after e4: stifle ) protesters definition -... Constant speed during symmetric relation in reasoning time periods, what is the relation R over set... Of thedevelopment of the sets a and b is commonly denoted by ⊖, or ⁡ taken! Knowledge graphs aims at predicting missing relationships between entities not process previously unseen entities Main Physics relations Functions... Or transitive m… relation prediction, most previous works are limited to a transductive setting and can not previously... 48 ] models the symmetric difference of the sets a and b is commonly denoted by,... X 3 bijective part of thedevelopment of the art reasoning methods are dominated by neu-ral module networks [ ]! ( e3: raged ) after government ’ s efforts to ( e4:,., then y= x each pair of pix-els b ) is the function f: R! is... Solutions is given below other Engineering Entrance Exam Preparation, JEE Main other Engineering Exam! Aren ’ t to be taken for granted ( aRb→ bRa ) Notice the previous example illustrates that function... Suppose f: R! R is Antisymmetric, i.e., aRb bRa! A set x is symmetric be the cause of e3 understanding of allthese Early visual works... Help students understand the concept very well R over a set Ais symmetricif! The function f: R → R such that f ( x, y ∈! Exam Preparation, JEE Main Physics relations and Functions Class 12 Maths with Answers prepared! State of the debate has consisted in the refinement of preciselythese distinctions with Solutions is given below ( x =... Stifle ) protesters in the refinement of preciselythese distinctions process previously unseen entities constant during. Denoted by ⊖, or ⁡ guided reasoning f ( x, y ) ∈ Z × Z y! And b is commonly denoted by ⊖, or exhibiting symmetry thesedistinctions aren t! 48 ] models the asymmetric relations by using a bilinear scoring function over a x... Relation prediction, most previous works are limited to a transductive setting and can process... Noticing if a relation that is associated with it b R a bijective function to be for! Is also symmetric, Antisymmetric or transitive: Causal relation dictated by temporal relation commutative!, e4 should be the cause of e3 different from that for relations irreversible! = m 7 = m 7, then the product of a and b is symmetric if: symmetric.... Taken for granted Answers to help students understand the concept very well to (:... ) is the union of an asymmetric and symmetric part of two symmetric matrices and follow. The union of an asymmetric and symmetric part about R: ∀a∈ a part of thedevelopment of the?. Visual reasoning works [ 26 ] generate dense relations for each point, state your reasoning in proper sentences associated... Symmetric and transitive then it is called equivalence relation after e4: stifle, e4 should be the cause e3! A path from one vertex to another relativelyspeaking, things got started has relation! A bijective function then a n is also symmetric: stifle, e4 be! Generate dense relations are then combined with language features for language guided reasoning s to. Asymmetric relations by using a bilinear scoring function in complex space are “ one-place ” or “ relation! 3 bijective product of a and b is commonly denoted by ⊖, exhibiting... ## Let's Get Started Want The Unfair Advantage Of High Quality Digital Marketing At An Affordable Price? Let's not waste more time, send us your name and email and our representative will reach out as soon as possible! Or schedule a FREE call immediatly to save even more time ## The Ultimate Guide To Level-Up Your E-Comm Store Experts Reveal Their Secrets! This guide will give you the tried and tested strategies that will skyrocket your sales! Want to know more? Schedule a FREE Strategy call immediatly to save even more time
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8351975083351135, "perplexity": 1860.9825216308227}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662588661.65/warc/CC-MAIN-20220525151311-20220525181311-00425.warc.gz"}
http://oak.snr.missouri.edu/nr3110/topics/schnabel.php
## Mark and Recapture Methods ### Schnabel Method: (Multiple marking) The Schnabel method is an extension of the Petersen method to a series of samples.  Each sampling event includes capture, examination for previous marks, marking and then release.  The marking time is not need so only one type of mark is needed.  For each sample we need to record the following data: • ct= Total number of individuals caught in samplet • Rt = Number of individuals already marked when caught in sample t • ut= Number of individuals marked for the first time and released in sample t , Accidental deaths are subtracted. Marks accumulate over time. So • Mt = the number marked individual in the population just before the current sample. or as a formula: $$M_t = \sum_{i=1}^{t-1} U_i$$ where S is the number of the sample in the series. Schnabel (1938) proposed the following weighted average of  Petersen estimates as a estimate of population density. $$\hat{N} = \frac{\sum_{t=1}^{S} (C_t M_t)}{\sum_{t=1}^{S} R_t}$$ If the each sample catch and the marked  population are less that 10 % of the population size the following is a better estimator of the population size. $$\hat{N} = \frac{\sum_{t=1}^{S} (C_t M_t)}{\left( \sum_{t=1}^{S} R_t\right) + 1}$$ where: S is the number of samples in the series. The variance estimator is calculated on the reciprocal of the population density 1/N as: $$Variance \left( \frac{1}{\hat{N}} \right) = \frac{\sum_{t=1}^{S} R_t}{\left( \sum_{i=1}^{S} (C_t M_t) \right)^2 }$$ And the standard error of the reciprocal population density is: $$Standard \space error \left( \frac{1}{\hat{N}} \right) = \sqrt{Variance \left( \frac{1}{\hat{N}} \right)}$$ Schumacher and Eschmeyer (1943) pointed out that one could use a regression a slope of 1/N passing through th origin.  A formula using the regression methodology is: $$\hat{N} = \frac{\sum_{t=1}^{S} (C_t M_t^2)}{\sum_{t=1}^{S} R_t M_t}$$ The variance of this density estimate can be estimated from the variance estimator or the slope parameter estimate (Zar, 1996) as: $$Variance \space of \frac{1}{\hat{N}} = \frac{ \sum_{t=1}^{S} \left( \frac{R_t^2}{C_t}\right) - \frac{ \left( \sum_{t=1}^S (R_t M_t) \right)^2}{\sum_{t=1}^S (C_t M_t^2)} }{s - 2}$$ And the standard error of the population density estimate can be determined by: $$Standard \space error \left( \frac{1}{\hat{N}} \right) = \sqrt{Variance \frac{\frac{1}{\hat{N}}}{\sum_{t=1}^2 (C_t M_t^2) } }$$ Confidence Intervals:  If the total number of recapture in < 50 the confidence limits for the population estimate should be obtained from a Poisson distribution (Krebs 1989, Appendix 1.2).   If the number of recaptures is >50 use the normal approximation as follows: $$\frac{1}{\hat{N}} \pm t_{\alpha} s_\bar{x}$$ where: S.E. is the standard error of  1/N t alpha is the value from Student's t-table for (100 - alpha) % confidence limits. Assumptions: • Assumptions of the Petersen method apply. • In the Schnabel method it is easier to pick up violation of the assumptions. Also See: Chapter 2 - Estimating Abundance: Mark-and-Recapture pages 30-37 in: Krebs, C. J. 1998. Ecological Methodology. Harper and Row, Publishers. New York. 654 pp. Schnabel, Z. E. 1938. The estimation of total fish populations of a lake. Am. Math. Monthly 45:348-352. Schumacher, F. X. and Eschmeyer, R. W. 1943. The estimation of fish populations in lakes and ponds. J. Tennessee Acad. Sci. 18:228-249. Zar, J. H. 2007. Biostatistical Analysis. Prentice-Hall, Inc. Englewood Cliffs, New Jersey. 718 pp.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9356564283370972, "perplexity": 2545.989045140871}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084893629.85/warc/CC-MAIN-20180124090112-20180124110112-00406.warc.gz"}
http://www.reference.com/browse/Constructive+Type+Theory
Definitions # Intuitionistic type theory Intuitionistic type theory, or constructive type theory, or Martin-Löf type theory or just Type Theory is a logical system and a set theory based on the principles of mathematical constructivism. Intuitionistic type theory was introduced by Per Martin-Löf, a Swedish mathematician and philosopher, in 1972. Martin-Löf has modified his proposal a few times; his early, impredicative formulations were inconsistent as demonstrated by Girard's paradox, and later formulations were predicative. He also proposed extensional and then intensional variants of intuitionistic type theory. Intuitionistic type theory is based on a certain analogy or isomorphism between propositions and types: a proposition is identified with the type of its proofs. This identification is usually called the Curry–Howard isomorphism, which was originally formulated for propositional logic and simply typed lambda calculus. Type Theory extends this identification to predicate logic by introducing dependent types, that is types which contain values. Type Theory internalizes the interpretation of intuitionistic logic proposed by Brouwer, Heyting and Kolmogorov, the so called BHK interpretation. The types of Type Theory play a similar role to sets in set theory but functions definable in Type Theory are always computable. ## Connectives of Type Theory In the context of Type Theory a connective is a way of constructing types, possibly using already given types. The basic connectives of Type Theory are: ### Π-types $Pi$-types, also called dependent product types, are analogous to the indexed products of sets. As such, they generalize the normal function space to model functions whose result type may vary on their input. E.g. writing $mbox\left\{Vec\right\}\left(\left\{mathbb R\right\},n\right)$ for $n$-tuples of real numbers, $Pi n:\left\{mathbb N\right\}.mbox\left\{Vec\right\}\left(\left\{mathbb R\right\},n\right)$ stands for the type of functions which given a natural number returns a tuple of real numbers of this size. The usual function space arises as a special case when the range type does not actually depend on the input, e.g. $Pi n:\left\{mathbb N\right\}.\left\{mathbb R\right\}$ is the type of functions from natural numbers to the real numbers, which is also written as $\left\{mathbb N\right\}to\left\{mathbb R\right\}$. Using the Curry-Howard isomorphism $Pi$-types also serve to model implication and universal quantification: e.g. a term inhabiting $Pi m,n:\left\{mathbb N\right\}.m+n = n+m$ is a function which assigns to any pair of natural numbers a proof that addition is commutative for that pair and hence can be considered as a proof that addition is commutative for all natural numbers. ### Σ-types $Sigma$-types, also called dependent sum types, are analogous to the indexed disjoint unions of sets. As such, they generalize the usual Cartesian product to model pairs where the type of the 2nd component depends on the first. E.g. the type $Sigma n:\left\{mathbb N\right\}.mbox\left\{Vec\right\}\left(\left\{mathbb R\right\},n\right)$ stands for the type of pairs of a natural number and a tuple of real numbers of that size, i.e. this type can be used to model sequences of arbitrary length (usually called lists). The conventional Cartesian product type arises as a special case when the type of the 2nd component doesn't actually depend on the first, e.g. $Sigma n:\left\{mathbb N\right\}.\left\{mathbb R\right\}$ is the type of pairs of a natural number and a real number, which is also written as $\left\{mathbb N\right\}times\left\{mathbb R\right\}$. Again, using the Curry-Howard isomorphism, $Sigma$-types also serve to model conjunction and existential quantification. ### Finite types Of special importance are $0$ (the empty type), $1$ (the unit type) and $2$ (the type of Booleans or classical truth values). Invoking the Curry-Howard isomorphism again, $0$ stands for False and $1$ for True. Using finite types we can define negation as $neg A = A to 0$. ### Equality type Given $a,b : A$ then $a = b$ is the type of equality proofs that $a$ is equal to $b$. There is only one (canonical) inhabitant of $a = b$ and this is the proof of reflexivity $mbox\left\{refl\right\} : Pi a:A.a = a$. ### Inductive types A prime example of an inductive type is the type of natural numbers $mathbb N$ which is generated by $0 : \left\{mathbb N\right\}$ and $mbox\left\{succ\right\} :\left\{mathbb N\right\} to \left\{mathbb N\right\}$. An important application of the propositions as types principle is the identification of (dependent) primitive recursion and induction by one elimination constant: $\left\{mathbb N\right\}-mbox\left\{elim\right\} : P\left(0\right) to \left(Pi n:\left\{mathbb N\right\}.P\left(n\right)to P\left(mbox\left\{succ\right\}\left(n\right)\right)\right)toPi n:\left\{mathbb N\right\}.P\left(n\right)$ for any given type $P\left(n\right)$ indexed by $n:\left\{mathbb N\right\}$. In general inductive types can be defined in terms of W-types, the type of well-founded trees. An important class of inductive types are inductive families like the type of vectors $mbox\left\{Vec\right\}\left(A,n\right)$ mentioned above, which is inductively generated by the constructors $mbox\left\{vnil\right\}:mbox\left\{Vec\right\}\left(A,0\right)$ and $mbox\left\{vcons\right\}:AtoPi n:\left\{mathbb N\right\}.mbox\left\{Vec\right\}\left(A,n\right)tombox\left\{Vec\right\}\left(A,mbox\left\{succ\right\}\left(n\right)\right)$. Applying the Curry-Howard isomorphism once more, inductive families correspond to inductively defined relations. ### Universes An example of a universe is $U_0$, the universe of all small types, which contains names for all the types introduced so far. To every name $a:U_0$ we associate a type $El\left(a\right)$, its extension or meaning. It is standard to assume a predicative hierarchy of universes: $U_n$ for every natural number $n:\left\{mathbb N\right\}$, where the universe $U_\left\{n+1\right\}$ contains a code for the previous universe, i.e. we have $u_n:U_\left\{n+1\right\}$ with $El\left(u_n\right)=U_n$. This hierarchy is often assumed to be cumulative, that is the codes from $U_n$ are embedded in $U_\left\{n+1\right\}$. Stronger universe principles have been investigated, i.e. super universes and the Mahlo universe. In 1992 Huet and Coquand introduced the calculus of constructions, a type theory with an impredicative universe, thus combining Type Theory with Girard's System F. This extension is not universally accepted by Intuitionists since it allows impredicative, i.e. circular, constructions, which are often identified with classical reasoning. ## Formalisation of Type Theory Type Theory is usually presented as a dependently typed lambda calculus, using the judgements: • $vdash Gamma, mbox\left\{Context\right\}$, $Gamma$ is a well-formed context of typing assumptions. • $Gammavdash sigma, mbox\left\{Type\right\}$, $sigma$ is a well-formed type in context $Gamma$. • $Gammavdash t : sigma$, $t$ is a well-formed term of type $sigma$ in context $Gamma$. • $Gammavdash sigmaequivtau$, $sigma$ and $tau$ are equal types in context $Gamma$. • $Gammavdash t equiv u: sigma$, $t$ and $u$ are equal terms of type $sigma$ in context $Gamma$. Of special importance is the conversion rule, which says that given $Gammavdash t : sigma$ and $Gammavdash sigmaequivtau$ then $Gammavdash t : tau$. ## Categorical models of Type Theory Using the language of category theory, Seely introduced the notion of a locally cartesian closed category (LCCC) as the basic model of Type Theory. This has been refined by Hofmann and Dybjer to Categories with Families or Categories with Attributes based on earlier work by Cartmell. A category with families is a category C of contexts (in which the objects are contexts, and the context morphisms are substitutions), together with a functor T : C^op -> Fam(Set). Fam(Set) is the category of families of Sets, in which objects are pairs (A,B) of an "index set" A and a function B: X -> A, and morphisms are pairs of functions f : A -> A' and g : X -> X', such that B' ° g = f ° B - in other words, f maps Ba to B'g(a). The functor T assigns to a context G a set Ty(G) of types, and for each A : Ty(G), a set Tm(G,A) of terms. The axioms for a functor require that these play harmoniously with substitution. Substitution is usually written in the form Af or af, where A is a type in Ty(G) and a is a term in Tm(G,A), and f is a substitution from D to G. Here Af : Ty(D) and af : Tm(D,Af). The category C must contain a terminal object (the empty context), and a final object for a form of product called comprehension, or context extension, in which the right element is a type in the context of the left element. If G is a context, and A : Ty(G), then there should be an object (G,A) final among contexts D with mappings p : D -> G, q : Tm(D,Ap). A logical framework, such as Martin-Löf's takes the form of closure conditions on the context dependent sets of types and terms: that there should be a type called Set, and for each set a type, that the types should be closed under forms of dependent sum and product, and so forth. A theory such as that of predicative set theory expresses closure conditions on the types of sets and their elements: that they should be closed under operations that reflect dependent sum and product, and under various forms of inductive definition. ## Extensional versus intensional A fundamental distinction is extensional vs intensional Type Theory. In extensional Type Theory definitional (i.e. computational) equality is not distinguished from propositional equality, which requires proof. As a consequence type checking becomes undecidable. In contrast in intensional Type Theory type checking is decidable, but the representation of many mathematical concepts is non-standard due to a lack of extensional reasoning. It is a subject of current discussion whether this tradeoff is unavoidable and whether the lack of extensional principles in intensional Type Theory is a feature or a bug. ## Implementations of Type Theory Type Theory has been the base of a number of proof assistants, such as NuPRL, LEGO, Coq, and Agda. Recently, dependent types also featured in the design of programming languages such as Dependent ML, Cayenne and Epigram.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 65, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9427672028541565, "perplexity": 502.6359420927286}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1409535923940.4/warc/CC-MAIN-20140909030438-00068-ip-10-180-136-8.ec2.internal.warc.gz"}
http://mathhelpforum.com/trigonometry/162523-trigonometry-ambiguous-case-print.html
# Trigonometry " The Ambiguous Case". • November 8th 2010, 03:04 AM ejaykasai Trigonometry " The Ambiguous Case". So, This is the problem, solve the triangle Quote: 1. A = 35° a = 5 b = 8 So it's gonna be like this right? Quote: - angle A is Acute a < b a__bsinA 5__8sin35° 5>4.59 therefore two triangles. So This is [COLOR="rgb(46, 139, 87)"]my answer for the 1st triangle.[/COLOR] Quote: B = 66.60° c = 8.54 C = 78.4° CORRECT ME IF I'M WRONG.. Can anyone teach me how to solve for the other triangle? :| • November 8th 2010, 04:59 AM earboth Quote: Originally Posted by ejaykasai So, This is the problem, solve the triangle So it's gonna be like this right? So This is my answer for the 1st triangle.[/COLOR] CORRECT ME IF I'M WRONG.. Can anyone teach me how to solve for the other triangle? :| You have to use the Sine rule (what you obviously did) 1. From $\dfrac{\sin(B)}{\sin(35.8^\circle)}=\dfrac85~\impl ies~\sin(B)\approx 0.9177...$. 2. Since $\sin(B) = \sin(180^\circ - B) \implies B\approx 66.6^\circ~\vee~ B \approx 113.4^\circ $ 3. You now can calculate the dimensions of the 2nd triangle.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 2, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9619561433792114, "perplexity": 3158.216311036908}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416400380683.31/warc/CC-MAIN-20141119123300-00219-ip-10-235-23-156.ec2.internal.warc.gz"}
https://discuss.codechef.com/questions/18662/does-codechef-have-any-code-of-conduct
× # Does CodeChef have any Code Of Conduct? 22 4 I am new to CodeChef. I am not sure what can be considered as an improper conduct on this website. I do not want to be penalized for doing something unknowingly which is not the accepted behavior. Is there any such Code of Conduct? asked 23 Jul '13, 21:31 1★anup ♦♦ 2.2k●26●38●42 accept rate: 23% 1 how to register for writing the exam on march 30th? please give me replay (26 Mar '14, 19:54) 11 I would suggest putting the Code of conduct in the Help (FAQ) Section on the Codechef site and providing a link to the same in the Rules section in each contest. answered 24 Jul '13, 14:23 5★n2n_ 1.8k●6●13●19 accept rate: 9% Indeed. We will do so. (24 Jul '13, 16:10) admin ♦♦0★ 8 My sincere advise to everyone please never use ideone. I have invested so much time and my energy to solve sept long challenge 2016 and one stupid brainless guy copied my code as it is and due to that stupid guy my rating goes down to 4K from 10K. His fake name was Rahul Bhatia and his original name was (Hardik Gulati) at Jaypee Institute of Information Technology and his handle is rahul1714. His fb profile is (https://www.facebook.com/smart.hardik.gulati) Beware of such stupid fellows. for proof I m giving my code which i have submitted on 9th Sept https://www.codechef.com/viewsolution/11392815 and he submitted exact code of mine on 15th Sept https://www.codechef.com/viewsolution/11525589 And request codechef please change some rules for plagiarism cases because many innocent peoples have to suffer huge loses because of someunkown useless guy.. :( :( answered 17 Sep '16, 00:49 4★kunnu96 81●3 accept rate: 0% 1 Are we allowed to put our code for past Contests say September 2014 on GitHub??? That's public too but the contest is long gone. answered 09 Nov '14, 17:26 16●1 accept rate: 0% Good Question.. But @animax001 i never used GitHub.. Can you please tell the advantages of doing so ? Thanks in Advance !! :) (10 Nov '14, 01:23) @rishabhprsd7 GitHub is basically a code repository with version control. You can keep your projects there and anyone can see your code unless you have marked the repository private. In US, employers prefer to look into candidate's GitHub profile to see their coding style, various independent projects done etc. Gives a better picture to them. (11 Nov '14, 17:32) @animax001: Yes. You are allowed to do so if it is a past contest. Giving attribute to the problem (having a link to the original problem) will be really good. (18 Nov '14, 22:27) anup ♦♦1★ I push all my code on github but only after contest. for example september 2016 long contest code is here. https://github.com/AshokRajpurohit/karani/tree/master/src/com/ashok/codechef/marathon/year16/sept (17 Sep '16, 19:30) 1 any of the point which you have given above is not violating my any submission to make me guilty. so please kindly check it out this problem and dont deactivate my account and dont drop my rating.... answered 20 Mar '15, 20:12 11●1 accept rate: 0% 1 i did'nt know how its happen"We will be dropping your CYPHER ratings and disqualifying all the submissions of your account on grounds of using false practice during CYPHER NOVEMBER 2016 . We found few of your codes similar with one another user and would like to have an explanation about the same from you" . but i do all the programme by my self.is there is any mean to prove me"like give different programme then make you realize that i am wrong panelized". answered 06 Nov '16, 13:54 11●1 accept rate: 0% 0 Really?? Is obfuscated code a sin?? :P answered 23 Jul '13, 22:36 4.1k●5●23●64 accept rate: 14% 4 No it is not a sin. But we do not see any reason behind obfuscation in a competitive programming environment. It also goes against the spirit of learning. Others will not be able to understand your code if you obfuscate it. (23 Jul '13, 22:51) admin ♦♦0★ 4 Right! Didn't give any second thought to it. :-) And yeah! (From the other post) Obfuscating to escape being caught is a real sin!! :) (23 Jul '13, 23:07) I am not able to see 2 problems in NOV13 SPOTWO and CHEFGM . It shows i am not authorized to see them ? what happened there. I just use the inline function to take input fast from http://discuss.codechef.com/questions/7394/help-on-fast-inputoutput . is it also not allowed ? rest of code is mine only .. Please replyy.. (06 Nov '13, 14:53) @shubham1402 Cheating cases are caught only after a contest is over. The not authorized comment is due to some server error and has nothing to do with cheating. Hopefully it will be fixed soon. (06 Nov '13, 14:55) gautam943★ 3 @shubham1402 your comment is off-topic in this thread, please read the forum too before posting something... (06 Nov '13, 14:58) 0 I sent a mail to help@codechef.com asking to deactivate my old codechef account. It was around a week back and the account is still active. Please do the needful. answered 11 Nov '13, 14:20 279●3●5●12 accept rate: 8% 0 0 any of the point which you have given above is not violating my any submission to make me guilty. so please kindly check it out this problem and dont deactivate my account and dont drop my rating.... answered 20 Mar '15, 20:11 11●1 accept rate: 0% 0 #include #include #define getcx getchar_unlocked inline void inp( int *n )//fast input function { *n=0; int ch=getcx();int sign=1; while( ch < '0' || ch > '9' ){if(ch=='-')sign=-1; ch=getcx();} while( ch >= '0' && ch <= '9' ) *n = (*n<<3)+(*n<<1) + ch-'0', ch=getcx(); *n=*n*sign; } inline void inplong( long *n )//fast input function { *n=0; int ch=getcx();int sign=1; while( ch < '0' || ch > '9' ){if(ch=='-')sign=-1; ch=getcx();} while( ch >= '0' && ch <= '9' ) *n = (*n<<3)+(*n<<1) + ch-'0', ch=getcx(); *n=*n*sign; } int main() { int i,T; long j,N,val; long long int cnt,cnt2; inp(&T); while(T--) { cnt=cnt2=0; //scanf("%ld",&N); inplong(&N); while(N--) { inplong(&val); // scanf("%ld",&val); if(val==2){ cnt2++; } else if(val>1){ cnt++; } } cnt+=cnt2; printf("%lld\n",((cnt*(cnt-1))/2)-((cnt2*(cnt2-1))/2)); } return 0; } answered 01 Apr '15, 21:58 1 accept rate: 0% 0 I am new user and can't understand what is the problem to solve in the compitition. Please help me in this aspect answered 05 May '15, 11:52 1 accept rate: 0% 0 Am new user sandhya ....right now Media.net off campus s gng on but i dono hw to do in this ....am confused plz help me answered 30 Mar '16, 15:47 1 accept rate: 0% 0 Passing it off as your own is. If you are taking your code from some other source, it is expected that you give the due attribution to the source in your code. It is kind of mandatory. Do not ask or discuss any aspect of any CodeChefs' problem during an ongoing contest on any other platform, online or offline. for c programming training institute please visit the following link c programming answered 14 Jul '16, 12:33 1 accept rate: 0% 0 O.k I understood but I am a new on this site so I am not understood where I get the problems of cc and c++ in this site answered 08 Oct '16, 19:44 1 accept rate: 0% 0 yes. if u break any rule,result will reflect in ur profile,so be carefull,dont do cheating keep learning... answered 17 Mar, 17:48 2★vivek96 508●1●7 accept rate: 7% 0 @admin, please if you can suspend my account. answered 18 Mar, 10:47 0★hafiz45 1 accept rate: 0% 0 How to register this competion? please share the link if you know? answered 01 Jun, 10:40 1 accept rate: 0% 0 I would suggest to upload some video tutorials of solution of contest problems after contest,so that one can fathom solution which he/she couldn't be able to do it. Thank you! answered 13 Jun, 00:44 3★dhruv42 11 accept rate: 0% -1 what is the average time , each question can take to be solved during the contest ? answered 20 Aug '14, 00:24 -1 accept rate: 0% toggle preview community wiki: Preview By Email: Markdown Basics • *italic* or _italic_ • **bold** or __bold__ • image?![alt text](/path/img.jpg "title") • numbered list: 1. Foo 2. Bar • to add a line break simply add two spaces to where you would like the new line to be. • basic HTML tags are also supported • mathemetical formulas in Latex between \$ symbol Question tags: ×1,557 ×457 ×167 ×11 question asked: 23 Jul '13, 21:31 question was seen: 80,389 times last updated: 13 Jun, 00:44
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3668212294578552, "perplexity": 6847.639685964238}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886104172.67/warc/CC-MAIN-20170817225858-20170818005858-00031.warc.gz"}
https://doubtnut.com/jee-mains-solutions/class-12/three-dimensional-geometry/MTEzODg=
X Doubtnut Math Doubt App Click photo & get instant video solution COURSE EXAM STUDY MATERIALS JEE MAINS PREVIOUS YEAR 2007 THREE DIMENSIONAL GEOMETRY Class 12th Maths Solutions Solutions for THREE DIMENSIONAL GEOMETRY Class 12th JEE MAINS PREVIOUS YEAR 2007 Let L be the line of intersection of the planes 2x+3y+z=1 and x+3y+2z=2 . If L makes an angles alpha with the positive x-axis, then cos alpha equals Get here the Maths video solutions of all the questions of of NCERT book. We provide you step by step solution for all the questions given in NCERT Maths book as per CBSE guidelines .The topics and subtopics of are as follows - Click on the link to Learn any Topic for Free: INTRODUCTION DIRECTION COSINES AND DIRECTION RATIOS ANGLE BETWEEN TWO LINES EQUATION OF LINE PASSING THROUGH TWO GIVEN POINT INTERSECTION OF TWO LINES PERPENDICULAR DISTANCE OF A POINT FROM A LINE SHORTEST DISTANCE BETWEEN TWO STARIGHT LINES PLANE ANGLE BETWEEN TWO PLANES DISTANCE BETWEEN PLANE DISTANCE BETWEEN PARALLEL PLANES DISTANCE OF A POINT FROM A PLANE IMAGE OF POINT (REFLECTION) LINE AND PLANE ANGLE BETWEEN LINE AND PLANE ANGLE BETWEEN A LINE AND A PLANE INTERSECTION POINT OF LINE AND PLANE CONDITION FOR A LINE TO LIE IN A PLANE DISTANCE BETWEEN SKEW LINES EQUATION OF SPHERE SIMILAR VIDEOS Latest from Doubtnut
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8017327189445496, "perplexity": 4134.811516623693}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247496855.63/warc/CC-MAIN-20190220230820-20190221012820-00070.warc.gz"}
http://mathhelpforum.com/calculus/21606-numerical-derivative.html
# Math Help - Numerical Derivative 1. ## Numerical Derivative If you know numerical derivative, that is what I'm finding. If not, that's okay. I can translate. Find the NDER of [tex]x^(3/5) Also x^(2/3) .... but I think learning the top will be enough to grasp the concept, unless there is something else worth mentioning for this. Thanks! 2. Originally Posted by Truthbetold If you know numerical derivative, that is what I'm finding. If not, that's okay. I can translate. Find the NDER of [tex]x^(3/5) Also x^(2/3) .... but I think learning the top will be enough to grasp the concept, unless there is something else worth mentioning for this. Thanks! wat do u mean by numerical derivative? if im not wrong, are you looking for the answer $\frac{3}{5}x^{\frac{-2}{5}}$ 3. if im not wrong, are you looking for the answer \ That looks correct, but I don't know how to do this. NUmerical derivative is this: $\frac {f(a+.001)- f(a - .001)}{.002}$ If you don't get it, just find the derivative. I don't know how to deal with (a + 0.001)^(3/5) It isn't shortcuts, meaning the power rule and stuff, because they are next lesson. Thanks! 4. Originally Posted by Truthbetold .. NUmerical derivative is this: $\frac {f(a+.001)- f(a - .001)}{.002}$ .. I don't know how to deal with (a + 0.001)^(3/5) It isn't shortcuts, meaning the power rule and stuff, because they are next lesson. Thanks! where did $\frac {f(a+.001)- f(a - .001)}{.002}$ come from? i just used power rule.. 5. I am taught the power rule next lesson. You asked for the numerical derivative, and that is what that is. I need to use that to figure out the equation. I prefer the power rule also, but I can't use it. I need to eventually know this for the test as well. 6. Originally Posted by Truthbetold I am taught the power rule next lesson. You asked for the numerical derivative, and that is what that is. I need to use that to figure out the equation. I prefer the power rule also, but I can't use it. I need to eventually know this for the test as well. ok, i just dont get where a, +- 0.001, and 0.002 came from.. sorry.. 7. Originally Posted by Truthbetold If you know numerical derivative, that is what I'm finding. If not, that's okay. I can translate. Find the NDER of [tex]x^(3/5) Also x^(2/3) .... but I think learning the top will be enough to grasp the concept, unless there is something else worth mentioning for this. Thanks! If I am correct your teacher is expecting you to approximate the value of the derivative of a function at a given point. $f^{\prime}(x) \approx \frac{f(x + 0.001) - f(x - 0.001)}{0.002}$ where we are approximating the formula $f^{\prime}(x) \approx \frac{f(x + \Delta h) - f(x - \Delta h)}{2 \Delta h}$ (If the 2 in the denominator looks a little funny remember that the denominator is the size of the interval over which we are applying the secant function.) So for $f(x) = x^{3/5}$ we get $f^{\prime}(x) \approx \frac{(x + 0.001)^{3/5} - (x - 0.001)^{3/5}}{0.002}$ This result only has value if you know the general form for the binomial theorem: $(x + \Delta h)^n = x^n + n \Delta h x^{n - 1} + ~ ...$ (The rest of the terms don't matter as $\Delta h$ is small.) So $f^{\prime}(x) \approx \frac{(x + 0.001)^{3/5} - (x - 0.001)^{3/5}}{0.002} \approx$ $\frac{x^{3/5} + \frac{3}{5} \cdot 0.001x^{3/5 - 1} - x^{3/5} + \frac{3}{5} \cdot 0.001x^{3/5 - 1}}{0.002}$ $= \frac{3}{5} \cdot \frac{2 \cdot 0.001 x^{-2/5}}{2 \cdot 0.001}$ $= \frac{3}{5}x^{-2/5}$ (If anyone objects to the concept of using the approximation form of the binomial theorem, which I believe is gotten by using differential Calculus, I agree that this makes my argument rather circular. I simply couldn't find a way to turn the approximate derivative into a function, rather than something that needed to be graphed point by point.) -Dan
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 14, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8992211818695068, "perplexity": 678.8325317431322}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257824230.71/warc/CC-MAIN-20160723071024-00263-ip-10-185-27-174.ec2.internal.warc.gz"}
https://export.arxiv.org/abs/gr-qc/0311070
gr-qc (what is this?) # Title: Gravitational Waves and the Sagnac Effect Abstract: We consider light waves propagating clockwise and other light waves propagating counterclockwise around a closed path in a plane (theoretically with the help of stationary mirrors). The time difference between the two light propagating path orientations constitutes the Sagnac effect. The general relativistic expression for the Sagnac effect is discussed. It is shown that a gravitational wave incident to the light beams at an arbitrary angle will not induce a Sagnac effect so long as the wave length of the weak gravitational wave is long on the length scale of the closed light beam paths. The gravitational wave induced Sagnac effect is thereby null. Comments: LaTeX format 1 *.eps figure Subjects: General Relativity and Quantum Cosmology (gr-qc) Cite as: arXiv:gr-qc/0311070 (or arXiv:gr-qc/0311070v1 for this version) ## Submission history From: Sivasubramanian Somu [view email] [v1] Fri, 21 Nov 2003 03:15:37 GMT (10kb) Link back to: arXiv, form interface, contact.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8511795997619629, "perplexity": 1807.415830082097}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945218.30/warc/CC-MAIN-20230323225049-20230324015049-00255.warc.gz"}
https://www.physicsforums.com/threads/electric-field-net-charge-on-a-cylindrical-shell.380633/
# Electric field / net charge on a cylindrical shell 1. Feb 22, 2010 ### csteff 1. The problem statement, all variables and given/known data A cylindrical shell of radius 8.1 cm and length 247 cm has its charge density uniformly distributed on its surface. The electric field intensity at a point 20.3 cm radially outward from its axis (measured from the midpoint of the shell ) is 41800 N/C. Given: ke =8.99×10^9 N·m^2 /C^2 . What is the net charge on the shell? Answer in units of C. 2. Relevant equations Gauss's Law 3. The attempt at a solution I've been using E=$$\lambda$$/2$$\pi$$$$\epsilon$$r I used the radius from the midpoint (20.3 cm) in place of the r in the equation and used the given E to solve for $$\lambda$$. Then I found the area of the cylindrical shell by doing 2.47*2$$\pi$$*.081+2$$\pi$$*.081 and multiplied that by the $$\lambda$$. I ended up with an answer of 6.1288E-7 the last time I did it. Is the r supposed to be distance from the surface, maybe? or am I using the completely wrong equation? Thank you! Last edited: Feb 22, 2010 2. Feb 22, 2010 ### collinsmark Are you supposed to approximate the answer using Gauss' law, or are you supposed to find the exact answer through detailed integration? (in other words, which section of your textbook is this problem given? And does your course require lots of calculus?) Either way, Go back to the original definition of Gauss' law, that gives you a relationship in terms of $$Q_{enc} / \epsilon _0$$. That way, you don't need to drag $$\lambda$$ into things (the question asks for the total charge, not the charge per unit length). If the particular problem is really looking for an approximation, then that can be done fairly easily. You don't know what the electric field is at the end-caps of the Gaussian surface, but I'm guessing that you are supposed to assume that the electric field component, parallel to the end-caps, contribution is minimal. In other words, I'm guessing that you don't need to include the end caps of the Gaussian surface into the equation. On the other hand, if you need the exact solution, you will need to integrate over the entire surface, without using the approximation that that the electric field is constant over most of the Gaussian cylinder (I'm guessing that you probably are allowed to make the approximation, but that's just my guess). And no, r is the center of the cylindrical shell, which also the center of the Gaussian surface too. Last edited: Feb 22, 2010 3. Feb 22, 2010 ### csteff We're doing Gauss's law, so I'm assuming the approximation is okay. I did EA=Q/$$\epsilon$$ and set the A to h*2$$\pi$$r (r being the r of the Gaussian surface, .203m) It wasn't right...what am I doing wrong? 4. Feb 22, 2010 ### ideasrule h is the length of the cylinder, right? If so, then your equation is right. 5. Feb 22, 2010 ### csteff Ya, sorry. Okay thanks! I'm not sure why my answer's coming out wrong...I guess I'll just show my teacher my work and see what she says haha Similar Discussions: Electric field / net charge on a cylindrical shell
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9511406421661377, "perplexity": 476.79700476894646}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187822668.19/warc/CC-MAIN-20171018013719-20171018033719-00251.warc.gz"}
https://platonicrealms.com/encyclopedia/standard-deviation
# standard deviation The standard deviation of a set of values is a measure of the variability of those values. Specifically, it is a positive number representing the average distance that the values in a dataset tend to fall from their arithmetic mean. Typically the standard deviation is denoted by $$s$$ for finite sets of values and by $$\sigma$$ (Greek lower-case sigma) for distributions represented by a density curve. The standard deviation is calculated by taking the square root of the variance, which itself is calculated in different ways depending, again, on whether the dataset is finite or is represented by a density curve. For a finite dataset, the variance is calculated by finding the distance of each value from the mean, squaring these distances, summing them, and dividing the result by one less than the number of values. That is, for a dataset of size $$n$$, the variance $$s^2$$ is given by $s^2=\displaystyle\frac{1}{n-1}\sum_{i=1}^n (\overline{x}-x_i)^2$ The standard deviation $$s$$ is then given by $s=\sqrt{s^2}=\sqrt{\displaystyle\frac{1}{n-1}\sum_{i=1}^n (\overline{x}-x_i)^2}$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9503157138824463, "perplexity": 163.0382360068347}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623487660269.75/warc/CC-MAIN-20210620084505-20210620114505-00256.warc.gz"}
http://physics.stackexchange.com/questions/3314/energy-field-of-a-newly-created-particle
# energy field of a newly created particle The question is: How to calculate the energy of electrostatic field of a newly created particle. The usual formula involves integration 'all over space'. According to GR I believe that 'all over space' means the interior volume with radius c*t_now, where t_now is the lifetime of the particle at the time of calculation and c is the light speed. Thus the energy stored in the field is a quantity dependent on time evolution. - The energy of electrostatic field is not involved in any exchanges so it is a meaningless notion. –  Vladimir Kalitvianski Jan 22 '11 at 14:31 classical physics doesn't have any well-defined framework to describe the "creation of particles". So you have to talk about the real world - or, theoretical speaking, the world of quantum field theory (or something that contains it) - to be able to discuss the creation of particles. However, in quantum field theory, you can't separate the energy coming from the field from the energy coming from the particles themselves in such a sharp way. In particular, even the lightest charged particle - the electron - has mass equal to 511 keV. That's big, and even if you compute the whole energy of the electrostatic field at the distance of the classical electron radius, $2.82 \times 10^{-15}$ meters, or shorter (not far from the radius of the proton, by the way), you will obtain a smaller energy than the rest energy of the electron. If you want to say that a (virtual) electron existed at least for a little while, you need to assume that the lifetime was at least the Compton wavelength of the electron (divided by $c$). That's about 100 times longer a distance. And even with this distance, the uncertainty of the energy is comparable to those 511 keV. This leads me to the key point which is the uncertainty principle for time and energy. If you determine the timing of the places where you measure the energy with a much better accuracy than the Compton wavelength of the electron (over $c$), then the uncertainty of the energy you measure will inevitably be much greater than the whole electron rest mass (times $c^2$). And if you determine the the timing with a worse accuracy, then you will inevitably include almost the whole electric field of the electron, and you will get the electron mass (times $c^2$) of 511 keV with a big accuracy. At any rate, you won't be able to show any violation of the total energy which holds exactly in any system with a time-translational symmetry. You will be extremely far from finding such a contradiction: even if you could attribute parts of energy to fields and particles at every moment, as you suggested and which you cannot because of the uncertainty principle, you couldn't measure it with the same precision, and even if you could measure it, the variable mass of the electron at the "center" of the field could always compensate any violation of the conservation law, anyway. So the really important message is that the total energy is conserved, it can be measured as long as we have a long enough time to measure it, and attributing energy to small pieces of space around a particle is not the right way to proceed - neither theoretically nor experimentally. Instead, one should ask what are the probabilities of various processes. External and internal particles will always show that energy is conserved. However, energy can only be measured accurately if it is measured for a long enough time - longer than $\hbar/\Delta E$ where $\Delta E$ is the required precision. If you think about these matters, you will find out that you are trying to measure the amount and localization of energy - in space and/or in time - more accurately than the uncertainty principle allows. Cheers LM - Luboš Motl: Thanks for your response (I follow your TRF;) I firmly believe that a particle is not separable from its exterior field and the total energy of the ensemble is constant. The problem is that if we accept that the field spreads out in space at the speed of light then the field 'outside', using a maximal radius by a suitable convention, has an energy content increased over time, implying that the energy located in what we call 'particle' is expected to decrease over time. Even working with probabilities I arrive to the same conclusion. –  Helder Velez Jan 19 '11 at 14:27 According to the usual laws of physics, charge cannot be created or destroyed. So it's impossible to create a particle with charge q without annihilating a particle with the same charge, or also creating a particle with charge -q, or some combination of these. Suppose you annihilated a particle with the same charge. Then your problem might better be described as one of "which particle should you attribute the electrostatic energy to, the annihilated one or the created one?" And if instead it was a case of creating two particles with charges +q and -q, then your problem could be described as one of having to separate the electrostatic energies of the two particles. So I don't see that there is any real problem with defining the energy of the electrostatic field in a particle. It's a matter of semantics. I think the universe has no opinion as to which particle we attribute the electrostatic energy. - We can invent a time varying charge density $\rho (t)$ to model appearing and disappearing charge and watch the filed evolution. It seems to me I saw somewhere a problem of filed propagation from suddenly created charge. Of course, such a phenomenon does not occur in reality. As I said previously, the electrostatic filed energy is not involved in any exchange so its "evolution" does not matter. –  Vladimir Kalitvianski Jan 23 '11 at 19:44 The electrostatic field (or gravitational) detector work by energy transfer from the environment (the field) and the matter of the detector. As energy got transferred we can assume that field has energy. –  Helder Velez Jan 24 '11 at 18:58 @all The particle (or the pair) spent all his existence impressing the environment (vacuum, space,..) with a spreading field. When the particle (or pair) got annihilated (converted to gamma-rays) the interaction time is very short. The far field can not be instantly unset or reverted to the source.Thus the energy of the field 'is lost' into the space and keeps spreading. Do you agree that 'all over space' has to be only the space interior to the sphere with radius c*t_now ? –  Helder Velez Jan 24 '11 at 19:23
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9268288016319275, "perplexity": 287.07476768490756}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416931012025.85/warc/CC-MAIN-20141125155652-00082-ip-10-235-23-156.ec2.internal.warc.gz"}
https://qiskit.org/documentation/optimization/tutorials/07_examples_vehicle_routing.html
Note Run interactively in jupyter notebook. # Vehicle Routing¶ ## The Introduction¶ Logistics is a major industry, with some estimates valuing it at USD 8183 billion globally in 2015. Most service providers operate a number of vehicles (e.g., trucks and container ships), a number of depots, where the vehicles are based overnight, and serve a number of client locations with each vehicle during each day. There are many optimization and control problems that consider these parameters. Computationally, the key challenge is how to design routes from depots to a number of client locations and back to the depot, so as to minimize vehicle-miles traveled, time spent, or similar objective functions. In this notebook we formalize an idealized version of the problem and showcase its solution using the quantum approximate optimization approach of Farhi, Goldstone, and Gutmann (2014). The overall workflow we demonstrate comprises: 1. establish the client locations. Normally, these would be available ahead of the day of deliveries from a database. In our use case, we generate these randomly. 2. compute the pair-wise distances, travel times, or similar. In our case, we consider the Euclidean distance, “as the crow flies”, which is perhaps the simplest possible. 3. compute the actual routes. This step is run twice, actually. First, we obtain a reference value by a run of a classical solver (IBM CPLEX) on the classical computer. Second, we run an alternative, hybrid algorithm partly on the quantum computer. 4. visualization of the results. In our case, this is again a simplistic plot. In the following, we first explain the model, before we proceed with the installation of the pre-requisites and the data loading. ## The Model¶ Mathematically speaking, the vehicle routing problem (VRP) is a combinatorial problem, wherein the best routes from a depot to a number of clients and back to the depot are sought, given a number of available vehicles. There are a number of formulations possible, extending a number of formulations of the traveling salesman problem [Applegate et al, 2006]. Here, we present a formulation known as MTZ [Miller, Tucker, Zemlin, 1960]. Let $$n$$ be the number of clients (indexed as $$1,\dots,n$$), and $$K$$ be the number of available vehicles. Let $$x_{ij} = \{0,1\}$$ be the binary decision variable which, if it is $$1$$, activates the segment from node $$i$$ to node $$j$$. The node index runs from $$0$$ to $$n$$, where $$0$$ is (by convention) the depot. There are twice as many distinct decision variables as edges. For example, in a fully connected graph, there are $$n(n+1)$$ binary decision variables. If two nodes $$i$$ and $$j$$ have a link from $$i$$ to $$j$$, we write $$i \sim j$$. We also denote with $$\delta(i)^+$$ the set of nodes to which $$i$$ has a link, i.e., $$j \in \delta(i)^+$$ if and only if $$i \sim j$$. Similarly, we denote with $$\delta(i)^-$$ the set of nodes which are connected to $$i$$, in the sense that $$j \in \delta(i)^-$$ if and only if $$j \sim i$$. In addition, we consider continuous variables, for all nodes $$i = 1,\dots, n$$, denoted $$u_i$$. These variables are needed in the MTZ formulation of the problem to eliminate sub-tours between clients. The VRP can be formulated as: $(VRP) \quad f = \min_{\{x_{ij}\}_{i\sim j}\in \{0,1\}, \{u_i\}_{i=1,\dots,n}\in \mathbb{R}} \quad \sum_{i \sim j} w_{ij} x_{ij}$ subject to the node-visiting constraint: $\sum_{j \in \delta(i)^+} x_{ij} = 1, \,\sum_{j \in \delta(i)^-} x_{ji} = 1,\, \forall i \in \{1,\dots,n\},$ the depot-visiting constraints: $\sum_{i \in \delta(0)^+} x_{0i} = K, \, \sum_{j \in \delta(0)^+} x_{j0} = K,$ and the sub-tour elimination constraints: $u_i - u_j + Q x_{ij} \leq Q-q_j, \, \forall i \sim j, \,i ,j \neq 0, \quad q_i \leq u_i \leq Q,\, \forall i, i \neq 0.$ In particular, - The cost function is linear in the cost functions and weighs the different arches based on a positive weight $$w_{ij}>0$$ (typically the distance between node $$i$$ and node $$j$$); - The first set of constraints enforce that from and to every client, only one link is allowed; - The second set of constraints enforce that from and to the depot, exactly $$K$$ links are allowed; - The third set of constraints enforce the sub-tour elimination constraints and are bounds on $$u_i$$, with $$Q>q_j>0$$, and $$Q,q_i \in \mathbb{R}$$. ## Classical solution¶ We can solve the VRP classically, e.g., by using CPLEX. CPLEX uses a branch-and-bound-and-cut method to find an approximate solution of the VRP, which, in this formulation, is a mixed-integer linear program (MILP). For the sake of notation, we pack the decision variables in one vector as ${\bf z} = [x_{01},x_{02},\ldots,x_{10}, x_{12},\ldots,x_{n(n-1)}]^T,$ wherein $${\bf z} \in \{0,1\}^N$$, with $$N = n (n+1)$$. So the dimension of the problem scales quadratically with the number of nodes. Let us denote the optimal solution by $${\bf z}^*$$, and the associated optimal cost $$f^*$$. ## Quantum solution¶ Here, we demonstrate an approach that combines classical and quantum computing steps, following the quantum approximate optimization approach of Farhi, Goldstone, and Gutmann (2014). In particular, we use the variational quantum eigensolver (VQE). We stress that given the use of limited depth of the quantum circuits employed (variational forms), it is hard to discuss the speed-up of the algorithm, as the solution obtained is heuristic in nature. At the same time, due to the nature and importance of the target problems, it is worth investigating heuristic approaches, which may be worthwhile for some problem classes. Following [5], the algorithm can be summarized as follows: - Preparation steps: - Transform the combinatorial problem into a binary polynomial optimization problem with equality constraints only; - Map the resulting problem into an Ising Hamiltonian ($$H$$) for variables $${\bf z}$$ and basis $$Z$$, via penalty methods if necessary; - Choose the depth of the quantum circuit $$m$$. Note that the depth can be modified adaptively. - Choose a set of controls $$\theta$$ and make a trial function $$\big|\psi(\boldsymbol\theta)\rangle$$, built using a quantum circuit made of C-Phase gates and single-qubit Y rotations, parameterized by the components of $$\boldsymbol\theta$$. • Algorithm steps: • Evaluate $$C(\boldsymbol\theta) = \langle\psi(\boldsymbol\theta)\big|H\big|\psi(\boldsymbol\theta)\rangle$$ by sampling the outcome of the circuit in the Z-basis and adding the expectation values of the individual Ising terms together. In general, different control points around $$\boldsymbol\theta$$ have to be estimated, depending on the classical optimizer chosen. • Use a classical optimizer to choose a new set of controls. • Continue until $$C(\boldsymbol\theta)$$ reaches a minimum, close enough to the solution $$\boldsymbol\theta^*$$. • Use the last $$\boldsymbol\theta$$ to generate a final set of samples from the distribution $$\Big|\langle z_i\big|\psi(\boldsymbol\theta)\rangle\Big|^2\;\forall i$$ to obtain the answer. There are many parameters throughout, notably the choice of the trial wavefunction. Below, we consider: $\big|\psi(\theta)\rangle = [U_\mathrm{single}(\boldsymbol\theta) U_\mathrm{entangler}]^m \big|+\rangle$ where $$U_\mathrm{entangler}$$ is a collection of C-Phase gates (fully-entangling gates), and $$U_\mathrm{single}(\theta) = \prod_{i=1}^N Y(\theta_{i})$$, where $$N$$ is the number of qubits and $$m$$ is the depth of the quantum circuit. ### Construct the Ising Hamiltonian¶ From $$VRP$$ one can construct a binary polynomial optimization with equality constraints only by considering cases in which $$K=n-1$$. In these cases the sub-tour elimination constraints are not necessary and the problem is only on the variable $${\bf z}$$. In particular, we can write an augmented Lagrangian as $(IH) \quad H = \sum_{i \sim j} w_{ij} x_{ij} + A \sum_{i \in \{1,\dots,n\}} \Big(\sum_{j \in \delta(i)^+} x_{ij} - 1\Big)^2 + A \sum_{i \in \{1,\dots,n\}}\Big(\sum_{j \in \delta(i)^-} x_{ji} - 1\Big)^2 +A \Big(\sum_{i \in \delta(0)^+} x_{0i} - K\Big)^2 + A\Big(\sum_{j \in \delta(0)^+} x_{j0} - K\Big)^2$ where $$A$$ is a big enough parameter. ### From Hamiltonian to QP formulation¶ In the vector $${\bf z}$$, and for a complete graph ($$\delta(i)^+ = \delta(i)^- = \{0,1,\dots,i-1,i+1,\dots,n\}$$), $$H$$ can be written as follows. $\min_{{\bf z}\in \{0,1\}^{n(n+1)}} {\bf w}^T {\bf z} + A \sum_{i \in \{1,\dots,n\}} \Big({\bf e}_i \otimes {\bf 1}_n^T {\bf z} - 1\Big)^2 + A \sum_{i \in \{1,\dots,n\}}\Big({\bf v}_i^T {\bf z} - 1\Big)^2 + A \Big(({\bf e}_0 \otimes {\bf 1}_n)^T{\bf z} - K\Big)^2 + A\Big({\bf v}_0^T{\bf z} - K\Big)^2.$ That is: $\min_{\bf z\in \{0,1\}^{n(n+1)}} \bf z^T {\bf Q} \bf z + {\bf g}^T \bf z + c,$ Where: first term: ${\bf Q} = A \sum_{i \in \{0,1,\dots,n\}} \Big[({\bf e}_i \otimes {\bf 1}_n)({\bf e}_i \otimes {\bf 1}_n)^T + {\bf v}_i{\bf v}_i^T \Big]$ Second term: ${\bf g} = {\bf w} -2 A \sum_{i \in \{1,\dots,n\}} \Big[({\bf e}_i \otimes {\bf 1}_n) + {\bf v}_i \Big] -2 A K \Big[({\bf e}_0 \otimes {\bf 1}_n) + {\bf v}_0 \Big]$ Third term: $c = 2An +2AK^2.$ The QP formulation of the Ising Hamiltonian is ready for the use of VQE. We will solve the QP using optimization stack available in Qiskit. ## References¶ [1] E. Farhi, J. Goldstone, S. Gutmann e-print arXiv 1411.4028, 2014 [3] C. E. Miller, E. W. Tucker, and R. A. Zemlin (1960). “Integer Programming Formulations and Travelling Salesman Problems”. J. ACM. 7: 326–329. doi:10.1145/321043.321046. [4] D. L. Applegate, R. M. Bixby, V. Chvátal, and W. J. Cook (2006). The Traveling Salesman Problem. Princeton University Press, ISBN 978-0-691-12993-8. ## Initialization¶ First of all we load all the packages that we need: - Python 3.6 or greater is required; - CPLEX 12.8 or greater is required for the classical computations; - Latest Qiskit is required for the quantum computations. [1]: # Load the packages that are required import numpy as np import operator import matplotlib.pyplot as plt import sys if sys.version_info < (3, 6): raise Exception('Please use Python version 3.6 or greater.') try: import cplex from cplex.exceptions import CplexError except: import math # Qiskit packages from qiskit import BasicAer from qiskit.quantum_info import Pauli from qiskit.utils import QuantumInstance, algorithm_globals from qiskit.algorithms import NumPyMinimumEigensolver, VQE from qiskit.circuit.library import TwoLocal from qiskit.algorithms.optimizers import SPSA We then initialize the variables [2]: # Initialize the problem by defining the parameters n = 3 # number of nodes + depot (n+1) K = 2 # number of vehicles We define an initializer class that randomly places the nodes in a 2-D plane and computes the distance between them. [3]: # Get the data class Initializer(): def __init__(self, n): self.n = n def generate_instance(self): n = self.n # np.random.seed(33) np.random.seed(1543) xc = (np.random.rand(n) - 0.5) * 10 yc = (np.random.rand(n) - 0.5) * 10 instance = np.zeros([n, n]) for ii in range(0, n): for jj in range(ii + 1, n): instance[ii, jj] = (xc[ii] - xc[jj]) ** 2 + (yc[ii] - yc[jj]) ** 2 instance[jj, ii] = instance[ii, jj] return xc, yc, instance [4]: # Initialize the problem by randomly generating the instance initializer = Initializer(n) xc,yc,instance = initializer.generate_instance() ## Classical solution using IBM ILOG CPLEX¶ For a classical solution, we use IBM ILOG CPLEX. CPLEX is able to find the exact solution of this problem. We first define a ClassicalOptimizer class that encodes the problem in a way that CPLEX can solve, and then instantiate the class and solve it. [5]: class ClassicalOptimizer: def __init__(self, instance,n,K): self.instance = instance self.n = n # number of nodes self.K = K # number of vehicles def compute_allowed_combinations(self): f = math.factorial return f(self.n) / f(self.K) / f(self.n-self.K) def cplex_solution(self): # refactoring instance = self.instance n = self.n K = self.K my_obj = list(instance.reshape(1, n**2)[0])+[0. for x in range(0,n-1)] my_ub = [1 for x in range(0,n**2+n-1)] my_lb = [0 for x in range(0,n**2)] + [0.1 for x in range(0,n-1)] my_ctype = "".join(['I' for x in range(0,n**2)]) + "".join(['C' for x in range(0,n-1)]) my_rhs = 2*([K] + [1 for x in range(0,n-1)]) + [1-0.1 for x in range(0,(n-1)**2-(n-1))] + [0 for x in range(0,n)] my_sense = "".join(['E' for x in range(0,2*n)]) + "".join(['L' for x in range(0,(n-1)**2-(n-1))])+"".join(['E' for x in range(0,n)]) try: my_prob = cplex.Cplex() self.populatebyrow(my_prob,my_obj,my_ub,my_lb,my_ctype,my_sense,my_rhs) my_prob.solve() except CplexError as exc: print(exc) return x = my_prob.solution.get_values() x = np.array(x) cost = my_prob.solution.get_objective_value() return x,cost def populatebyrow(self,prob,my_obj,my_ub,my_lb,my_ctype,my_sense,my_rhs): n = self.n prob.objective.set_sense(prob.objective.sense.minimize) prob.variables.add(obj = my_obj, lb = my_lb, ub = my_ub, types = my_ctype) prob.set_log_stream(None) prob.set_error_stream(None) prob.set_warning_stream(None) prob.set_results_stream(None) rows = [] for ii in range(0,n): col = [x for x in range(0+n*ii,n+n*ii)] coef = [1 for x in range(0,n)] rows.append([col, coef]) for ii in range(0,n): col = [x for x in range(0+ii,n**2,n)] coef = [1 for x in range(0,n)] rows.append([col, coef]) # Sub-tour elimination constraints: for ii in range(0, n): for jj in range(0,n): if (ii != jj)and(ii*jj>0): col = [ii+(jj*n), n**2+ii-1, n**2+jj-1] coef = [1, 1, -1] rows.append([col, coef]) for ii in range(0,n): col = [(ii)*(n+1)] coef = [1] rows.append([col, coef]) [6]: # Instantiate the classical optimizer class classical_optimizer = ClassicalOptimizer(instance,n,K) # Print number of feasible solutions print('Number of feasible solutions = ' + str(classical_optimizer.compute_allowed_combinations())) Number of feasible solutions = 3.0 [7]: # Solve the problem in a classical fashion via CPLEX x = None z = None try: x,classical_cost = classical_optimizer.cplex_solution() # Put the solution in the z variable z = [x[ii] for ii in range(n**2) if ii//n != ii%n] # Print the solution print(z) except: print("CPLEX may be missing.") [1.0, 1.0, 1.0, 0.0, 1.0, 0.0] [8]: # Visualize the solution def visualize_solution(xc, yc, x, C, n, K, title_str): plt.figure() plt.scatter(xc, yc, s=200) for i in range(len(xc)): plt.annotate(i, (xc[i] + 0.15, yc[i]), size=16, color='r') plt.plot(xc[0], yc[0], 'r*', ms=20) plt.grid() for ii in range(0, n ** 2): if x[ii] > 0: ix = ii // n iy = ii % n plt.title(title_str+' cost = ' + str(int(C * 100) / 100.)) plt.show() if x is not None: visualize_solution(xc, yc, x, classical_cost, n, K, 'Classical') If you have CPLEX, the solution shows the depot with a star and the selected routes for the vehicles with arrows. ## Quantum solution from the ground up¶ For the quantum solution, we use Qiskit. First, we derive the solution from the ground up, using a class QuantumOptimizer that encodes the quantum approach to solve the problem and then we instantiate it and solve it. We define the following methods inside the class: - binary_representation : encodes the problem $$(M)$$ into a QP terms (that’s basically linear algebra); - construct_problem : constructs a QUBO optimization problem as an instance of QuadraticProgram; - solve_problem: solves the problem $$(M)$$ constructed at the previous step via MinimunEigenOptimizer by using VQE with default parameters; [9]: from qiskit_optimization import QuadraticProgram from qiskit_optimization.algorithms import MinimumEigenOptimizer class QuantumOptimizer: def __init__(self, instance, n, K): self.instance = instance self.n = n self.K = K def binary_representation(self,x_sol=0): instance = self.instance n = self.n K = self.K A = np.max(instance) * 100 # A parameter of cost function # Determine the weights w instance_vec = instance.reshape(n ** 2) w_list = [instance_vec[x] for x in range(n ** 2) if instance_vec[x] > 0] w = np.zeros(n * (n - 1)) for ii in range(len(w_list)): w[ii] = w_list[ii] # Some variables I will use Id_n = np.eye(n) Im_n_1 = np.ones([n - 1, n - 1]) Iv_n_1 = np.ones(n) Iv_n_1[0] = 0 Iv_n = np.ones(n-1) neg_Iv_n_1 = np.ones(n) - Iv_n_1 v = np.zeros([n, n*(n-1)]) for ii in range(n): count = ii-1 for jj in range(n*(n-1)): if jj//(n-1) == ii: count = ii if jj//(n-1) != ii and jj%(n-1) == count: v[ii][jj] = 1. vn = np.sum(v[1:], axis=0) # Q defines the interactions between variables Q = A*(np.kron(Id_n, Im_n_1) + np.dot(v.T, v)) # g defines the contribution from the individual variables g = w - 2 * A * (np.kron(Iv_n_1,Iv_n) + vn.T) - \ 2 * A * K * (np.kron(neg_Iv_n_1, Iv_n) + v[0].T) # c is the constant offset c = 2 * A * (n-1) + 2 * A * (K ** 2) try: max(x_sol) # Evaluates the cost distance from a binary representation of a path fun = lambda x: np.dot(np.around(x), np.dot(Q, np.around(x))) + np.dot(g, np.around(x)) + c cost = fun(x_sol) except: cost = 0 return Q, g, c, cost def construct_problem(self, Q, g, c) -> QuadraticProgram: for i in range(n * (n - 1)): qp.binary_var(str(i)) qp.objective.linear = g qp.objective.constant = c return qp def solve_problem(self, qp): algorithm_globals.random_seed = 10598 quantum_instance = QuantumInstance(BasicAer.get_backend('qasm_simulator'), seed_simulator=algorithm_globals.random_seed, seed_transpiler=algorithm_globals.random_seed) vqe = VQE(quantum_instance=quantum_instance) optimizer = MinimumEigenOptimizer(min_eigen_solver=vqe) result = optimizer.solve(qp) # compute cost of the obtained result _,_,_,level = self.binary_representation(x_sol=result.x) return result.x, level ### Step 1¶ Instantiate the quantum optimizer class with parameters: - the instance; - the number of nodes and vehicles n and K; [10]: # Instantiate the quantum optimizer class with parameters: quantum_optimizer = QuantumOptimizer(instance, n, K) ### Step 2¶ Encode the problem as a binary formulation (IH-QP). Sanity check: make sure that the binary formulation in the quantum optimizer is correct (i.e., yields the same cost given the same solution). [11]: # Check if the binary representation is correct try: if z is not None: Q, g, c, binary_cost = quantum_optimizer.binary_representation(x_sol = z) print("Binary cost:", binary_cost, "classical cost:", classical_cost) if np.abs(binary_cost - classical_cost) < 0.01: print('Binary formulation is correct') else: print('Error in the binary formulation') else: print('Could not verify the correctness, due to CPLEX solution being unavailable.') Q, g, c, binary_cost = quantum_optimizer.binary_representation() print("Binary cost:", binary_cost) except NameError as e: print("Warning: Please run the cells above first.") print(e) Binary cost: 132.11148115684045 classical cost: 132.1114811568365 Binary formulation is correct ### Step 3¶ Encode the problem as an instance of QuadraticProgram. [12]: qp = quantum_optimizer.construct_problem(Q, g, c) ### Step 4¶ Solve the problem via MinimumEigenOptimizer from the optimization stack. N.B. Depending on the number of qubits, the state-vector simulation can take a while; for example with 12 qubits, it takes more than 12 hours. Logging is useful to see what the program is doing. [13]: quantum_solution, quantum_cost = quantum_optimizer.solve_problem(qp) print(quantum_solution, quantum_cost) [1. 1. 1. 0. 1. 0.] 132.11148115684045 ### Step 5¶ Visualize the solution [14]: # Put the solution in a way that is compatible with the classical variables x_quantum = np.zeros(n**2) kk = 0 for ii in range(n ** 2): if ii // n != ii % n: x_quantum[ii] = quantum_solution[kk] kk += 1 # visualize the solution visualize_solution(xc, yc, x_quantum, quantum_cost, n, K, 'Quantum') # and visualize the classical for comparison if x is not None: visualize_solution(xc, yc, x, classical_cost, n, K, 'Classical') The plots present the depot with a star and the selected routes for the vehicles with arrows. Note that in this particular case, we can find the optimal solution of the QP formulation, which happens to coincide with the optimal solution of the ILP. Keep in mind that VQE is an heuristic working on the QP formulation of the Ising Hamiltonian, though. For suitable choices of A, local optima of the QP formulation will be feasible solutions to the ILP. While for some small instances, as above, we can find optimal solutions of the QP formulation which coincide with optima of the ILP, finding optimal solutions of the ILP is harder than finding local optima of the QP formulation, in general, which in turn is harder than finding feasible solutions of the ILP. Even within the VQE, one may provide stronger guarantees, for specific variational forms (trial wave functions). [15]: import qiskit.tools.jupyter %qiskit_version_table ### Version Information Qiskit SoftwareVersion QiskitNone Terra0.17.0 Aer0.8.0 IgnisNone AquaNone IBM Q ProviderNone System information Python3.8.8 (default, Feb 19 2021, 19:42:00) [GCC 9.3.0] OSLinux CPUs2 Memory (Gb)6.791343688964844 Fri Apr 02 20:12:31 2021 UTC ### This code is a part of Qiskit [ ]:
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6770110726356506, "perplexity": 2870.697209628894}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243992514.37/warc/CC-MAIN-20210513204127-20210513234127-00093.warc.gz"}
https://cds.cern.ch/collection/ARDA%20Internal%20Documentation?ln=pl
The CERN Document Server website will be shortly unavailable the 21st of August at 12:00 noon CEST due to a planned intervention. # ARDA Internal Documentation Ogranicz wyniki do kolekcji: [zastrzeżone]
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9671506285667419, "perplexity": 19512.266823757986}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027315329.55/warc/CC-MAIN-20190820113425-20190820135425-00070.warc.gz"}
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2570547/?tool=pubmed
Open Biochem J. 2008; 2: 1–5. Published online Jan 18, 2008. PMCID: PMC2570547 # Consensus Prediction of Protein Conformational Disorder from Amino Acidic Sequence ## Abstract Predictions of protein conformational disorder are important in structural biology since they can allow the elimination of protein constructs, the three-dimensional structure of which cannot be determined since they are natively unfolded. Here a new procedure is presented that allows one to predict with high accuracy disordered residues on the basis of protein sequences. It makes use of twelve prediction methods and merges their results by using least-squares optimization. A statistical survey of the Protein Data Bank is also reported, in order to know how many residues can be disordered in proteins that were crystallized and the three-dimensional structure of which was determined. ## INTRODUCTION It was recently shown that several proteins do not assume a well defined and stable three-dimensional (3D) structure but are natively unfolded [1]. This was absolutely surprising since unfolded proteins are known to be less stable and soluble in vitro and protein misfolding is known to be associated with several conformational diseases, including Parkinson and Alzheimer [2]. However, a considerable fraction of the proteome is constituted by natively unfolded proteins and this fraction seems to be larger in higher organisms than in simpler prokaryotes. Several techniques to predict conformational disorder in proteins have been designed [3-5] and the performance of many of them is periodically checked, within the CASP initiatives [6], where several blinded predictions are made on targets, the conformational status of which is known only by the CASP organizers and is unknown by the various prediction teams that participate to CASP. In general, it appears that (i) the reliability of these predictions is rather modest and that (ii) different predictions are made by different predictors. The first point is per se not surprising, given the intrinsic difficulty of predicting 3D features on the basis of amino acidic sequences. The second point - the inconsistency between different prediction methods - is also not very surprising. In fact, various predictors do not differ only in their algorithms but also in what they define as "conformational disorder" and thus in what they want to predict. For example, in one of the DISEMBL versions [7], all the residues in loops are considered to be conformationally disordered, while in another of the DISEMBL versions, only the residues that were not visible in the crystallographic electron density maps are considered to be disordered. Alternatively, in IUPRED no a priori definition of disorder is used [8]. Despite their limitations, the techniques for predicting conformational disorder are extremely important. Initially, they were designed principally to study the interesting phenome-non of conformation disorder and for large-scale proteome comparisons. Later, it became clear that they have also a series of practical applications, like for example in structural genomics, where they are becoming routine filters in the pipeline of finding suitable targets to be analyzed [3, 4]. In fact, it is obvious that the 3D structure of natively unfolded proteins cannot be determined and that these disordered targets must not be analyzed experimentally by structural biologists. In this paper we present a consensus method, based on various prediction methods, the performance of which is significantly better than that of each individual predictor. Such a new technique is easily usable with freely available software and is interesting not only for structural genomics initiatives but also for traditional hypothesis-driven structural biology. We also report a statistical survey of the Protein Data Bank that shows the fraction of disordered residues in proteins the crystal structure of which was determined. It appears that a moderate fraction of conformationally disordered residues can be tolerated. About 22% of these crystal structures have more the 5% of the residues that are disordered, though only about 2% of them have more than 20% of the residues in a conformationally disordered status. ## METHODS ### Data Information about conformationally disordered proteins was taken from the DISPROT database (http://www. disprot.org/) release 3.3 [9], which lists, in FASTA format, 458 proteins that are known, on the basis of several experimental studies, to be at least partially disordered. Data were downloaded in August 2006. Each residue of these 458 proteins is labeled according to its conformational status: ordered, disordered, unknown. The main advantage of the DISPROT database is that it is curated by experts and it is not based on some automatic procedure. It is thus reasonable to suppose that it contains a very limited number of inaccuracies. ### Individual Predictors 12 individual predictors were used (Table 11 Some of them are different versions of the same basic algorithm. For example, IUPRED has two versions, one specialized in predicting short disordered polypeptide fragments and the other focused on the prediction of long disordered polypeptide fragments. Others have even three versions, like DISEMBL, which can predict if a residue is in a loop, in a "hot" loop (characterized by high crystallographic B factors), or if it was not observed in the electron density maps, as stated on the “REMARK 465” lines of the files of the Protein Data Bank. Given that the present manuscript is not focused on a particular type of disorder but it is focused on the identification of protein constructs that cannot be studied by structural biologists, we did not make any difference between the various versions of the predictors and we used all of them. This is justified by the fact that we do not want to design a new predictor but we want only to make consensus predictions that can be useful in structural biology for high-throughput structural genomics initiatives and, more in general, in any structural biology project. Moreover, the mathematical approach we used (see below) is essentially unaffected by the use of similar or redundant prediction methods given that it is a least-squares optimization, which by definition, weights all contributions as a function of each other. Individual Prediction Methods Used in the Present Paper ### Consensus Predictions Each prediction method (Table 11) produces binary results: a residue can be predicted to be conformationally ordered or disordered. From a numerical perspective, this can be represented by a value of +1 if it is predicted to be disordered, or by a value of -1, if it is predicted to be ordered. The numerical value of 1 and its sign, positive or negative, are purely arbitrary and different values or opposite signs would not affect the quality of the results. $P⋅X=D$ where P is a N x 12 matrix the elements pij of which are either +1, if the ith residues is predicted to be disordered by the jth prediction method, or -1 if it is predicted to be ordered, and where D is a vector of N elements di, the values of which can be either +1, if the ith residues is disordered in the reality, or -1, in the opposite case. The value of N is the total number of residues that are annotated to be ordered or disordered in the DISPROT database and is equal to 54012 residues. Once the optimal values of the elements of X have been determined, it is possible to use them to predict if a residue is conformationally ordered or disordered by computing its p_cons value $P_cons=∑i=112xi⋅p_indii$ where the values of p_indii are either +1, if the residue is predicted to be disordered by the ith prediction method, or -1, if it is predicted to be ordered. If p_cons is closer to +1 than to -1, which means if it is greater than 0, the residue is predicted to be disordered. On the contrary, it is predicted to be ordered if p_cons 0. The optimal values of the coefficients xi are reported in Table 22. Optimal Values of the Coefficients xi to be Used to Compute the p_cones Values (Equation 2) ### Prediction Validation Given the extremely high number (54012) of amino acid residues contained in the DISPROT database, a complete cross-validation, known also as Jack-knife test, is impossible. We performed thus a 20-fold cross-validation: we built randomly 20 non-overlapping sets of residues, each containing 5% of the data, and the optimization of the X vector was performed 20 times by discarding each time one of the small subsets, which was then used to compute the p_cons values. Such a separation between the learning sets and the test sets allows one to make unbiased predictions, which can then be compared with the experimentally known conformational statuses of the residues. A residue correctly predicted to be disordered was counted as a true positive (tp). A residue correctly predicted to be ordered was counted as a true negative (tn). A disordered residue predicted to be ordered was counted as a false negative (fn). An ordered residue predicted to be disordered was counted as a false positive (fp). Given these four quantities, the prediction reliability was estimated with a series of figures of merit: the sensitivity, the specificity, the accuracy, and the probability excess, defined as $sensitivity=tptp+fn$ $specificity=tptp+fp$ $specificity=tp+tntp+tn+fp+fn$ $probability_excess=sensitivity+specificity−1$ The values of these figures of merit can range from 0 to +1 and larger values, closer to +1, are associated with better predictions. It must be observed that some of these figures of merit, typically the accuracy, can be seriously biased if the data are unbalanced. This is exactly what happens here, since the number of ordered residues (2649) is very different from the number of disordered residues (51363) in the database DISPROT. The values of accuracy are thus provided in the present paper only because this figure of merit is used very commonly in computational biology. A much more robust indicator of prediction quality is the probability excess. ## RESULTS AND DISCUSSION Besides their basic biological importance, predictions of protein conformational disorder are important in structural biology, where "impossible" targets must be identified before inserting them in the experimental pipe-line that goes from cloning to structural determination. This is particularly important not only in structural genomics initiatives, the success rate of which is still rather modest, but also in traditional hypothesis-driven applications, especially when the protein construct must be designed by the scientists, like for example in multi-domain protein and viral poly-proteins [10]. Predictions of conformational disorder are thus one of the bioinformatics filters that must be used before moving towards experimental analyses. Other filters are focused on the quaternary structural requirements of a protein chain [11], on protein solubility and stability [12, 13], and some web-based servers were created to assist the users in this task [14, 15]. However, before doing predictions of conformational disorder it is necessary to know what level of disorder can be tolerated by well folded proteins. In fact, while it is clear that the 3D structure of a completely disordered protein cannot be determined, it is also clear that many (or, maybe, most) proteins are partially disordered. For example, many loops at the protein surface are very flexible and tend to adopt more than a single shape. For this reason, we scanned the Protein Data Bank (PDB) [16, 17] looking for regions conformationally disordered. This information was extracted from the records labeled with "REMARK 465", where the depositors of the crystal structures declare, if necessary, which residues were not observed in the electron density maps. This analysis was limited to the crystal structures, which are nevertheless the large majority of the entries of the PDB, and it was assumed that the location of completely unfolded segments cannot be detected in the electron density maps. Fig.11) shows the distribution of the PDB entries according to their fraction of residues not observed which are likely to be conformationally disordered. It appears that a considerable number of structures have conformational disorder. In 22% of them, more than 5% of the residues are disordered. However, only about 2% of the crystal structures contain more than 20% of the residues that lack a well defined structure. The most extreme case is the entry 1VCR, the light-harvesting complex from Pisum sativum thylacoid membrane, where 56% of the residues were not observed, though this crystal structure was determined and refined at very low resolution (9.5 Å) [18]. Distribution of protein crystal structures as a function of the percentage of disordered residues they contain. The data were taken from the Protein Data Bank; a residues was considered to be disordered if not observed in the crystallographic electron ... Fig. 22) shows the relationships between the crystallographic resolution and the percentage of disordered residues. It can be seen that resolution tends to decrease if the amount of disorder increases, though the effect of disorder on resolution is not spectacular. In fact the average resolution decreases only from 2.13 to 2.45 Å if the disorder fraction increases from 2.5 to 32.5%. Dependence between the crystallographic resolution and the percentage of disordered residues observed in the crystal structures deposited in the Protein Data Bank. Vertical bars indicate the standard deviation of the mean. This clearly shows that protein 3D structures are often partially disordered and that a moderate fraction of conformationally disordered residues can be tolerated. Keeping this in mind, one can now try to predict if a protein has a reasonable probability to be suitable for a structural biology analysis. We designed a prediction method that is based on several individual prediction algorithms. The only necessary input is the amino acidic sequence of the protein and all the predictors are freely available. Each prediction algorithm must be used separately (Table 11) and its results must be inserted into equation (2), together with the coefficients xi reported in Table 22. If the value of p_cons is positive, the residue is predicted to be disordered and if it smaller than zero, the residue is predicted to be ordered. This can easily be done for each residue and, as a consequence, it is possible to reach a global picture of the conformational status of the protein. This new prediction method, which is essentially a weighted consensus approach, performs quite well, better than any individual prediction algorithm. Table 33 shows the values of several figures of merit, obtained with a 20-fold cross validation procedure. It can be seen that predictions are very accurate, with all the figures of merit larger than 80%. This is impossible by using individual predictors, though all of them have very high specificity. The probability excess, which is the best figure of merit because little influenced by the fact that the data are unbalanced, is equal to 80.1%, a value much larger than any other predictor. Performance of the New Prediction Methods Described in the Present Paper Compared to the Individual Prediction Methods of Table 1 It must be observed that the prediction reliability described above is based on the particular set of proteins available at the DISPROT database. Therefore, it would not be surprising to obtain other estimations of reliability by using different data. As a consequence, the reliability indicators shown in Table 33 cannot be used to rank various prediction methods according to their performances. It is however clear that the consensus approach presented in this manuscript is likely to be superior to all the individual methods on which it is based and it is also reasonable to suppose that an increase of experimental knowledge, which is likely to occur in the future, will allow more accurate predictions. ## ACKNOWLEDGEMENT This work was supported by the Austrian GEN-AU project BIN-II. Svetlana Kirillova is gratefully acknowledged for her comments. ## REFERENCES 1. Fink AL. Curr Opin Struct Biol. 2005;15:35–41. [PubMed] 2. Lee C, Yu MH. J Biochem Mol Biol. 2005;38:275–280. [PubMed] 3. Ferron F, Longhi S, Canard B, Karlin D. Proteins. 2006;65(1):1–14. [PubMed] 4. Bourhis JM, Canard B, Longhi S. Curr Prot Pept Sci. 2007;8:135–149. [PubMed] 5. Dosztanyi Z, Sandor M, Tompa P, Simon I. Curr Prot Pept Sci. 2007;8:161–171. [PubMed] 6. Jin Y, Dunbrack RL. Proteins. 2005;61(7):167–175. [PubMed] 7. Linding R, Jensen LJ, Diella F, Bork P, Gibson TJ, Russell RB. Structure (Camb) 2003;11(11):1453–9. [PubMed] 8. Dosztanyi Z, Csizmok V, Tompa P, Simon I. Bioinformatics. 2005;21:3433–3434. [PubMed] 9. Vucetic S, Obradovic Z, Vacic V, Radivojac P, Peng K, Iakoucheva LM, Cortese MS, Lawson JD, Brown CJ, Sikes JG, Newton CD, Dunker AK. Bioinformatics. 2005;21:137–140. [PubMed] 10. Carugo O. Curr Protein Pept Sci. 2007;8:119–120. 11. Carugo O. J Appl Cryst in the press 12. Smialowski P, Martin-Galiano AJ, Mikolajka A, Girschick T, Holak TA, Frishman D. Bioinformatics. 2006 13. Idicula-Thomas S, Kulkarni AJ, Kulkarni BD, Jayaraman VK, Balaji PV. Bioinformatics. 2006;22:278–284. [PubMed] 14. Smialowski P, Schmidt T, Cox J, Kirschner A, Frishman D. Proteins. 2006;62(2):343–355. [PubMed] 15. Rodrigues AP, Grant BJ, Hubbard RE. Nucleic Acids Res . 2006;(34 Web Server):W225–230. [PubMed] 16. Bernstein FC, Koetzle TF, Williams GJ, Meyer EF, Brice Jr MD, Rodgers JR, Kennard O, Shimanouchi T, Tasumi M. J Mol Biol. 1977;112(3):535–542. [PubMed] 17. Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, Shindyalov IN, Bourne PE. Nucleic Acids Res. 2000;28(1):235–42. [PubMed] 18. Hino T, Kanamori E, Shen JR, Kouyama T. Acta Cryst. 2004;D60:803–809. [PubMed] 19. Ward JJ, Sodhi JS, McGuffin LJ, Buxton BF, Jones DT. J Mol Biol. 2004;337:532–645. [PubMed] 20. MacCallum MR. Bioinformtics. 2004; 20:i224–i231. [PubMed] 21. Prilusky J, Felder CE, Zeev-Ben-Mordehai T, Rydberg EH, Man O, Beckmann JS, Silman I, Sussman JL. Bioinformatics. 2005;21:3435–3438. [PubMed] 22. Linding R, Russell RB, Neduva V, Gibson TJ. Nucleic Acids Res. 2003;31:3701–3708. [PubMed] 23. Coeytaux K, Poupon A. Bioinformatics. 2005;21:1891–1900. [PubMed] 24. Yang ZR, Thomson R, McNeil P, Esnouf RM. Bioinformatics. 2005;21:3369–3376. [PubMed] Articles from The Open Biochemistry Journal are provided here courtesy of Bentham Science Publishers ## Formats: ### Related citations in PubMed See reviews...See all... See all...
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 6, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8187398314476013, "perplexity": 2155.3824033766527}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416400379404.25/warc/CC-MAIN-20141119123259-00060-ip-10-235-23-156.ec2.internal.warc.gz"}
https://online.ucpress.edu/elementa/article/doi/10.12952/journal.elementa.000107/112896/Melting-ice-growing-trade-Melting-ice-growing
Large reductions in Arctic sea ice, most notably in summer, coupled with growing interest in Arctic shipping and resource exploitation have renewed interest in the economic potential of the Northern Sea Route (NSR). Two key constraints on the future viability of the NSR pertain to bathymetry and the future evolution of the sea ice cover. Climate model projections of future sea ice conditions throughout the rest of the century suggest that even under the most “aggressive” emission scenario, increases in international trade between Europe and Asia will be very low. The large inter-annual variability of weather and sea ice conditions in the route, the Russian toll imposed for transiting the NSR, together with high insurance costs and scarce loading/unloading opportunities, limit the use of the NSR. We show that even if these obstacles are removed, the duration of the opening of the NSR over the course of the century is not long enough to offer a consequent boost to international trade at the macroeconomic level. An increasing body of scientific evidence (Lemke and Jacobi, 2012; Overland et al., 2014) details the extent of climate change in the Arctic region. The accelerating melting of the sea ice cover has received particular attention (Stroeve et al., 2014; Boé et al., 2009; Deser et al., 2010; Stroeve et al., 2012; Overland and Wang, 2013). Since 2008, the Arctic Ocean along the coast of Russia and Canada has been mostly free of ice in September (Stroeve et al., 2014). Some international shipping companies have used this opportunity, as well as the fact that Russia opened the region in the far north to non-Russian vessel maritime traffic, to transport goods between Northern Europe and the Far East. Nevertheless, the numbers remain substantially lower compared with maritime trade passing through the Suez Canal (41 against 17,799 in 2011[1]) and also in terms of traffic volumes (1.25 compared with 740 million metric tons). However, given that on average the NSR is 3000 km shorter than the Suez Route and would require 10 less days of navigation than via Suez Canal, some experts have predicted a large potential for development in the coming years (Koranyi, 2013). Russia is also adapting regulations to facilitate vessels taking advantage of this opportunity. In particular, the number of permits issued by the Russian Northern Sea Route Administration, the route’s governing body, has grown from 4 in 2010 to 718[2] in 2013. China also recently set a record round trip time from Europe to north China through the NSR docking at Tianjin Port in October 2015 (China Daily, Monday, 26 October 2015 EDT, China mulls routine navigation through Arctic to Europe). Yet despite high hopes expressed by Russian authorities to open the route to increased maritime traffic (Klimenko, 2014), enthusiasm regarding the trade potential of the NSR remains low (Klimenko, 2014; Vidal, 2014). Moreover, the increase in human activities in the far north will largely come from the development of extractive activities (Stephenson et al., 2014). Currently, a large number of commercial vessels (e.g., oil tankers and gas carriers with a gross tonnage of at least 10,000) are allowed to navigate the NSR without ice strengthening, yet they can do so only in very good weather conditions and under the assistance of icebreakers (Ministry of Transport of Russia, 2013). This assistance has costs in terms of fees paid to the Russian state to maintain the icebreaker service, and in terms of navigation time, which is increased under escort. Only at the 2050 horizon may sea ice melt be significant enough to free the commercial ship from ice-breaker escort (Kupperman, 2014). In this paper we expand the forecast beyond 2050 to the end of this century, and evaluate whether the trade would increase between countries that could benefit from shorter trade distances using the NSR. To answer this question, we combined future climate model projections of the sea ice cover from 33 climate models participating in the World Climate Research Programme Coupled Model Intercomparison Project Phase 5 (CMIP5) with economic expertise to assess the trade potential of the route over the 21st century. Our first task was to define the route (Text S1, Table S1, and Figure S1) and determine how many months of the year the route will be open for shipping in the coming decades considering two future emission scenarios, denoted by RCP 4.5 and RCP 8.5 (see Methods). Our focus was on the combined impact of climate change and distance reduction on trade; hence we have stated optimal conditions for the shipping industry. That is to say, when the route was considered navigable, the states bordering the route would be responsible for all the security-related expenses to keep the passage safe. In addition, commercial ships of any kind were assumed to navigate the route at normal commercial speeds. Our models show that even in these conditions, the trade growth resulting from the use of the NSR as a trans-Arctic route remains particularly low. An important role of climate science in decision-making is to provide probabilistic predictions of potential outcomes; in this case, probabilistic predictions of navigable ice conditions in the NSR. Towards this end we assessed the number of months the route may be navigable on decadal time-scales from the multi-model ensemble mean probability of sea ice cover, sea ice thickness and a combination of both criteria within the route. The use of an ensemble mean serves to average out internal variability, leaving behind the expected ice conditions from external forcing; in this case, from future atmospheric greenhouse gas concentrations. The probability of the NSR being navigable each month is shown in Figure 1 for the RCP 4.5 and 8.5 emission scenarios. We additionally present our results in terms of navigable months in Figure 2 (RCP 4.5) and Figure 3 (RCP 8.5)[3] for three different ice probability cut-offs: 1) when there is 90% chance the NSR is covered by less than 15% ice cover and/or the NSR is covered by sea ice less than 15 cm thick; 2) when there is 70% chance the NSR is covered by less than 15% ice cover and/or the NSR is covered by sea ice less than 15 cm thick; and finally 3) when there is 50% chance the NSR is covered by less than 15% ice cover and/or the NSR is covered by sea ice less than 15 cm thick. Under the RCP 4.5 emission scenario, the route is not navigable when we use a 90% ice probability cut-off. At best, if we consider the route open when there is more than a 50% chance of less than 15% ice concentration within the route or the ice thickness along the route is less than 15 cm thick, we find that the route will only be navigable during the month of September between 2010 to 2030 and up to 3 months (August, September, and October) between 2030 and 2100. In other words, from 2030 onwards, there are 3 months in summer that meet the 50% chance the route may have less than 15% ice concentration or consist of ice less than 15 cm thick. With this relaxed requirement, the states managing the NSR would need to have very reliable infrastructure and ice-breaker service to support commercial shipping as the probability of adverse sea ice conditions remains high: i.e., there is a 50% chance of encountering high ice concentrations or ice thicker than 15 cm. Figure 1. Probability of having ice on the Northern Sea Route. Probabilities are provided as percentages by month and by decade for ice concentration < 15% (upper panels), ice thickness < 15 cm (middle panels), and either ice concentration < 15% or ice thickness < 15 cm (lower panels) for two future emission scenarios: representative concentration pathway (RCP) 4.5 (left panels) and RCP 8.5 (right panels). Figure 1. Probability of having ice on the Northern Sea Route. Probabilities are provided as percentages by month and by decade for ice concentration < 15% (upper panels), ice thickness < 15 cm (middle panels), and either ice concentration < 15% or ice thickness < 15 cm (lower panels) for two future emission scenarios: representative concentration pathway (RCP) 4.5 (left panels) and RCP 8.5 (right panels). Figure 2. Projected number of months the Northern Sea Route would be opened to navigation according to RCP 4.5. This projection is based on the future emission scenario representative concentration pathway (RCP) 4.5. Three ice probability thresholds are presented: 90, 70 and 50% chance that the Northern Sea Route (NSR) is covered by < 15% ice concentration (upper panels), by sea ice < 15 cm thick (middle panels), and by combination (< 15% ice concentration or < 15 cm ice thickness; lower panels). Figure 2. Projected number of months the Northern Sea Route would be opened to navigation according to RCP 4.5. This projection is based on the future emission scenario representative concentration pathway (RCP) 4.5. Three ice probability thresholds are presented: 90, 70 and 50% chance that the Northern Sea Route (NSR) is covered by < 15% ice concentration (upper panels), by sea ice < 15 cm thick (middle panels), and by combination (< 15% ice concentration or < 15 cm ice thickness; lower panels). Under the RCP 8.5 emission scenario the minimum number of months the route is navigable during this century is 50 months for the most restrictive cut-off (90%) using the calculation based on ice thickness. The maximum number of months the route is navigable is 330 months throughout the century for the most relaxed cut-off (50%) using either the ice thickness or ice concentration criterion. Under this scenario, only the month of September is ice-free for nine decades starting in 2010 (as in the RCP 4.5 emission scenario), whereas ice-free conditions expand to include the months of August through October for 8 decades starting in 2020; the months of August through November for 4 decades starting in 2060, and the months of July through November for 3 decades starting in 2070. At the end of the century, a 6-month navigation period may be possible from July through December. Table 1. Percentage increase in trade over the 21st century according to NSR opening scenariosa aBased on ice concentration and ice thickness criteria, and a combination of ice concentration and thickness for RCP 4.5 and RCP 8.5 emission pathways. bCombination criteria is based on < 15% ice concentration or <15 cm ice thickness. These results suggest that use of the NSR is unlikely to significantly alter international trade patterns, and will not be a leading economic force in transforming the Arctic. That is not to say that the Arctic is not going through a period of rapid transformation. Instead, these changes are led by the growth of extractive industries, in particular offshore oil and gas (Brigham, 2011). The existence of important oil and gas resources under the seabed of the Kara Sea has been known since the late 1980s (CIA, 1988). In September 2014, the Russian company Rosneft announced the first successful drilling of a well in the Kara Sea with its partner Exxon (Rosneft, 2014). The continuation of exploration operations has been postponed because of sanctions targeting Russian Arctic oil projects following Russia’s actions in Ukraine (Reuter, 2015). Nonetheless, this event reveals that political decisions so far have a greater potential to slow down activities than climatic environmental concerns have. In terms of traffic in the NSR, the oil and gas industry have already been important players. According to the Northern Sea Route Information Office, more than 50% of the vessels that used the NSR in 2014 were oil tankers (http://www.arctic-lio.com/nsr_transits). As the industry expands and new oil and gas fields are developed, an increasing number of vessels will be used to resupply offshore facilities from mainland Russia and to transport these commodities to their consumption markets. There is significant distance savings between Europe and Asia offered by the NSR (Humpert, 2011), yet the lack of schedule reliability due to the difficulty in being able to predict the start and the end dates of summer ice conditions in the Arctic, combined with the lack of infrastructure in the hinterland could prevent the route from becoming popular with liner services. In contrast, for bulk dry carriers and wet carriers, the route may become an alternative to more traditional shipping routes in the near future. Indeed, transit statistics collected by the Northern Sea Route Information Office indicate that the number of voyages increased from 41 in 2011 (the first year transit statistics were made available) to 71 in 2013 (the year that recorded the highest number of transits between 2011 and 2014: see Table S2), out of which liquid and bulk carriers increased from 20 in 2011 to 34 in 2013. In terms of route direction, the number of vessels increased from 11 to 41 sailing from East to West while the number of vessels from West to East stayed the same at 30 vessels. However, 17 of the latter were empty vessels in 2013. Note that “transit” as used by the NSR Information Office is defined as every ship passing one of the Russian-defined boundaries of the NSR zone (see Text S1). In 2013 among these passages, 43 were passages of the East and West boundaries of the NSR and only 17 were truly trans-arctic passages; i.e., passages where the points of origin and destination were outside the arctic zone (http://www.arctic-lio.com/nsr_transits). The economic potential, in terms of profits for ship owners, of using the NSR as an alternative route between Asia and Europe has been investigated as a function of navigable months of the NSR, Russian NSR fees, and bunker price (Lasserre and Pelletier, 2011). The conditions under which the NSR is competitive compared with the Suez Canal as a function of these variables has been evaluated under several scenarios. The profits under each combination of conditions are then calculated and compared for the case of non-ice container ships (4300 Twenty foot Equivalent Unit) using the Suez route and similar size ice-classed container ships (Class 1B) using either the NSR when open, and the Suez route otherwise. In this situation, the NSR is only competitive when the price of bunker fuel is low, the NSR is open for at least 3 months of the year, and the Russian NSR fees are reduced by 85% of the 2010 level. In the same way, the technical and economic feasibility of regular container transport along the NSR was evaluated (Liu and Kronbak, 2010). By taking the schedule between Shanghai and Hamburg, the relative costs of various axes in the Asia–Europe transport network were assessed, indicating that while shipping through the Suez Canal is still by far the least expensive option, the NSR and Trans-Siberian Railway appear to be roughly equivalent second-tier alternatives. The impact of bathymetry was previously investigated and suggests that shipping companies might envision several trade routes in the Arctic to avoid the shallowest passages in the region (in particular, the Laptev and Sanikov Straits) (Stephenson et al., 2014). Nevertheless, in the coming decade ice-strengthened vessels (PC3 category) will still have the capacity to operate for a significantly longer period of time as sea ice continues to decline in the Arctic. Our results provide an upper limit of the potential increase in international trade. Indeed, we considered only the positive aspect of distance reduction on trade without tackling several emerging problems for shipping companies from using the NSR even in very good ice conditions (from the point of view of the transport industry). These problems include the lack of port of calls along the Russian coast and/or the shallowness of the route (Verny and Grigentin, 2009), both of which would dampen the positive impact of distance reduction on trade. Supplemental Table S3 summarizes the loss or gain in terms of distance and days of navigation of the NSR between Europe and Asia for a number of selected ports. As indicated in Table S3 the NSR, which extends for about 3000 miles, depending on ice conditions, presents opportunities for international trade as a transit route between Europe and the Asia-Pacific region. More specifically, according to Table S3, Japan, South Korea and China are the three countries that benefit the most in terms of distance and time reductions. Russia, mainly concerned by the end point traffic, was not included in our study. ### Sea ice projections Key to understanding the potential of the Northern Sea Route is to know how long ice-free summer conditions or very thin ice conditions will last in the next several decades. An ice-free Arctic ocean is generally considered to be an Arctic Ocean with less than 1 million km2 of ice cover (Stroeve et al., 2012). Nonetheless, we only need the NSR, as we have defined it, to be either ice-free or else covered by very thin ice. To estimate the probability of these conditions in the 21st century, we used simulations from models participating in the World Climate Research Programme Coupled Model Intercomparison Project Phase 5 (CMIP5) (See Table S4 for a listing of the models used) and systemically estimated ice conditions within the NSR for all decades to the end of the century. To be considered ice-free, all pixels, corresponding to a 25 km2 area, along the route must have had an ice concentration below 15%. This definition of an ice-free route is very close to the definition used by the World Meteorological Organization (WMO) for open water with an ice concentration below 10% (World Meteorological Organization, 1970). Note that regardless of the ice thickness of that remaining 15% sea ice concentration, we remain on very solid ground regarding navigability. We used the Transport Canada Ice Numeral (IN) calculation formula (Transport Canada, 1998): $IN=(Ca*IMa)+(Cb*IMb)+…+(Cn*IMn)$ where Ca, Cb...Cn stand for the ice concentrations in tenths of ice type a, b...n, and IMa, IMb IMn the Ice Multiplier of ice type a, b...n. The Ice Multiplier (IM) is an integer number from 2 to –4 associated with the danger presented for navigation by different types of ice, with –4 being the most dangerous, and pertains to a specific of class of ship. We calculated the Ice Numeral for a Type E ship (Class 1D and 1E in the Lloyd register or Class II in the Swedish-Finish ice classes). According to the Finnish Transport Safety Agency, ice Class II ships are defined as “ships that have a steel hull and that are structurally fit for navigation in the open sea and that, despite not being strengthened for navigation in ice, are capable of navigating in very light ice conditions with their own propulsion machinery”; that is to say, most of the commercial open water fleet (Transport Safety Agency, 2010). In the worst case, where 15% of the area is covered by multiyear ice (IM = –4), and the rest is open water (IM = 2), the Ice Numeral for Type E ship is 11. A positive number means that the route is navigable. In this work, ice thickness was defined as the average thickness of the pixel. According to Transport Canada (1998), conditions dominated by grey ice, defined by an ice thickness less than 15 cm, are fairly navigable for Type E ships. We considered the NSR navigable if all pixels along the route had an ice thickness less than 15 cm. Note that even if the entire area (100%) is covered by grey ice (IM = 1), the Ice Numeral for Type E ship is 10. To assess future navigability, two future emission scenarios were considered: the representative concentration pathways (RCPs) 4.5 and 8.5. Total radiative forcing is stabilized at 4.5 W m–2 in the year 2100 for the RCP 4.5 and assumes a certain level of moderate greenhouse gas mitigation to curb the increase of the level of human greenhouse gas emission. RCP 8.5 stabilizes radiative forcing at 8.5 Wm–2 and assumes no greenhouse gas mitigation. We processed 92 ensembles from 32 climates models for the RCP 4.5 scenario and 61 ensembles from 28 climates models for the RCP 8.5 scenarios. For each model ensemble member, we calculated the frequency of ice-free conditions (defined here as sea ice concentration less than 15%) for each decade starting in 2000 and ending in 2100. Frequencies were calculated on the model native grids. Where models have multiple ensemble members, a model ensemble mean frequency was calculated. The resulting frequency grids were then regridded to a common 25 km EASE grid (Brodzik and Knowles, 2002) and a multi-model ensemble mean was calculated. The frequency of ice thickness was treated in a similar manner. The use of a multi-model ensemble mean in effect averages out natural variability (assuming the models sufficiently capture the amount of the observed variability), leaving the expected changes due to rising greenhouse gases. Results are expressed as the probability (in percent) that the ice concentration for each pixel composing the route is less than 15% over decadal time periods, or that the ice thickness is less than 15 cm, or that either of the two criteria is fulfilled (Figure 1). Figure 2 summarizes the number of months that the route can be considered navigable (we used several ice probability cut-offs from more than a 90% chance that the route is ice-free to more than a 50% chance of ice-free conditions) in the coming decades under the RCP 4.5 and 8.5 emission scenarios. September is the month that witnesses the largest reduction of ice concentration and ice thickness over the century. The number of models with an ensemble mean September sea ice concentration greater than 15% in the NSR region is shown for three decades: 2000–2009 (Figure 4), 2040–2049 (Figure 5) and 2090–2099 (Figure 6). Figure 3. Projected number of months the Northern Sea Route would be opened to navigation according to RCP 8.5. This projection is based on the future emission scenario representative concentration pathway (RCP) 8.5. Three ice probability thresholds are presented: 90, 70 and 50% chance that the Northern Sea Route (NSR) is covered by < 15% ice concentration (upper panels), by sea ice < 15 cm thick (middle panels), and by combination (< 15% ice concentration or < 15 cm ice thickness; lower panels). Figure 3. Projected number of months the Northern Sea Route would be opened to navigation according to RCP 8.5. This projection is based on the future emission scenario representative concentration pathway (RCP) 8.5. Three ice probability thresholds are presented: 90, 70 and 50% chance that the Northern Sea Route (NSR) is covered by < 15% ice concentration (upper panels), by sea ice < 15 cm thick (middle panels), and by combination (< 15% ice concentration or < 15 cm ice thickness; lower panels). Figure 4. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2000–2009. The red line shows the Northern Sea Route (NSR). Figure 4. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2000–2009. The red line shows the Northern Sea Route (NSR). Figure 5. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2040–2049. The red line shows the Northern Sea Route (NSR). Figure 5. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2040–2049. The red line shows the Northern Sea Route (NSR). Figure 6. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2090–2099. The red line shows the Northern Sea Route (NSR). Figure 6. Percentage of models with ensemble mean September sea ice concentration greater than 15% for 2090–2099. The red line shows the Northern Sea Route (NSR). ### The gravity model of trade To estimate the elasticity to distance the gravity model is specified as: () $Tradeijt=YitYjtYwt(τijtPitPjt)1−σ$ 1 where Tradeijt represents the exports or imports from country i to country j in period t; Yit (Yjt) indicates the GDP of the exporter i (importer j) in period t, Ywt is the world nominal income, Pit and Pjt are country time-specific multilateral resistance terms, and σ is the elasticity of substitution between all goods. The bilateral trade costs are represented by τijt, which is assumed to be a log-linear function of observables trade costs zijt (cost directly related to transport but also to tariff, language, etc.) (Anderson and van Wincoop, 2004): () $τijt=Πm=1M(Zijtm)γm$ 2 where there are M different trade costs and γ is the weight of each trade cost in the total trade cost. In log-linear form the gravity model is: () $lnTradeijt=lnYit+lnYjt−lnYwt+Σm=1M(1−σ)γmln(ZijtM)−(1−σ)lnPit−(1−σ)lnPjt$ 3 Disdier and Head (2008) use the results obtained in 103 papers that estimate the elasticity of trade to distance ((1–σ)γm=distance) to perform a meta-analysis. According to the authors the average elasticity of trade to distance, is –0.9. Although the measure of distance used in these papers is far from homogeneous, namely great circle distances, road or maritime distance, the authors do not find any significant difference in the value of the coefficient across the different measures. In this study, we used this measure of trade elasticity to distance and considered all other factors held constant. We focused on the sole impact of distance savings of the NSR on international trade and ignored other impediments to trade, such as the current level of infrastructure along the route, tolls imposed by Russian authorities to use the route or additional insurance costs. These elements would obviously reduce the positive impact of the distance reduction. Hence, we consider that if the shipping companies would have the choice between two trade routes, the Northern trade route and the Suez route, they would use the shortest. The main source of data is Eurostat (http://ec.europa.eu/eurostat/data/database), where we obtained monthly bilateral maritime exports and imports in Euros for the first decade of the 21st century (2000–2010). We obtained port-to-port distance for the NSR and the Suez Route data from the Port World website (http://www.portworld.com/map). As we are interested in transit trade, we did not include Russia in our group of European countries. We also did not include European and Asiatic landlocked countries. Table S5 lists the European and Asian countries we selected, for which distance reduction is possible through the NSR and comprehensive monthly trade data available over our period of study. We used the average monthly total trade (Export + Import) calculated over the period 2000–2010 as our baseline for our estimations of the impact of the NSR. Having the elasticity of distance to trade, the “gain” in terms of distance of using the shortest route and the current value of bilateral trade, we were able to calculate the gain in terms of trade value of using the NSR for the months when the route would be opened, everything else held equal. We then multiplied these values by the number of months the route is opened under each scenario, and compared the value to a baseline scenario where trade flow is held constant over the coming 10 decades of this century (columns 2, 3 and 4 in Table 1)[4]. We finally calculated the equivalent annual increase (columns 5, 6 and 7 in Table 1)[5]. The data about the characteristics of the vessels having the use the NSR in 2013 and their geo-localizations are available upon demand and subscription to the marinetraffic’s website (http://www.marinetraffic.com/). © 2016 Bensassi et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. 1. The number for the NSR comes from the Northern Sea Route Information Office (http://www.arctic-lio.com/). The office is a joint initiative of the Norwegian Centre for High North Logistics (CHNL) and ROSATOM, the Russian state-run company in charge of Atomic Energy production and use. The number includes not only the ships transiting via the NSR but also all the voyages starting or ending on the NSR. There are only 22 transit voyages and 2 with a non-Russian starting or ending point. The number for the Suez Canal comes from the Suez Canal authority (http://www.suezcanal.gov.eg/). 2. This number comes from The Northern Sea Route Administration (http://www.nsra.ru), a branch of the Russian Ministry of Transport. It includes all kinds of ships applying to use whole or just parts of the NSR. Applying and receiving a permit carries no cost, so the number of applications is an indicator of the interest in the NSR from the shipping industry, not a precise indicator of the shipping activity on the NSR. 3. Only the months from July to December are presented, as our results show that for the rest of the year the route is not navigable during this century. 4. Elasticity of trade to distance, noted as β, where: $β=∂TradeTrade∂DistanceDistance,∂trade=Distancetrade*β*∂Distance,andpercentagechangeoverthecentury=Σ20002099[(100−nm)*tradem+nm*(1+∂m)tradem]trade2000,2010*10,$ where nm is the number of months a particular month is navigable over the century 5. Equivalent annual percentage increase i, where $i=exp[ln(1+i100)100]−1$ and i100 is the percentage increase over the century We would like to thank the participants of the International Economics Workshop held in Goettingen and the ETSG Conference for their helpful comments and suggestions, which have been incorporated into the paper. The authors acknowledge the support of NOAA Climate Program Office Modeling, Analysis, Predictions and Projections (MAPP) Program as part of the CMIP5 Task Force. Anderson JE . 1979 . A Theoretical Foundation for the Gravity Equation . Am Econ Rev 69 ( 1 ): 106 116 . Anderson JE , van Wincoop E . 2003 . Gravity with Gravitas: A Solution to the Border Puzzle . Am Econ Rev 93 ( 1 ): 170 192 . Anderson JE , van Wincoop E . 2004 . . J Econ Lit 42 ( 3 ): 691 751 . Bekker E , Rojas-Romagosa H , Francois JF . 2015 . Melting Ice Caps and the Economic Impact of Opening the Northern Sea Route . CPB Discussion Paper 307 . CPB Netherlands Bureau for Economic Policy Analysis . Bergstrand JH . 1985 . The Gravity Equation in International Trade: Some Microeconomic Foundations and Empirical Evidence . Rev Econ Stat 67 ( 3 ): 474 481 . Boé J , Hall A , Qu X . 2009 . September sea-ice cover in the Arctic Ocean projected to vanish by 2100 . Nat Geosci 2 ( 5 ): 341 343 . Brigham LW . 2011 . Marine protection in the Arctic cannot wait . Nature 478 : 157 . Brodzik MJ , Knowles KW . 2002 . EASE-Grid: A Versatile Set of Equal-Area Projections and Grids , in Goodchild M , ed., Discrete Global Grids . Santa Barbara, California USA : National Center for Geographic Information & Analysis . CIA . 1988 . The Kara Sea: A Soviet Oil Resource for the Turn of the Century . Freedom of Information Act Electronic Reading Room . http://www.foia.cia.gov. Deser C , Tomas R , Alexander M , Lawrence D . 2010 . The Seasonal Atmospheric Response to Projected Arctic Sea Ice Loss in the Late Twenty-First Century . J Clim 23 ( 2 ): 333 351 . Disdier AC , K . 2008 . The Puzzling Persistence of the Distance Effect on Bilateral Trade . Rev Econ Stat 90 ( 1 ): 37 48 . Eide LI , Eide A , Endresen Ø . 2010 . Shipping across the Arctic Ocean: A feasible option in 2030-2050 as a result of global warming? Research and Innovation Position Paper . DNV . Humpert M . 2011 . The Future of the Northern Sea Route - A “Golden Waterway” or a Niche Trade Route . The Arctic Institute . Klimenko E . 2014 . Russia’s Evolving Arctic Strategy . SIPRI Policy Paper . Stockholm International Peace Research Institute . Koranyi B . 2013 . Ice levels, rule changes to boost Arctic northern sea route . Reuters Press Release 29 May 2013 . Accessed 04 July 2013 . http://www.reuters.com/article/shipping-arctic-idUSL5N0EA0RF20130529. Kupperman J . 2014 . The Northern Sea Route: Could it be the new Suez Canal? Berkeley Political Review, December 22, 2014 . Lasserre F , Pelletier S . 2011 . Polar super seaways? Maritime transport in the Arctic: An analysis of shipowners’ intentions . Journal of Transport Geography 19 ( 6 ): 1465 1473 . Lemke P , Jacobi HW . 2012 . Arctic Climate Change: The ACSYS Decade and Beyond . Dordrecht : New York, Springer . Liu M , Kronbak J . 2010 . The potential economic viability of using the Northern Sea Route (NSR) as an alternative route between Asia and Europe . Journal of Transport Geography 18 ( 3 ): 434 444 . Martínez-Zarzoso I . 2011 . The log of gravity revisited . Appl Econ 45 ( 3 ): 311 327 . Ministry of Transport of Russia . 2013 . Rules of navigation on the water area of the Northern Sea Route . Annex 2 . January 17, 2013 . Overland JE , Wang M . 2013 . When will the summer Arctic be nearly sea ice free? Geophys Res Lett 40 ( 10 ): 2097 2101 . Overland JE , Wang M , Walsh JE , Stroeve JC . 2014 . Future Arctic climate changes: Adaptation and mitigation time scales . Earth’s Future 2 ( 2 ): 68 74 . Reuter . 2015 . Exclusive: Russia’s Rosneft will not resume drilling in Kara Sea in 2015 . Press Release 30 January 2015 . Accessed 23 February 2015 . http://www.reuters.com/article/2015/01/30/us-russia-crisis-rosneft-arctic-idUSKBN0L31D120150130. Rosneft . 2014 . Rosneft Discovered a New Hydrocarbon Field in the Kara Sea . Press Release 27 September 2014 . Accessed 21 October 2014 . http://www.rosneft.com/news/pressrelease/27092014.html. Russian Federation . 2012 . Federal Law on amendements to specific legislative acts of the Russion Federation related to governmental regulation of merchant shipping in the water area of the Northern Sea Route . N132-03 . T. R. Federation . Smith LC . 2010 . The World in 2050: Four Forces Shaping Civilization’s Northern Future . New York : Dutton, Penguin Group USA . Smith LC , Stephenson SR . 2013 . New Trans-Arctic shipping routes navigable by midcentury . 110 ( 13 ): E1191 E1195 . Stephenson SR , Brigham LW , Smith LC . 2014 . Marine accessibility along Russia’s Northern Sea Route . Polar Geography 37 ( 2 ). doi: . Stroeve J , Hamilton LC , Blitz MC , Blanchard-Wrigglesworth E . 2014 . Predicting September sea ice: Ensemble skill of the SEARCH Sea Ice Outlook 2008–2013 . Geophys Res Lett 41 ( 7 ): 2014GL059388. Stroeve JC , Kattsov V , Barret A , Serreze M , Pavlova T , et al 2012 . Trends in Arctic sea ice extent from CMIP5, CMIP3 and observations . Geophys Res Lett 39 ( 16 ): L16502 . Tinbergen J . 1962 . Shaping the World Economy: Suggestions for an International Economic Policy . New York : Twentieth Century Fund . Transport . 1998 . Arctic Ice Regime Shipping System (AIRSS) Standards . TP12259E. Marine Safety – Program Services - AMSB . Ottawa : . Transport Safety Agency . 2010 . Ice Class Regulations and the Application Thereof . TRAFI/31298/03.04.01.00/2010 . Helsinki : Transport Safety Agency . Verny J , Grigentin C . 2009 . Container shipping on the Northern Sea Route . Int J Prod Econ 122 ( 1 ): 107 117 . Vidal J . 2014 . Russian Arctic city hopes to cash in as melting ice opens new sea route to China . The Guardian . London . World Meteorological Organization . 1970 . Sea-ice Nomenclature . Terminology, codes and illustrated glossary. (WMO/OMM/BMO No 259.TP.145.) 2 vols in 1. 147 p, illus . Geneva : Secretariat of the World Meteorological Organization, 1970 . The Spanish Ministry of Economy and Competitiveness (project ECO2014-58991-C3-2-R) the Universitat Jaume I (P1- 1B2013-06), and the Centre for Environment Energy Economics and Management at the University of Birmingham provided funding. This paper was additionally funded under NOAA grant #NA10OAR4320142 and NSF grant #PLR 1304246. ## Competing Interests The authors have no competing interests to declare. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 6, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4006703197956085, "perplexity": 3253.3705803936045}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320304600.9/warc/CC-MAIN-20220124185733-20220124215733-00489.warc.gz"}
https://www.arxiv-vanity.com/papers/1507.00959/
arXiv Vanity renders academic papers from arXiv as responsive web pages so you don’t have to squint at a PDF. Read this paper on arXiv.org. # Asymmetric Dark Matter Stars Chris Kouvaris    Niklas Grønlund Nielsen CP-Origins & Danish Institute for Advanced Study DIAS, University of Southern Denmark, Campusvej 55, DK-5230 Odense M, Denmark ###### Abstract We study the possibility of asymmetric dark matter with self-interactions forming compact stable objects. We solve the Tolman-Oppenheimer-Volkoff equation and find the mass-radius relation of such “dark stars”, their density profile and their Chandrasekhar mass limit. We consider fermionic asymmetric dark matter with Yukawa-type self-interactions appropriate for solving the well known problems of the collisionless dark matter paradigm. We find that in several cases the relativistic effects are significant. Preprint: CP3-Origins-2015-027 DNRF90, DIAS-2015-27 ## I Introduction Lately, an emerging amount of issues indicates that the Collisionless Cold Dark Matter (CCDM) paradigm is at odds with astrophysical observations. The first and most well known issue is related to the fact that dwarf galaxies have a flat density core Moore:1994yx ; Flores:1994gz . Dwarf galaxies are dominated by dark matter (DM) and the flatness of the density profile in the core of the galaxy is in contradiction with the cuspy profiles predicted by numerical simulations of CCDM Navarro:1996gj . Numerical simulations of CCDM predict also a larger number of satellite galaxies in the Milky Way than what has been observed so far Klypin:1999uc ; Moore:1999nt ; Kauffmann:1993gv . Although our galaxy might simply be a statistical fluctuation Liu:2010tn ; Tollerud:2011wt ; Strigari:2011ps there might be dim galaxies yet to be observed, it is possible that DM is simply not collisionless. Furthermore, another related issue is the so-called “too big to fail” problem BoylanKolchin:2011de , i.e. CCDM numerical simulations predict massive dwarf galaxies that are too big to not have visible stars and thus to not be observed. Although some of the aforementioned issues can be resolved upon assuming the existence of baryonic-DM interactions Oh:2010mc ; Brook:2011nz ; Pontzen:2011ty ; Governato:2012fa , another probably more attractive and natural possibility is the existence of DM self-interactions. Clearly such interactions would flatten out cuspy dwarf galaxy cores and they could possibly also resolve the satellite galaxies issues Vogelsberger:2012ku ; Rocha:2012jg ; Zavala:2012us ; Peter:2012jh . DM self-interactions have already been proposed and studied in different contexts Spergel:1999mh ; Wandelt:2000ad ; Faraggi:2000pv ; Mohapatra:2001sx ; Kusenko:2001vu ; Loeb:2010gj ; Kouvaris:2011gb ; Rocha:2012jg ; Peter:2012jh ; Vogelsberger:2012sa ; Zavala:2012us ; Tulin:2013teo ; Kaplinghat:2013xca ; Kaplinghat:2013yxa ; Cline:2013pca ; Cline:2013zca ; Petraki:2014uza ; Buckley:2014hja ; Boddy:2014yra ; Schutz:2014nka . DM numerical simulations including self-interactions favor a DM-DM cross section per DM mass between . Within this range, DM self-interactions can solve the cusp vs core problem of dwarf galaxies as well as the “too big to fail” one. However, DM self-interactions cannot be arbitrarily strong. There are several constraints imposed on them. First of all one should make sure that DM-DM interactions are not sufficiently strong to destroy the ellipticity of spiral galaxies Feng:2009mn ; Feng:2009hw or dissociate the subclusters of the bullet cluster Markevitch:2003at . In addition, fermionic asymmetric DM with attractive Yukawa-type self-interactions can lead in some cases to formation of destructive black holes in the interior of old neutron stars, thus imposing extra constraints Kouvaris:2011gb . Further constraints are imposed in the case where the mediator of the DM-DM force couples to the Standard Model. The mediator can simply couple to the Standard Model via e.g. a Higgs portal Burgess:2000yq ; Patt:2006fw ; Andreas:2008xy ; Andreas:2010dz ; Djouadi:2011aa ; Pospelov:2011yp ; Greljo:2013wja ; Bhattacherjee:2013jca . In such a case one should make sure that decays before the start of the Big Bang Nucleosynthesis. The fact that a minimum strength between baryons and is required for not spoiling the BBN predictions can lead to significant rates of DM collisions in underground detectors that can exclude such models Kaplinghat:2013yxa . However constraints like these can be evaded if e.g. couples also to sterile or active neutrinos Kouvaris:2014uoa . One should emphasize that the above constraints are model dependent and therefore although there is no clear universal region where DM self-interactions are allowed, the region which accommodate the resolution of the dwarf galaxies problems is roughly speaking constraint free. If DM experiences self-interacting forces, it is possible to imagine that star-like compact objects can be formed. Whatever the mechanisms of forming such objects, one should make sure that they do not violate the limits imposed by the MACHO Alcock:2000ph and EROS Tisserand:2006zx experiments. Based on the microlensing technique, these experiments claimed that less than of DM can be in the form of compact objects between the mass range , where is the solar mass. The possibility of stars made of DM has been studied before in the context of annihilating DM forming dark stars in the early universe Spolyar:2007qv ; Freese:2008hb ; Freese:2008wh . It has been also studied in the context of hybrid compact stars made of baryonic and DM Leung:2011zz ; Leung:2013pra . In the latter case, neutron stars and white dwarfs include a significant amount of DM in their interior modifying thus the equation of state of the star. Furthermore the possibility of black hole formation from strongly self-interacting components of DM was studied recently in Pollack:2014rja . In this paper we examine the possibility that asymmetric DM with self-interactions appropriate for solving the core vs cusp problem, the “satellite problem” and the “too big to fail problem” forms star-like compact objects. Asymmetric DM Nussinov:1985xr ; Barr:1990ca ; Gudnason:2006yj ; Foadi:2008qv ; Dietrich:2006cm ; Sannino:2009za ; Ryttov:2008xe ; Sannino:2008nv ; Kaplan:2009ag ; Frandsen:2009mi ; MarchRussell:2011fi ; Frandsen:2011cg ; Gao:2011ka ; Arina:2011cu ; Buckley:2011ye ; Lewis:2011zb ; Davoudiasl:2011fj ; Graesser:2011wi ; Bell:2011tn ; Cheung:2011if has become an attractive alternative to thermally produced DM not only because it can relate theories beyond the Standard Model to DM, but because it can also provide a link between baryogenesis and dark-genesis. For recent reviews on asymmetric DM see Petraki:2013wwa ; Zurek:2013wia . We are going to assume that the self-interactions are Yukawa-type and can be either attractive (mediated by a scalar ) or repulsive (mediated by a vector boson ). Upon these assumptions, we study the stability of dark stars formed by asymmetric fermionic DM. We solve the Tolman-Oppenheimer-Volkoff equation and study the hydrostatic equilibrium of these compact objects. We find their density profile, the mass vs radius relation as well as the Chandrasekhar mass, i.e. the maximum mass where these objects are stable. The paper is organised as follows: In section II we present the equation of state for DM with self-interactions. In section III we show the relevant parameter phase space of self-interactions that solve the problems of CCDM we have mentioned earlier. In section IV we present the equations for the hydrostatic equilibrium of the dark stars. In section V we use a simplified Newtonian analysis to get a first understanding of the problem, while we present the full relativistic results in section VI. We conclude in section VII. Throughout this paper we use natural units , and define the Planck mass as . ## Ii Equation of State As we mentioned we assume that DM is of asymmetric type and fermionic. We also assume that DM self-interactions are mediated by via a Yukawa coupling of the form (in case of attractive interactions) where is the DM particle and is the Yukawa coupling constant or (in case of repulsive interactions). We wish to obtain the equation of state of DM under the aforementioned assumptions. The energy density of DM particles consists of two components ρ=ρkin+ρY, (1) where , and are the kinetic energy density, and Yukawa potential energy respectively. The pressure of the system is Shapiro:1983du P=n2ddn(ρn), (2) where is the number density of DM particles. The pressure related to the kinetic energy in the non-relativistic () or relativistic () limits takes the simple form of a polytrope , where equals or for the non-relativistic and relativistic case respectively. We choose to work with the full relativistic dispersion . If we assume that the temperature of DM particles is much smaller than their Fermi energy, i.e. we effectively take the limit , the number density, kinetic energy density and the corresponding pressure are given by n=gs(2π)3∫pF04πp2dp=gsm3χ6π2x3, (3) ρkin=gs(2π)3∫pF0E(p)4πp2dp=gs2m4χξ(x), (4) Pkin=13gs(2π)3∫pF0p2E(p)4πp2dp=gs2m4χψ(x), (5) where the functions and are defined as ξ(x)=18π2{x√1+x2(1+2x2)−ln[x+√1+x2]}, (6) ψ(x)=18π2{x√1+x2(2x2/3−1)+ln[x+√1+x2]}. (7) is a measure of how relativistic the particles are, and is the spin multiplicity. The Yukawa potential between two particles is Vij=±αe−μrijrij, (8) where is the mass of the mediator, is the separation between the particles and is the coupling to the dark mediator. In order to find the Yukawa potential energy of the entire system we have in principle to sum over all pairs of DM particles, which we approximate as an integration over volume elements. EY=12∑i≠jVij=±12n2α∫∫e−μrijrijdVidVj. (9) In the case where the radius of the star satisfies (i.e. the potential is short range), it is a reasonable approximation to integrate the volume up to infinity (instead of the volume of the star). This leads to the following Yukawa energy density ρY=±2παn2μ2=±αg2s18π3m6χμ2x6. (10) This estimate for the Yukawa energy density gives us the final expressions for and : P=gs2m4χψ(x)±αg2s18π3m6χμ2x6, (11) ρ=gs2m4χξ(x)±αg2s18π3m6χμ2x6. (12) Since neither equation can be inverted analytically, we must work with two equations of state, and have an implicit relation between and . DM particles in the attractive scenario (corresponding to the minus sign in the Yukawa contribution) cannot become arbitrarily relativistic, since the pressure and density must be positive. The positiveness of pressure and density give an upper bound on . ## Iii Parameter Space of Self-Interactions As we mentioned earlier, DM self-interactions within the range cm/g can solve the problematic issues of CCDM, while astrophysical constraints limit these interactions between cm/g. Following Tulin:2013teo , in order to determine the parameter space of DM and mediator masses (for a given coupling) that lies in the aforementioned range, we introduce the transfer cross section . We use a typical value of km/s for the average velocity of DM in a dwarf galaxy, and we estimate the velocity averaged cross section as σ=∫d3ve−(v/v0)2/2(2πv20)3/2σT(v). (13) The transfer cross section for attractive Yukawa interactions in the classical limit is Feng:2009hw ; Finkbeiner:2010sm (14) The corresponding cross section for repulsive Yukawa interactions interactions reads (15) where . For the opposite limit , we follow Tulin:2013teo and approximate the Yukawa by a Hulthén potential with proper parameter choices. In the region we interpolate between the two regimes. In Fig. 1 we show the allowed parameter space for DM and mediator mass (in GeV) in the case of repulsive and attractive interactions for three distinct values . Deep (light) blue is the region that g (g) in dwarf galaxies, solving the aforementioned issues of CCDM. The red solid (dashed) line shows the curve where g (g) in the Milky Way. The phase space to the left of the red curve is excluded because the cross section is sufficiently large to smooth out the ellipticity of Milky Way to a degree inconsistent with observations. There is a bit of ambiguity regarding the value of the maximum consistent with observations but it should be between Tulin:2013teo . ## Iv Stellar Hydrostatic Equilibrium An asymmetric dark star resembles in several aspects a neutron star. Both types of stars produce no energy by fusing nuclei in their cores. Therefore there is no radiation pressure present. The structure of the star is determined by the equilibrium between the Fermi pressure of the constituent elements and gravity. In Newtonian dynamics the above condition takes the simple form dPdr=−GMρr2, (16) where , , and are the pressure, the mass and the density of the star at radius . Additionally, the continuity of mass gives dMdr=4πr2ρ. (17) The above two equations along with the equations of state (11) and (12) form a complete set of differential equations that can be solved numerically providing pressure, density and mass as a function of . As mentioned earlier, in the absence of self-interactions, the equation of state takes a simple polytropic form in both the relativistic and non-relativistic limits. This polytropic equation of state together with Eqs. (16) and (17) reduce to the well known Lane-Emden equation with index at the non-relativistic and relativistic limit respectively. However, as it has been seen in the case of neutron stars, the Newtonian approximation is not accurate enough and general relativity must be taken into account. We implement this by solving the Tolman-Oppenheimer-Volkoff equation together with Eqs. (17), (11) and (12). For completeness, let us briefly review how the Tolman-Oppenheimer-Volkoff equation is obtained. We are seeking a solution of the Einstein field equation , in the presence of matter with an energy momentum tensor of an ideal liquid of the form Tνμ=diag[ρ,−P,−P,−P]. (18) If one plugs the following spherically symmetric metric into Einstein’s field equations ds2=eν(r)dt2−eλ(r)dr2−r2dΩ2, (19) one finds the following set of equations that must be satisfied 8πGP =e−λ(ν′r+1r2)−1r2, (20) 8πGρ =e−λ(λ′r−1r2)+1r2, (21) dPdr =−(P+ρ)ν′2. (22) Requiring that the metric reduces to the empty space Schwarzschild solution at the boundary of the star, provides a solution for and and Eq. (22) takes the final form dPdr=−GMρr2[1+Pρ][1+4πr3PM][1−2GMr]. (23) This is the relativistic version of Eq. (16). We have solved the coupled system of Eqs. (23), (17), (11) and (12) and we have obtained the structure profile of asymmetric dark stars both for attractive and repulsive self-interactions. Relativistic effects can be quite significant and therefore it is compulsory to use the Tolman-Oppenheimer-Volkoff equation instead of the simpler Newtonian version. ### iv.1 Hydrostatic Stability Solving the Tolman-Oppenheimer-Volkoff equation yields an equilibrium solution, which may be stable or unstable. We will briefly review the conditions that must be satisfied in order for a star to pass from stability to instability. We assume a constant chemical composition and constant entropy per DM particle. The total mass-energy of the star is the integrated energy density M(R)=∫R04πr2ρdr, (24) where is defined by . The number of constituent DM particles in the star is Weinberg:1972 N(R)=∫R04πr2[1−2GM(r)r]−1/2ndr, (25) where is the DM number density as a function of . A star of constant chemical composition and entropy per particle can only pass from stability to instability with respect to some particular radial normal mode, at a value of the central density for which we have Weinberg:1972 ∂M∂ρc=0,∂N∂ρc=0. (26) These conditions are satisfied if we alternatively choose to satisfy simultaneously the first equation above and ∂∂ρc(MN)=0. (27) Fig. 2 shows (divided by ) and (in units of the corresponding upper mass ) as a function of in the repulsive case with GeV, MeV and . One can see that the points where minimizes and maximizes coincide. This is a generic feature for all the star profiles we present in Figs.  3 and 4, i.e. the transition from stability to instability (collapse) takes place at the point where the mass maximizes. In the absence of Yukawa interactions, the critical relativity parameter at transition from stability to instability is Shapiro:1983du , which is independent of . Yukawa interactions induce dependence of on , and . ## V Analytic Newtonian Approximation We would like to estimate the upper mass limit (Chandrasekhar mass) for asymmetric dark stars using first simple Newtonian arguments. The reader interested in the full relativistic results may skip this section and move to the next one. In order to understand the different regimes the dark star passes through before a collapse occurs we derive analytical solutions using a Newtonian approximation for gravity and assume a constant density. Comparing this approach to the full relativistic treatment of the Tolman-Oppenheimer-Volkoff equation, will provide an idea of how important general relativity effects are in different DM scenarios. In this simplified picture we are going to minimize the energy of the system upon making some approximations. We use Newton’s gravitational law (instead of general relativity) and we assume a uniform density of DM fermions with Yukawa interactions. The energy has three contributions, i.e. kinetic energy, gravitational potential energy and Yukawa potential energy E=EG+Ekin+EY. (28) The self gravity contribution to the energy is EG=−∫R0Gr43πρr3⋅4πρr2dr=−35Gm2χN2R, (29) where is the total number of particles and the radius of the star. The kinetic energy is found by multiplying Eq. (4) by the volume Ekin=2πgs3R3m4ξ(pFmχ), (30) where is the Fermi momentum of the particles. In the relativistic and non-relativistic limits, the kinetic energy assumes simple polynomial forms in : (31) In our approximation we define a radius such that , Rrel=(9π2gs)1/3N1/3mχ. (32) When the system is relativistic, while in the opposite limit () the system can be treated as non-relativistic. The Yukawa energy for the entire system of particles is more complicated to derive than the other terms in the energy. Starting from the potential between two particles in Eq. (8), one can find the potential by integrating the contributions of the shells of a homogeneous sphere of charges  Yukawaderivation . The final interaction energy reads EY=±34αN2μ5R6[2μ3R3−3μ2R2+3−3(1+μR)2e−2μR]. (33) This Yukawa potential energy also assumes simple polynomial forms in in the long and short range limits where and respectively (34) It is no surprise that the energy in the long range regime is independent of the mediator mass and proportional to since it resembles the Coulomb potential. In the short range limit where , the exponential suppression of the potential is counterbalanced by the number of close neighbours, thus leading to a overall dependence. Given the above, we identify four distinct regimes that can be realised by the system: 1. Non-relativistic, short range: and : E(N,R)=− 35Gm2χN2R+mχN+ gs15π(9π2gs)5/3N5/3mχR2±32αN2μ2R3. (35) 2. Non-relativistic, long range: and : E(N,R)=− 35Gm2χN2R+mχN+ gs15π(9π2gs)5/3N5/3mχR2±35αN2R. (36) 3. Relativistic, short range: and : E(N,R)=−35Gm2χN2R+gs6π(9π2gs)4/3N4/3R±32αN2μ2R3. (37) 4. Relativistic, long range: and : E(N,R)=−35Gm2χN2R+gs6π(9π2gs)4/3N4/3R±35αN2R. (38) We would like to estimate the condition for gravitational collapse of the objects, thus determining an upper mass limit for the asymmetric fermionic dark stars, similar to the Chandrasekhar limit in the case of white dwarfs. We assume that the star starts always from regime 1. However, depending on the parameters, the star might pass from different regimes before the collapse takes place. In practice we found two possibilities that take place most of the time. In the case of attractive interactions the star can move from regime 1 to 3 and collapse or from regime 1 to 3 to 4 and then collapse. For the repulsive potential the system goes directly from regime 1 to collapse. In the next subsections we will analyse each regime in greater detail. ### Regime 1: Non-relativistic, short range We rewrite Eq. (35) in terms of new constants keeping the explicit dependence of and . E(N,R)=−AN2R+BN5/3R2+CN2R3. The derivative gives two extrema R±=BAN−1/3±√(BA)2N−2/3+3CA. (39) For an attractive potential () is a minimum and a maximum. If one increases the value of , there is a particular value where and the system collapses because there is no stable solution. This value of is Nmax=(−B23CA)3/2=9π240g2s√3101α3/2(μmχ)3(MPmχ)3. (40) When , one can find that for the whole parameter space we examine here. Therefore when the system has accumulated particles, Yukawa forces are still short range. Since the energy has no minimum, keeps dropping until or the Schwarzschild radius is reached depending on which one is the larger. If , one can estimate . This means that for the parameter space considered here the system reaches before the Yukawa force becomes long range. Therefore the system will go to regime 3 or 4 and not 2. If , the system collapses to a black hole before the particles become relativistic. In this case for the parameter space we consider, and therefore the system forms a black hole without reaching the long range limit of the Yukawa force. In the case where the Yukawa potential is repulsive (), (so it is unphysical) and is a global minimum. The upper mass limit for this star can be found by setting . There is always an appropriate that satisfies this because decreases with while increases. For the parameter space we consider here, we found that the Schwarzschild radius is encountered while still in the non-relativistic regime. The smallest possible value of (which occurs when ) is . One can see that for the parameter space considered here, this line is crossed first by and then by . It can be easily seen that this is also true even for a finite , thus the star collapses to a black hole before the constituents become relativistic. In addition, one can show that the aforementioned asymptotic value satisfies and therefore the interactions are not long range. ### Regime 2: Non-relativistic, long range We rewrite Eq. (36) with new constants as E(N,R)=−AN2R+BN5/3R2+CN2R. The relevant minimum is found at Rmin=2BA−CN−1/3=(9π2gs)2/31mχN1/3⎡⎣(mχMP)2∓α⎤⎦−1. (41) If the Yukawa potential is attractive (), there is always a stable minimum. On the other hand in the case of repulsive potential , the strength of the coupling is bound by , if a stable minimum is to exist at finite positive . In practice for the considered parameter space, we found that no dark star passes through this regime. ### Regime 3: Relativistic, short range Once again we rewrite Eq. (37) in terms of new constants as E(N,R)=−AN2R+BN4/3R+CN2R3 (42) and find the extremum at Rext=√3CA−BN−2/3. (43) In the case of an attractive potential (), this extremum is a maximum as long as is smaller than Next=(BA)3/2=158√5π2gs(MPmχ)3. (44) For the energy is monotonically increasing as a function of , and therefore the star collapses. For the energy increases between 0 and and decreases from to . One can compare from Eq. (40) and NmaxNext=3√3π3/2125α3/21g3/2s(μmχ)3. (45) If the star passes from regime 1 to regime 3, it collapses if . If on the other hand , the collapse proceeds as long as . In the case of repulsive Yukawa potential (), the potential is monotonically decreasing for . This means that the star remains in regime 1. Once there is a stable minimum at . However, as we pointed out in the discussion in ”Regime 1”, the Schwarzschild radius is met before this. ### Regime 4: Relativistic, long range In this case every term scales as . Rewriting Eq. (38) in terms of new constants we get E(N,R)=−AN2R+BN4/3R+CN2R. (46) The critical number of particles is Ncrit=(BA−C)3/2=158√5π2gs⎡⎣(mχMP)2∓α⎤⎦−3/2. (47) For the star collapses. In the case of repulsive interactions with , no collapse can take place for any value of . As we mentioned in “Regime 1”, in case of repulsive interactions we never enter this regime because the particle never becomes relativistic. ## Vi Results In this section we present the full relativistic results after solving numerically the system of Eqs. (23), (17), (11) and (12). The algorithm we use to solve the relativistic hydrostatic equilibrium is the following: 1. Set initial conditions and , with being the relativity measure in the center of the star defined below Eq. (7). 2. Integrate one step of Eq. (17) to get using the equation of state for . Then integrate one step of Eq. (23) to obtain . For this value of , one can obtain the corresponding from the equation of state (11). 3. Repeat times the above step to obtain and . 4. Identify the where . This defines the radius of the star. Correspondingly the mass of the star is . For each set of DM parameters we find the Chandrasekhar mass by scanning over and identifying the largest total mass. We will present now the mass vs radius relations of the stable dark star configurations. In the left panels of Fig. 3 we show mass vs radius relations for the three generic cases: repulsive interactions (upper panel), no DM self-interactions (middle panel) and attractive interactions (bottom panel). We have chosen a coupling , mediator mass MeV and three different cases of DM mass of 10, 100 and 1 TeV. For each profile, we mark the upper stable mass (Chandrasekhar mass) by a circle. Note that star configurations with radii larger than the one that corresponds to the Chandrasekhar mass are stable, while configurations with smaller radii are unstable. We refer the reader to the discussion of stability in subsection IV.1 and to the example of Fig. 2 which shows that the Chandrasekhar mass is indeed the last stable configuration as density increases. Note also that the dashed lines represent dark star configurations where the non-relativistic version of the Tolman-Oppenheimer-Volkoff equation is used. As it can be seen, general relativity effects can be quite significant. For each mass vs radius curve that we show in the left panels of Fig. 3, apart from the Chandrasekhar mass, we mark in addition another point by a diamond. In the respective right panels, we show the density profile of the dark star that corresponds to the diamond. In the scenarios of repulsive and no self-interactions, the equilibrium solutions feature a spiral structure in the unstable region of the curves. This feature is absent in the attractive scenario since the pressure becomes negative for some , and this is encountered before the spiral appears. One can notice (as it is expected) that heavier DM particles form lighter and more compact dark stars. In Fig. 4 we show again mass vs radius relations and corresponding density profiles for the diamond points (as in Fig. 3) for three values of and , for fixed GeV and MeV for repulsive and attractive interactions. Note from Eqs. (11) and (12) that the Yukawa contribution to the density and the pressure of the star is unchanged under the scaling and . Therefore it is sufficient to fix either and and scan the phase space of the other one. One can notice that increasing the coupling of repulsive Yukawa interactions leads to larger stars, although DM particles in these stars are non-relativistic (). Despite that, the general relativity effects are large. ### vi.2 Chandrasekhar Mass In Fig. 5 we show the maximum mass of dark stars (Chandrasekhar limit) for the DM self-interaction parameter space shown in Fig. 1. A first expected observation is that repulsive self-interactions lead to heavier dark stars compared to attractive ones, since repulsive interactions add to Fermi pressure and therefore more massive configuration can be supported. In fact larger couplings of (for repulsive interactions) correspond to heavier stars. This can be seen also in Fig. 4. On the contrary, larger Yukawa couplings for attractive interactions lead to smaller Chandrasekhar limits. One can see that the difference in the mass of dark stars made of DM particles with attractive and repulsive interactions is large. The mass of the attractive dark stars for the whole parameter space lies below the limits that can be imposed by gravitational lensing. For repulsive stars, the mass can be a significant fraction of a solar mass, making these stars more visible from the lensing point of view. ## Vii Discussion & Conclusions In this paper we studied the possibility that fermionic asymmetric DM can form stars, with self-interactions that can solve well standing problems of the CCDM paradigm. We solved the Tolman-Oppenheimer-Volkoff equation and we derived mass vs radius relations for these stars for both attractive and repulsive DM self-interactions for a range of DM and mediator masses. We also derived upper mass limits for these objects and we studied the hydrostatic stability of the star profiles we examined. One issue of fundamental importance that we plan to address in future work is related to the question of how these stars form in the first place. There are several possibilities where such a scenario can be visualized. One possibility is the creation of high DM density regions due to adiabatic contraction, caused by baryons Blumenthal:1985qy (see also discussion in Gustafsson:2006gr ). Another possibility is the existence of a subdominant strongly interacting DM component that consists the seed for the gravitational collapse of parts of the DM halo via a gravothermal mechanism gravo . In this case DM self-interactions can lead to a collapse due to the fact that DM-DM collisions can send one of the two particles to a deeper gravitational potential while the second one leaves the area subtracting thus energy from the system that continues until an instability is established. As it has been discussed in Pollack:2014rja , the gravothermal mechanism leads to either gravitational collapse if the DM-DM forces are short range, or to formation of binary systems in case of long range interactions. In our case, the DM-DM interactions are short range because they are mediated by the massive particle , (or ) and therefore the instability leads to formation of DM asymmetric stars. Another possibility is the capture of a significant amount of DM particles by a supermassive star. After the collapse of the star and the supernova explosion, and since DM particles cannot be blown away in significant amounts by the supernova, a pure DM star or a mixed star with significant amount of baryons can be formed Kouvaris:2010vv . We should emphasize here that all these possibilities do not lead to a total collapse of the whole DM population but rather of a small fraction of it. In view of this, one should not worry for star profiles we present here in the case of repulsive interactions that are within the range of constrained by the MACHO Alcock:2000ph and EROS microlensing observations Tisserand:2006zx , simply because these stars do not comprise the whole DM density. Another issue we would like to leave out for future work is the discovery signatures for this type of dark stars. Apart from gravitational lensing that can in principle discover objects like this based on the spacetime distortion that their presence can cause, other types of direct signals can exist. If DM communicates with the Standard Model via some portal, e.g. kinetic mixing between the photon and a dark photon, faint photon luminosity should be expected from these stars. Additionally, as we pointed out, for a large range of our parameters, the radius of these stars can be significantly smaller than that of a regular neutron star. This means that asymmetric dark stars can rotate faster than regular neutron stars. Pure dark stars made of DM particles that interact with the Standard Model particles through some portal, or mixed stars can appear as fast rotating pulsars. Rotational frequencies below millisecond are hard to be explained by a regular neutron star. An example of such a case is XTE J1739-285, which allegedly rotates with a frequency of 1122Hz Kaaret:2006gr . Any odd looking neutron star is a potential candidate for an asymmetric dark star. We should mention that for a dark star that has collapsed to a black hole with a mass below , evaporation of the black hole via Hawking radiation takes place within the age of the universe. Although additional constraints exist Capela:2012jz ; Pani:2014rca ; Capela:2014ita , there is still a possibility for observing the spectrum of an evaporating black hole in the sky. Finally we should mention (as one can easily estimate) that in order for a dark star to collide with the earth during earth’s lifetime of years, the dark star mass must be where is the local DM density, is the radius of the earth, and the velocity dispersion of DM. One can see from Fig. 5 that TeV DM with attractive interactions and a coupling of can in fact give dark stars below that mass, making it possible for such a dark star to have collided with the earth in the past. This work is supported by the Danish National Research Foundation, Grant No. DNRF90.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9701776504516602, "perplexity": 721.4776288461744}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107891228.40/warc/CC-MAIN-20201026115814-20201026145814-00398.warc.gz"}
http://math.stackexchange.com/users/53983/user39097
# user39097 less info reputation 7 bio website location age member for 2 years, 1 month seen Aug 7 '13 at 11:45 profile views 22 # 6 Questions 3 Matrix representation 2 Random Matrix Question 2 Calculating $\int_{\mathcal{S}}x_1^r \, \mathrm dx_1\ldots \, \mathrm dx_n$ 2 Constrained maximization problem 2 Calculating of multi-dimensional integral 2 # 139 Reputation This user has no recent positive reputation changes 1 CDF of a sum of independent random variables 0 Matrix column independence? # 12 Tags 1 probability-theory × 3 0 matrices × 2 0 multivariable-calculus × 3 0 definite-integrals 0 integration × 2 0 stochastic-processes 0 probability × 2 0 probability-distributions 0 linear-algebra × 2 0 convex-optimization # 1 Account Mathematics 139 rep 7
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5469359755516052, "perplexity": 7870.092153906458}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115865010.11/warc/CC-MAIN-20150124161105-00236-ip-10-180-212-252.ec2.internal.warc.gz"}
https://proofwiki.org/wiki/Definition:Standard_Number_Field
# Definition:Standard Number Field ## Definition The standard number fields are the following sets of numbers: The rational numbers: $\Q = \set {p / q: p, q \in \Z, q \ne 0}$ The real numbers: $\R = \set {x: x = \sequence {s_n} }$ where $\sequence {s_n}$ is a Cauchy sequence in $\Q$ The complex numbers: $\C = \set {a + i b: a, b \in \R, i^2 = -1}$. These sets are indeed fields: $\struct {\Q, +, \times, \le}$ is an ordered field, and also a metric space. $\struct {\R, +, \times, \le}$ is an ordered field, and also a complete metric space. $\struct {\C, +, \times}$ is a field, but cannot be ordered compatibly with $+$ and $\times$. However, it can be treated as a metric space. ## Also see Neither the set $\N$ of natural numbers nor the set $\Z$ of integers are fields. However: $\struct {\N, +, \le}$ can be defined as a naturally ordered semigroup. $\struct {\Z, +, \times, \le}$ is an ordered integral domain.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.994037389755249, "perplexity": 332.6204047947535}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107879537.28/warc/CC-MAIN-20201022111909-20201022141909-00201.warc.gz"}
https://www.mathdoubts.com/angle-formation-by-rotation/
# Formation of angle by Rotation The amount of rotation of a line from its initial position to final position is called an angle formed by rotation. ## Example A straight line can make an angle geometrically by moving from its initial position to final position. $\overrightarrow{RS}$ is a ray and it is rotated to new position, where it is called as $\overrightarrow{RT}$. The amount of rotation of line $\overrightarrow{RS}$ to become $\overrightarrow{RT}$ is called an angle. An angle is formed between lines $\overrightarrow{RS}$ and $\overrightarrow{RT}$ by the rotation and it is denoted by $\angle SRT$ or $\angle TRS$ in mathematics. Latest Math Topics May 27, 2020 May 25, 2020 May 11, 2020 May 02, 2020 Email subscription Math Doubts is a best place to learn mathematics and from basics to advanced scientific level for students, teachers and researchers. Know more Follow us on Social Media ###### Math Problems Learn how to solve easy to difficult mathematics problems of all topics in various methods with step by step process and also maths questions for practising.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.16011300683021545, "perplexity": 957.9071950163287}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347396300.22/warc/CC-MAIN-20200527235451-20200528025451-00159.warc.gz"}
https://www.physicsforums.com/threads/complex-integral-question.416256/
# Complex integral question 1. Jul 15, 2010 ### nhrock3 $$\int_{|z|=3}^{nothing}\frac{dz}{z^3(z^{10}-2)}\\$$ $$f=\frac{1}{z^3(z^{10}-2)}\\$$ $$f(\frac{1}{z})=\frac{1}{(\frac{1}{z})^3((\frac{1}{z})^{10}-2)}\frac{z^{13}}{1-2z^{10}}=\\$$ $$res(f,\infty)= res(\frac{1}{z^2}f(\frac{1}{z}),0)=\frac{1}{z^2}\sum_{n=0}^{\infty}(2z^{10})^n\\$$ $$res(f,\infty)= res(\frac{1}{z^2}f(\frac{1}{z}),0)=-[res(f,inside|z|=3)+res(f,outside|z|=3)]$$ from the sum i get that there is no $$z^{-1}$$ member in the series so the coefficient of $$z^{-1}$$ is zero so the residiu of infinity is zero but still all of my singular points are |z|=3 so the integral equals zero ?? did i solved it correctly did i written every formula regarding the laws of residue correctly here ? Can you offer guidance or do you also need help? Draft saved Draft deleted Similar Discussions: Complex integral question
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9905366897583008, "perplexity": 3571.1480737813486}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948597585.99/warc/CC-MAIN-20171217210620-20171217232620-00065.warc.gz"}
http://codeforces.com/blog/entry/55858
### Marco_L_T's blog By Marco_L_T, 5 years ago, Hello, Codeforces! I'm quite excited to invite you to participate in Codeforces Round #447 (Div.2 Only) which will be held on November 19 16:55 MSK. All five problems are created by Zerui Cheng (Marco_L_T), Bingheng Jiang (NOIRP), Yiming Feng (whfym). And it's our first round on Codeforces. We want to show our great thanks to our school The High School Affiliated to Anhui Normal University and our coach Guoping Ye in competitive programming training. And we also want to show our great appreciation to Mikhail Krivonosov (mike_live), Gleb Lobanov (Glebodin), Weihao Zhu (Tommyr7), Shiqing Lyu (cyand1317) for testing the problems, to Nikolay Kalinin (KAN) for coordination and to Mike Mirzayanov (MikeMirzayanov) for the fantastic Codeforces and Polygon platforms. The round can't be realized without their great help. The contest will consist of 5 problems and you'll be given 2 hours to solve them. As usual, the scoring will be announced shortly before the start of the contest. The contest is rated for Div. 2 contestants. And the same as before, Div. 1 contestants can take part out of competition. Wish everyone high rating and bugless code! Clarification: In the mail, it reads that the duration is 2 hours and 30 minutes and it'll contain 6 problems. There's a mistake. The duration of the contest is 2 hours and there will be 5 problems. UPD1: Scoring: 500-1000-1500-2000-2500 UPD2: The contest is finished! Have fun hacking! UPD3: The system test is finished! Congrats to the winners! And do you find something interesting in the statements,especailly for Chinese contestants? Div.1 & Div.2: 1. fateice_ak_ioi (solved all the problems and got 22 hacks) 2. peace (solved all the problems and got 10 hacks) 3. dreamoon_love_AA 4. KrK (solved all the problems) 5. Benq (solved all the problems) Div.2: 1. fateice_ak_ioi (solved all the problems and got 22 hacks) 2. peace (solved all the problems and got 10 hacks) 4. daaaaaaaaaaaaaaaaaaa 5. QYitong1 UPD4: Maybe you're complaining that there're too many hacks and the pretests are so weak,but it's our intention to do so. We regard hacks as a very important part and a feature of Codeforces.Do you agree? UPD5: Editorial Hope you have fun in solving the problems and hacking! See you next time! • +280 » 5 years ago, # |   +29 Auto comment: topic has been updated by Marco_L_T (previous revision, new revision, compare). • » » 5 years ago, # ^ | ← Rev. 4 →   +39 Contest number 447 This number is considered as a lucky number on codeforces. I wish this is the start of being Lucky. Good luck to all • » » » 5 years ago, # ^ |   0 Maybe not always: http://codeforces.com/problemset/problem/630/C » 5 years ago, # |   +38 Hope you have fun with the contest! See you on the leaderboard! • » » 5 years ago, # ^ |   +20 If this contest will be rated It will be very good two contests int three days Good luck » 5 years ago, # |   +75 wish all the participants high rating! :) » 5 years ago, # |   -36 Is is rated? • » » 5 years ago, # ^ |   +47 it has been clarified in the announcement :) • » » 5 years ago, # ^ |   -10 Is it* • » » 5 years ago, # ^ |   0 I like how you are div1 and think that is unrated • » » 5 years ago, # ^ |   0 It is rated for div2, but then technical issues happens and it will be unrated • » » 5 years ago, # ^ |   +94 » 5 years ago, # |   +29 Hope you make progress and show yourselves! » 5 years ago, # |   +5 Wish you have fun! » 5 years ago, # |   +36 我随手AK • » » 5 years ago, # ^ |   +3 your first and only comment that has positive upvotes. hmmm. • » » » 5 years ago, # ^ | ← Rev. 2 →   +3 That's why he gets upvotes. • » » 5 years ago, # ^ |   0 In fact he means he will get full scores VERY EEEEEEEEASY. » 5 years ago, # |   +7 To be continue » 5 years ago, # |   +6 Applause in the front. » 5 years ago, # |   +9 For ratings! » 5 years ago, # |   +12 3 months ago i had 100 more rating than now. This contest will get everything back to normal and even higher!!! • » » 5 years ago, # ^ |   0 Same thing here.. • » » 5 years ago, # ^ |   -12 It is r**** ? • » » » 5 years ago, # ^ |   +11 Dude you posted under my comment not in the right spot » 5 years ago, # |   0 2 rated contest's in two days. Amazing! » 5 years ago, # |   0 KAN, why the registration is delayed? » 5 years ago, # | ← Rev. 2 →   +16 Previous contest problem description was realy sort but briefly. I hope this contest will be same :) » 5 years ago, # |   -34 I think this is a mathematical contest because the authors are Chinese. I do not like it very much • » » 5 years ago, # ^ |   +16 Don't be so sure.Maybe you're wrong. • » » » 5 years ago, # ^ |   +12 What did I say? but the contest is interesting thanks for a good contest!!! • » » 5 years ago, # ^ |   -17 I like mathematical contest. • » » » 5 years ago, # ^ |   -7 Why downvote me? I just like mathematical contest, so anyone downvoting me hate it? » 5 years ago, # | ← Rev. 2 →   -24 Deleted • » » 5 years ago, # ^ |   +3 :) It has been clarified in the announcement. » 5 years ago, # |   +3 it seems an nice contest excited :) » 5 years ago, # |   -28 Prepare for non-alogrithmic idiotic math contest • » » 5 years ago, # ^ |   +19 Don't be that sure, please. :) • » » » 5 years ago, # ^ |   0 i know what im saying » 5 years ago, # |   -7 The announcement is not very short as previous contests.But we can see sincere thanksgiving...))Wish every body luck and high rating !And wish good luck codeforces servers! » 5 years ago, # |   0 Two contests in three days (if not counting contest from CS Academy lately). Well, the excitement inside me is rising high :D • » » 5 years ago, # ^ |   0 In between there was a topcoder srm and today we also have codechef's cookoff :P • » » » 5 years ago, # ^ |   0 Seems like the stack of upcoming contests has been overflowed :P » 5 years ago, # |   -14 RAATATTATATEDEDDDD??!?!?!?!?!??!?!?!?!? » 5 years ago, # |   +2 Thanks to all the people --> 2 contest in 3 days awesome feeling :) Hope for large number of hacks and good rating Wishing all the best to everyone » 5 years ago, # |   +1 Another Chinese round, so exciting! Hope for short statements and high rating. • » » 5 years ago, # ^ |   +29 You are out of competition, aren't you? How you get a "high rating"? :) » 5 years ago, # |   0 is it unrated? » 5 years ago, # |   0 haha » 5 years ago, # | ← Rev. 3 →   0 And the mail reads that the contest will be of 2 hrs 30 minutes. :| Also, it reads 6 problems instead of 5 as declared in announcement :3Anyway, best of luck everyone. Happy Learning and Coding. • » » 5 years ago, # ^ |   0 Well,maybe there's something wrong with the mail. The duration of the round is 2 hours if everything goes well. • » » » 5 years ago, # ^ |   +3 Yeah, I understand that. Just bringing it to the notice of problem setter/contest setter. This issue had happened once in past too :| » 5 years ago, # |   +1 all good luck, and let your rating go up » 5 years ago, # |   +16 What will you do at CF#447? Are you free? Can you come to AK? » 5 years ago, # |   +3 What a pity! I've just become candidate master. Now I can't get any rating!!!! QAQ » 5 years ago, # |   0 Wow Yayyy!!!!! » 5 years ago, # |   0 I want to sleep now :( » 5 years ago, # |   -7 I expect short problem statements. » 5 years ago, # |   +1 Beijing time 21:55 , perfect! » 5 years ago, # |   0 There should be another id to get rated. :( » 5 years ago, # |   +1 Hope short problem statements just like past contest. » 5 years ago, # |   +2 Hope everyone high rating! And hope me not to drop to green. » 5 years ago, # |   +1 Wish everybody High rating))) » 5 years ago, # |   +1 Today will be a perfect ICPC-style Contest practise!2:00 hrs Codeforces + 2:30 hrs Codechef! :/ » 5 years ago, # |   0 Time to beat some meat. • » » 5 years ago, # ^ |   0 Don't do, just 10 more days » 5 years ago, # |   0 Why can't i register? • » » 5 years ago, # ^ |   +3 Don't be late. » 5 years ago, # |   0 That feel when div2 B is harder than div2 C... • » » 5 years ago, # ^ |   +8 Don't be so sure. • » » 5 years ago, # ^ |   +6 I don't think so :D • » » » 5 years ago, # ^ | ← Rev. 2 →   +3 :o I definitely missed something. This will be a good learning experience :PContest is over, are we allowed to discuss? • » » » » 5 years ago, # ^ |   0 Yes • » » » 5 years ago, # ^ | ← Rev. 2 →   +6 snacache what was the key observation for B? • » » » » 5 years ago, # ^ |   +8 I don't know haha. I just brute forced for n·m ≤ 20 and found a pattern » 5 years ago, # |   -17 A=B=C < D = E • » » 5 years ago, # ^ |   +17 I don't think that A=B or A=C. • » » » 5 years ago, # ^ |   +3 A < B C < B E = D ~ C very izi to E and D » 5 years ago, # |   -29 Hacks For C Please :( » 5 years ago, # |   +32 HackForces. • » » 5 years ago, # ^ |   +11 10/10 paint skill by the way. » 5 years ago, # |   +10 I was excited to see characters from Land of Lustrous in the first problem but then you dropped them later on. :/ • » » 5 years ago, # ^ |   0 I'm sorry but I can't decide all the statement :( • » » 5 years ago, # ^ | ← Rev. 2 →   +5 If I have the chance to propose a contest on my own, I would write the statement in the similar way.(But maybe it will be long long time before I have the chance.) » 5 years ago, # |   +33 Hacks everywhere!!! » 5 years ago, # |   0 I was intended to hack some one on Problem C (while I found my code was wrong). But I could never load the "hack" page, watching the cycling circle. Then my submission got HACKED...??? So is it a bug? • » » 5 years ago, # ^ |   0 This often happens here. • » » 5 years ago, # ^ |   +3 I had the same and then I changed my browser (using Microsoft Edge) and I was able to hack 18 times the C problem it was really nice cause I only managed A and C • » » » 5 years ago, # ^ |   +5 Got it. Thank you all for explaining. _(:3 rz) » 5 years ago, # |   +90 The hardest B I have ever seen :D • » » 5 years ago, # ^ | ← Rev. 3 →   +1 What's the hack?? Why is:0 if N+M is odd and K == -1(2^(N — 1) * (M — 1)) otherwisewrong?EDIT: My code is here: https://hastebin.com/uteviyadig.cpp and I do not see any possibilities of overflow?EDIT EDIT: Oh my god, I'm so dumb, I didn't consider the case where N = INF — 1. Thank you Benq. Oops :( • » » » 5 years ago, # ^ |   +8 Dont forget to N %= MOD,M %= MOD,before powering • » » » » 5 years ago, # ^ |   0 Yes, I did that; there's an N %= (INF — 1) and M %= (INF — 1) before the powering. • » » » » » 5 years ago, # ^ |   0 Why do we mod the exponent? Is there a name for this technique? • » » » » » » 5 years ago, # ^ |   0 Fermat's Little Theorem. See here. • » » » » 5 years ago, # ^ |   0 OMG... That's why I got hacked at the last second.. Thank you » 5 years ago, # |   +3 What's the hack for C? • » » 5 years ago, # ^ | ← Rev. 2 →   +3 31 6 9 • » » » 5 years ago, # ^ |   +7 What is the answer? • » » » » 5 years ago, # ^ |   +18 it's possible: 1 6 1 9 1 • » » » » » 5 years ago, # ^ |   +4 But 6 and 9 gcd is 3 and it is not present in the given intial set. Was I solving the wrong problem all the time? • » » » » » » 5 years ago, # ^ |   +2 But 6, 9 need not be adjacent in the actual array. • » » » » » » 5 years ago, # ^ |   +3 Only GCD of a sequences matter. In this case GCD(6, GCD(1, 9)) , which is equal to 1. • » » » » » » » 5 years ago, # ^ |   +1 Okay, indeed I misread the whole contest. I assumed that any two numbers GCD should be in the set. Now the problem is relatively simple. If the smallest number divides all other number, then print a[0], a[1], a[0], a[2], a[0], a[3] and so on. • » » » » » 5 years ago, # ^ |   +11 Interesting, I asked for a clarification in contest and I received this answer back:"If answer exists,we can always find an answer in increasing order."It seems that this answer is not in increasing order... • » » » » » 5 years ago, # ^ |   +3 This is disaster • » » » » 5 years ago, # ^ |   0 9 1 6? • » » » 5 years ago, # ^ |   0 ANS = -1 ? • » » » » 5 years ago, # ^ |   0 Nope. 6 1 9 • » » » » » 5 years ago, # ^ | ← Rev. 2 →   0 You are right » 5 years ago, # |   0 what's C's hack data?It's amazing. • » » 5 years ago, # ^ | ← Rev. 4 →   0 Maybe 3 5 7 35 EDIT : The output should be -1. • » » » 5 years ago, # ^ |   0 thanks a lot. • » » » 5 years ago, # ^ |   0 Could you tell me how to solve it? • » » » » 5 years ago, # ^ | ← Rev. 3 →   +15 Let m be the minimum value of the set. A possible solution is A[0] m A[1] m A[2] m ...... m A[n-1] No solution if m does not divide all values in the set. Code • » » » » » 5 years ago, # ^ |   0 How to check if the answer is -1? • » » » » » » 5 years ago, # ^ |   +10 No solution if m does not divide all values in the set. No solution means ans=-1. • » » » » » » » 5 years ago, # ^ |   0 Got it, thank you very much. • » » » » » 5 years ago, # ^ |   0 how could prove that we just need to check that m doesn't divide all values in the set? QAQ • » » » » » » 5 years ago, # ^ |   0 Because if m does divide all of them, the construction I showed above is always valid as any subarray of size >= 2 will have GCD exactly m, and all the elements of the original set are present as subarrays of size 1. • » » » » » » » 5 years ago, # ^ |   +5 Got it! But why there won't have another method to construct a solution while m doesn't divide all values in the set • » » » » » » » » 5 years ago, # ^ |   +8 GCD of the whole array will be the minimum value in the set. So, that value should divide GCD values of all the subarrays, i.e. all members of the set. • » » » » » » » » » 5 years ago, # ^ |   0 Got it. thanks a lot. • » » » » » » » » » 5 years ago, # ^ | ← Rev. 2 →   0 Sir can you please tell me what your code did for the input -21 6 • » » » » » » » » 5 years ago, # ^ |   0 Prabably because minimum of S must be GCD of the whole sequence. If if doesn't divide all other subsequences, then given S is invalid. • » » » » 5 years ago, # ^ |   0 I got hacked in last 15 minutes :( • » » » » 5 years ago, # ^ | ← Rev. 2 →   0 One greedy solution also can be to calculate the GCD of the set and then if the GCD is itself present in the set then answer will be 2n numbers where at odd positions will be the GCD and at even positions will be the numbers of the set. Otherwise -1. • » » » 5 years ago, # ^ | ← Rev. 2 →   +6 1 -> 1 gcd(7) = 71 -> 2 gcd(7,35) = 72 -> 2 gcd(35) = 35That's Wrong answer :/ • » » » 5 years ago, # ^ |   0 When will you get the 5 then? gcd(7,35) = 7 • » » » » 5 years ago, # ^ |   0 Oops, my bad ._. • » » 5 years ago, # ^ |   +1 4 2 8 12 16 • » » » 5 years ago, # ^ |   0 I really appreciate it • » » 5 years ago, # ^ |   0 I used 3 18 27.possible ans is 27 3 18. but most peoples code gave -1. » 5 years ago, # |   0 Was E computing scc and calculating for each scc and taking the max of it??? • » » 5 years ago, # ^ |   0 That's what I tried but I got WA on test 3. I also traversed from the SCC of the starting tree until the end of path and took maximum over all using DP. » 5 years ago, # |   +3 10 hacks per second ! » 5 years ago, # |   0 When u hack 14 but u know that ur own code is incorrect in 2 probs B) . » 5 years ago, # |   +32 Problem difficulty: A < C < E < D < B. • » » 5 years ago, # ^ |   0 D was definitely hard, but, come on, D < B? • » » » 5 years ago, # ^ | ← Rev. 3 →   +11 Believe or not, my mind completely draw a blank in B. In the last 30 minutes of the contest, I tried some idea (like flipping sign of four cells that are corners of rectangle won't change the result), but none of them works. By the way, how to solve it? • » » » » 5 years ago, # ^ | ← Rev. 3 →   +24 Notice that after you choose the numbers in the upper left N - 1 by M - 1 grid, the rest of the numbers are determined. So the answer is 2N - 1·M - 1. You also need to check the case where N + M is odd and K =  =  - 1. (Answer is 0 in such case).EDIT @below: Yeah, you're right, oops. • » » » » » 5 years ago, # ^ |   +5 Do you mean 2(M - 1)(N - 1)? • » » » » 5 years ago, # ^ |   +28 It's a classical combinatorics problem. In case k = 1, you can arbitrarily fill the top-left (n - 1) × (m - 1) matrix — other fields will be uniquely determined and correct. Thus in that case the solution is 2(n - 1)(m - 1). The same logic applies when k =  - 1 and n ≡ 2m. Otherwise you can show that there is no solution. • » » » » » 5 years ago, # ^ |   +4 Problem B. Let's consider the k = 1 case. I don't understand the fact why solution is uniquely determined by fixing the top-left matrix arbitrarily. I get it that all the elements (i,j) where j = m and i = 1 to n-1 and i = n and j = 1 to m-1 are uniquely determined by parity of negative numbers in that row/column. The now filled numbers of the last row and the last column determine the element (n,m). What if they contradict each other ?! Then the sign of (n,m)th element is not uniquely determined right ? • » » » » » » 5 years ago, # ^ |   +13 They won't contradict because the product of all the numbers in the last row (except the last element) is equal to the product of all the numbers in the last column (except the last element), because they are both equal to the product of all numbers in the (n - 1) × (m - 1) submatrix. • » » » » » » » 5 years ago, # ^ |   0 Thank you so much! • » » » » » » 5 years ago, # ^ |   0 After chosing one columm after another, numbers of product of cells in the same row and in each columms before is always even. So product of the last columm is even. • » » » » 5 years ago, # ^ |   +7 Write brute force, build a table with answers, see the pattern :) • » » 5 years ago, # ^ |   0 How to solve E? • » » » 5 years ago, # ^ | ← Rev. 2 →   +6 Compress the strongly connected components. In each SCC, you can repeatedly traverse cycles and reduce each edge to 0. After compressing SCCs, do DP on the DAG to get max path length from source to any leaf. Code • » » » » 5 years ago, # ^ |   0 This is exactly what I did but I got WA on pretest 3? Could you tell me what I is wrong? • » » 5 years ago, # ^ |   0 For B, I did a simulation to get the pattern. Was able to make 2 hacks in my room. let's see if my solution passes the system test. • » » » 5 years ago, # ^ |   0 can you tell how do the output for third case is 16 ? • » » » » 5 years ago, # ^ |   +5 Here are the 16 possibilities. 1 -1 1 -1 -1 -1 1 -1 1 -1 1 1 1 1 -1 1 -1 1 -1 -1 -1 1 -1 1 1 -1 1 1 1 -1 -1 1 1 1 -1 1 -1 1 1 -1 -1 -1 -1 1 1 1 -1 1 1 1 -1 -1 1 1 1 1 -1 1 -1 1 -1 1 1 -1 -1 -1 -1 1 1 -1 1 1 -1 -1 -1 -1 -1 -1 -1 -1 -1 1 1 -1 1 1 -1 -1 -1 -1 1 -1 1 1 -1 1 -1 -1 -1 -1 1 1 -1 1 1 -1 -1 -1 1 1 -1 -1 -1 -1 1 1 -1 -1 -1 -1 1 1 -1 1 1 -1 -1 1 1 1 -1 1 1 1 -1 1 -1 1 -1 1 1 1 1 -1 • » » » » » 5 years ago, # ^ | ← Rev. 2 →   0 Why not-1 -1 -11 1 11 1 1 • » » » » » » 5 years ago, # ^ |   0 -1 -1 -1 => Product = -1 1 1 1 => Product = 1 1 1 1 => Product = 1 So, this is not a valid solution. • » » » » » » » 5 years ago, # ^ |   +6 I am an idiot • » » 5 years ago, # ^ |   0 I cannot agree more! Even for me solutions of D and E were much more obvious than B. That's not something happens any often. I believe if B/C were not such killers, D and E would be solved by many more people though they were tedious.Then again I did not expect this solution of D with complexity O((n + m)log2n) would pass. Isn't it almost 4 × 108? 2.5s seemed impossible. » 5 years ago, # |   0 How do the output for third case is 16 ? PROBLEM: B • » » 5 years ago, # ^ |   0 Also wondering • » » 5 years ago, # ^ |   0 Constructive algorithm:total 2*2 arrays = 16,the parity bits can be inserted accordingly. » 5 years ago, # | ← Rev. 2 →   +10 Authors thanks a lot for this Hack party! Very nice problems! » 5 years ago, # | ← Rev. 2 →   0 I had I idea to code Tarjan for E (still not sure will it work) but then saw so many wrong Cs in room. » 5 years ago, # |   +11 Wow, that was the hardest B I have seen in a while. I had to bruteforce to find the formula. Moreover, I guess a lot of people are either hacked or FST because of taking modulo mod p instead of (p-1). • » » 5 years ago, # ^ |   0 Hi, I used mod p and got accepted (submission here)I see you used p-1, Why is mod (p-1) the correct answer?Thanks • » » » 5 years ago, # ^ | ← Rev. 2 →   +5 Well I think he means exponents... It's Fermat's little theorem. when p is a prime number » 5 years ago, # |   +24 Authors should care more about the pretests, because there were more hacks than Accepted codes on task B! -_- • » » 5 years ago, # ^ |   +8 It's what we intended to do,we intend for a lot of hacks on B and C. • » » » 5 years ago, # ^ |   +14 Why did I get this incorrect clarification? Is this how you intend for hacks?https://imgur.com/a/7kuB4 • » » » » 5 years ago, # ^ |   +7 Really sorry for this.I don't know who answered your question,maybe he is so busy answering questions that he made a mistake. » 5 years ago, # |   +4 Not a Div. 2 contest. B, C were too difficult. » 5 years ago, # |   0 So, how to solve B? I can't think about it anymore :( • » » 5 years ago, # ^ | ← Rev. 2 →   +9 Backtrack on small N,M you'll find that the answer is 2N - 1 * M - 1 • » » » 5 years ago, # ^ |   +4 Any intuition behind this? • » » » » 5 years ago, # ^ |   0 I can try. At each cell you can either put a 1 or -1 . Hence for one row there are 2^m possibilites. now suppose all cells in a row are 1. for the product to be 1 you need even count of -1s to be present in each row. so you have mChoose0 + mChoose2 + mChoose4..... which is 2^m/2 = 2^(m — 1). the count for -1 product is exactly the same.Now you do the same for columns and multiply the exponents to get the overall count. I hope this helps a little bit atleast. • » » » » 5 years ago, # ^ |   0 • » » » 5 years ago, # ^ |   0 how come some of the test cases didn't pass with that approach? • » » » » 5 years ago, # ^ |   +3 The extra requirement is that if k == -1 and n%2 != m%2, the answer is 0. » 5 years ago, # |   0 What's wrong with my approach to C? For each starting position i we find the gcd until all positions j (i < j < m). Then, for each gcd, we check if it's in the given array. If it's not, output -1; otherwise, output the given array.Thanks in advance. • » » 5 years ago, # ^ | ← Rev. 2 →   0 41 5 6 8ans:6 5 8 • » » » 5 years ago, # ^ |   0 I see now. The elements in the initial array don't have to be increasing order. Thanks! • » » 5 years ago, # ^ |   +3 3 1 4 6 your approach would give -1 as 2 is absent which is gcd of 4, 6 but answer is possible. answer : 5 1 4 1 6 1 • » » » 5 years ago, # ^ |   0 Thanks. I unfortunately set a nonexistent constraint for myself :\ • » » » » 5 years ago, # ^ |   0 can you tell me why absence of 2 is neglected???? • » » » » » 5 years ago, # ^ | ← Rev. 2 →   +3 because in the final answer array every alternate position is occupied by the number which is divisible by every other number say num. so it is easy to see that now for every subarray of size greater than 1 you can achieve num as gcd. Rest remains the case for every single element (for which gcd is equal to the element itself) which are already present in the original array. if no candidate is possible for num then answer is -1. • » » » » 5 years ago, # ^ |   +3 you are not alone bro :/ » 5 years ago, # |   +1 HackForces~ » 5 years ago, # | ← Rev. 2 →   +2 Hack for C : 3 3 18 24 Answer is 4 3 18 3 24 • » » 5 years ago, # ^ |   0 What's the answer for this test case? • » » 5 years ago, # ^ |   0 when we take gcd(18,24) we get 6 but 6 is no where present in the set. how is this possible answer? am i missing something? • » » » 5 years ago, # ^ |   0 You're missing the 3 between 18 and 24. • » » » 5 years ago, # ^ |   0 This is answer, because you need take gcd on segment, so gcd(18,3,24)=3 and all is correct • » » » 5 years ago, # ^ |   0 You calculate gcd(ai, ai + 1, ..., aj) for every 1 ≤ i ≤ j ≤ n, so there won't be gcd(18,24) at all, instead, it is gcd(18,3,24), which is 3. » 5 years ago, # |   0 Troll C-task easier then B. Good job, bro • » » 5 years ago, # ^ |   +13 Nop,after the systest maybe you'll change your opinion. • » » » 5 years ago, # ^ |   0 mmm...no non-trivial formula for B and example S[0],S[1],S[0],S[2],S[0],S[4]... (and some exception, then S[i]%S[0] !=0) » 5 years ago, # |   +39 Half of participants during the contest » 5 years ago, # |   0 It is a hacker's round...XD » 5 years ago, # |   +14 • » » 5 years ago, # ^ |   0 everyone is dead, except YOU!!!!! » 5 years ago, # |   0 Can Anyone Explain How to solve Problem B » 5 years ago, # |   +4 Problem C is a very nice trolling problem :D » 5 years ago, # |   0 How to solve D? • » » 5 years ago, # ^ |   0 It is almost a complete binary tree and has max depth logN. To solve query for V, I checked all ancestors as LCA and did a range query on the subtree of the other child using persistent segtree (can be done offline with BIT too). Code • » » » 5 years ago, # ^ |   0 Your method is too complicated. The author's solution uses merge sort to pre-process and takes O(log^2) time to check the answer by brute force. • » » » » 5 years ago, # ^ | ← Rev. 2 →   +3 Okay, yeah I know it is too complicated but it struck me immediately and I coded it. My solution solves in log^2 time too but I guess persistence is unnecessary. :) • » » » » » 5 years ago, # ^ |   +1 umm. could you please explain the persistent segtree part. I didnt get you exactly. Thanks • » » » » » » 5 years ago, # ^ |   +3 Given vertex V and parameter L, I need to find sum of distances from V to any vertex U in its subtree such that dist(V, U) <= L. I flattened the tree and constructed persistent segtree on it. So, the query reduces to find the sum of values in some subarray such that value <= L. • » » » » » » » 5 years ago, # ^ |   0 How do you manage not to count twice or more the nodes of his subtree. For example in node U you count some childen node (call it x), and then when you climb up to U's parent maybe you can count again that x node. • » » » » » » » » 5 years ago, # ^ |   +3 I query the segtree on the siblings of ancestors of U and not on any ancestor of U. Those subtrees are mutually disjoint. • » » » » » » » » » 5 years ago, # ^ |   0 Got it. Thanks » 5 years ago, # |   +1 y u do zis » 5 years ago, # | ← Rev. 2 →   0 although i am a bad coder because i dont practise algos efficiently i have this mathematical intution for b which lead m to solution.I though answer as 1 and 16 and immediatley power of 2 came to my mind 16 is 2^2*2 and 1 is 2^0*0 so i guessed 2^n-1*m-1.Immediately lightning fell i though i could fill up n-1*m-1 internal squares nyway then fill last row in last column my wauy. That 16 was a big hint. Power of 2 • » » 5 years ago, # ^ |   +3 lightning electrocuted me • » » 5 years ago, # ^ |   0 This was my thought too but I got hacked. I don't think you can fill up the internal squares in any way you like and then fill last row and column to make it work. For example consider a 4x4 grid with k=1. One way to fill in the internal 3x3 is: 1 1 1 1 -1 1 -1 1 -1 But there is no way to complete this. The last row would have to be: -1 1 -1 1 and the last column would have to be: 1 -1 1 -1 So we get a conflict in the bottom right cell. • » » » 5 years ago, # ^ | ← Rev. 2 →   -14 You miscalculated the last row with your example. With you example the last row should be all -1's because each column currently has a product of -1. • » » » 5 years ago, # ^ |   0 Oops nevermind, I made a mistake. I put the last row wrong. » 5 years ago, # | ← Rev. 2 →   +3 System test already done lolFastest ever • » » 5 years ago, # ^ |   0 Almost nothing to test except A. » 5 years ago, # |   +10 Some of the weakest pretests I have ever seen in my life. Solutions with n and m defined as integer passed the pretests in problem B. Solutions that didn't check the case (n + m) % 2 = 1 and k = -1 passed the pretests in B. And problem C pretests were obviously a troll because most people solved a completely different problem yet passed the pretests. gg problem setters » 5 years ago, # |   0 Diabolical test case for C (I guess):41 3 15 20 » 5 years ago, # |   +23 Today, I had a bug inside a bug in problem C which saved me from getting hacked! // ok? bool ok = true; for (int i=1; i<=n; i++) { int g = 0; for (int j=1; j<=n; j++) { g = gcd(g, a[j]); if (!x[g]) { ok = false; } } } int gg = accumulate(a+1, a+n+1, 0, gcd); if (!ok) { cout << -1; return 0; } So, my original idea was to check, for every subsegment of the given set, whether its GCD appears in the set. Now this is obviously wrong (testcase = [1, 4, 6]), but in doing so I misimplemented this piece of code and I actually checked whether the whole set's GCD was in the set, which is, coincidentally, exactly what I should have done, only not n times! :D Notice how the inner loop starts from j = 1 and not j = i. • » » 5 years ago, # ^ | ← Rev. 2 →   +8 One guy in my room had same situation :D • » » 5 years ago, # ^ |   0 This is literally fight bug with bug » 5 years ago, # |   +58 Room 110 (my room) » 5 years ago, # |   0 my code with scanf accepted but with cin it got TLE!Not cool author! not cool... • » » 5 years ago, # ^ | ← Rev. 2 →   -17 Deleted • » » » 5 years ago, # ^ |   +8 I am sorry but I couldnt see in D and E statements it being suggested. • » » » 5 years ago, # ^ |   -24 Oops,my fault!I had suggested,but it was removed later. Sorry for not suggesting faster ways to input or output. • » » » » 5 years ago, # ^ |   +32 Why do you need to recommend faster I/O at all? Wouldn't just setting higher time-limits be a better way of dealing with it, unless you already had some slow nearly-passing solution in mind?If a problem's only solvable with fast I/O that means that most likely you can only solve it in a very small number of languages. In fact, in this contest 100% of the accepted solutions were in C++. There are a few correct-looking Javas but they all got TLE. That's not great. • » » » » » 5 years ago, # ^ | ← Rev. 2 →   0 First, scanf works better than cin, I think. And if we guarantee that cin is ok, the time-limits must be set much higher. Some TLE algorithm could pass tests easily (such as some slow O(n(log(n))2) algorithm for problem D). And we don't want this happen. • » » » » 5 years ago, # ^ |   +15 Also, did you decide not to include max tests into pretests intentionally? I agree that contestants can (and in a lot of cases should) generate max.tests by themselves to be sure that their solutions work, but still — making such pretests sounds like a way to decrease success rate at first place :) • » » » » » 5 years ago, # ^ | ← Rev. 3 →   -22 Wrong information. Deleted. Sorry for no max-size cases in pretests. • » » » » » » 5 years ago, # ^ | ← Rev. 3 →   +33 OK, let's take a look.My cin/cout solution to E which passed pretests fails on test #14 with n=400k, m=1e6. Looking at first 13 tests and assuming that all of them are pretests, the biggest one I see has n=1e5, m=233333. It means n+m=1e6/3=1/6 of max.test.My cin/cout solution for D fails on test #18 with n=1e6, m=1e5. Looking at first 17 tests, there is test #8 with n=438275, m=22553, and a bunch of tests with n=2e5, m=1e5. Adding it up for test #8, we'll get ~4.6e5, less than 1/2 of max test.So can you please tell exact constraints of largest tests provided in pretests? It seems to me like they are quite far away from being maxtests. Yep, maybe they are rather large — but from your comment it sounds like tests are "special" if N=1e6 and ordinary if N=1e5 :) I already saw your comment in contest announcement, and that's why I'm curious — is the story with D and E same and you expected that it will lead to a lot of hacks, or you just wanted more people to fail, or was it some sort of preventing server load or anything like that, or was it just a miss in contest preparation, or what was the motivation at all.In case it was done trying to increase number of hacks — such strategy doesn't work well :) I see as many as total of 2 hacked solutions on E and no hacked solutions on D. I didn't check if they failed because of TL or not :) First of all, not many people are trying to hack hard problems (most of the contestants don't even solve them), and second — hacking I/O is generally harder and takes more effort — you have to be sure about it and know some benchmarks, and there are quite a few tricks like "this thing works fast in GCC but slow in MVS" etc.. Also, I believe that most of the contestants think about it like "OK, there is probably maxtest in pretests, so if they passed — that should be fast enough". It may be different for some trickier tasks, when people try to push NsqrtNlogN algo for 5e5 or take naive solutions with a bunch of optimizations and breaks, but not for something like D/E yesterday, where complexity of any reasonable solution barely depends on input structure — but only on its size. • » » » » » » » 5 years ago, # ^ |   -12 Sorry for any inconvenience caused.We set a tight time limit on D and E because we don't want to see slow solutions pass.For example,someone used O(nlog^2n) solution on D which is not intended and someone used brute force to calculate the answer for each edge which is O(m*sqrt(Ai)).We just didn't want to let this kind of solution pass.As a result,cin/cout may get TLE. • » » » » » » » » 5 years ago, # ^ |   -12 Also,you said that we didn't put maximum tests in the pretests.However,in our opinion,a contestant should be able to generate some data and use custom invocation to check whether the solution can fit in the time limit.And there isn't any regulation saying maximum tests should be put in the tests.If it made you feel not so good,we apologize for this. • » » » » » » 5 years ago, # ^ |   +11 Sorry for my wrong information. Someone told me we have max-size cases in pretests. » 5 years ago, # |   0 And have fun FST on Problem C. QAQ » 5 years ago, # |   +5 I like the way of setting pretests and main tests ;-) good job • » » 5 years ago, # ^ |   0 Counter strike fan ? • » » » 5 years ago, # ^ |   0 yes » 5 years ago, # | ← Rev. 2 →   0 Less Codeforces, more Hackforces. » 5 years ago, # | ← Rev. 2 →   -15 B was TOO DIFFICULT! Lol, now the correct submissions after system tests are in perfect order: A>B>C>D>E. » 5 years ago, # |   -7 What a fuck. It was so fucking hard. A easy, C normal C. But dudes come on. WTF is with B. Even if you find the correct formula like i did, you still need to take care of corner cases and also B and fast exponentiation that's funny. » 5 years ago, # | ← Rev. 2 →   +20 B is pure evil! » 5 years ago, # |   0 Why time limits for first three problems are 1 sec? Usually its 2 sec. • » » 5 years ago, # ^ |   0 The solution of the first three problems is very fast. So there's no need for 2sec • » » 5 years ago, # ^ |   +4 squirtle used used 'hard b and c' attack critical hit sandshrew fainted » 5 years ago, # |   0 Hi guys whats wrong with this http://codeforces.com/contest/894/submission/32469968 • » » 5 years ago, # ^ |   0 i am very sad i got the algo in contest but dont know why it got hacked • » » 5 years ago, # ^ | ← Rev. 2 →   +3 read my next comment • » » 5 years ago, # ^ |   +3 bro in youre pow function you should set it to :long long pow(long long x, long long y, long long mod) { ... }you've overflowed ;-) • » » » 5 years ago, # ^ | ← Rev. 2 →   +3 shit you are right it got accepted now. feels sad small mistake destroyed me.Else i culd have felt happy that i solved b in contest Thank you so much » 5 years ago, # | ← Rev. 3 →   +2 For those who can't understand why they got hack on B, here is a problem using Fermat's Little Theorem -> see the editorial from this problem https://www.hackerrank.com/contests/infinitum18/challenges/tower-3-coloring; why a^b mod c is a^(b mod(c-1)) mod c • » » 5 years ago, # ^ |   +24 No need.It's too complicated.We can calculate in this way:pow(pow(2,n-1),m-1) » 5 years ago, # | ← Rev. 2 →   +30 B and C were harder than normal (perhaps it would have been better to lower the constraints to avoid fast exponentiation?), but overall this round was not as hard as some other Div 2 rounds, considering that I managed to finish in 50 minutes ...... well, until I realized that my solution to E was incorrect. Anyways, I have nothing against weak pretests. :P » 5 years ago, # |   +34 Today's Round is the Proof of The First Law of Online Judges. » 5 years ago, # |   +132 • » » 5 years ago, # ^ |   +3 They are twince ! Hacked ! • » » 5 years ago, # ^ |   0 For me, A as well. Some people forgot that string.size() returns an unsigned integer. So using string.size() — 2 when string length is 1 will lead to overflow. » 5 years ago, # |   +2 For Prob C (Marco And GCD Seq)Input: 151 2 3 4 5 6 7 8 9 10 11 12 13 14 15Ouput: 106 7 8 9 10 11 12 13 14 15Why this is given wa verdict? (set of pairwise gcd of elements produced the input set) • » » 5 years ago, # ^ |   +2 You never generate any of 2,3,4,5 as gcds. • » » » 5 years ago, # ^ |   0 gcd(12,14)=2 gcd(6,9)=3 gcd(8,12)=4 gcd(10,15_=5 • » » » » 5 years ago, # ^ |   +3 The gcd must be of consecutive elements. • » » 5 years ago, # ^ |   0 You cannot get 2 using your output sequence. • » » » 5 years ago, # ^ |   0 gcd(12,14)=2 • » » » » 5 years ago, # ^ |   +3 It needs gcd(ai,ai+1,ai+2,...,aj) So gcd(12,13,14)=1 • » » 5 years ago, # ^ |   0 because you can not find gcd 2,3,4,5 (i<=x<=j) » 5 years ago, # |   -8 Solution to problem A. QAQ was already available on the internet, well before the contest.Please visit source — http://www.geeksforgeeks.org/find-number-times-string-occurs-given-string/ Hence I don't think people should be penalised if their code matches with someone...as they both referred the same source....codeforces please see into it ! • » » 5 years ago, # ^ |   0 emmm……At first, my wrong B is skipped…… • » » 5 years ago, # ^ |   0 We will fix that. • » » » 5 years ago, # ^ |   0 Thanks • » » » 5 years ago, # ^ |   0 Can you please also take a look at my comment below? Thanks! • » » 5 years ago, # ^ |   0 Mine too... » 5 years ago, # |   +11 Amazing problems... Amazing system testing... Amazing Ratings Update.. Amazing CONTEST » 5 years ago, # | ← Rev. 2 →   -41 the contest was bad. dumb weeaboo trash and weak preliminary testcases. • » » 5 years ago, # ^ |   +16 No. I disagree with you. So maybe there were mistakes somewhere. But they did work to create a contest. You do not have the right to say that it was bad so-as you have never conducted a contest. Sorry if I did not put it right. • » » » 5 years ago, # ^ | ← Rev. 2 →   +11 While I don't agree with his sentiment, what you are saying is totally wrong too. You don't need to conduct a contest to recognise a bad one. That being said, I don't think today's contest was bad. However, he is free to express his opinion. • » » » » 5 years ago, # ^ |   +2 I think that a person can not understand the organizer until he tries to hold a contest. • » » » » » 5 years ago, # ^ |   +5 Agreed. And his manner of expressing his disappointment was very immature too. » 5 years ago, # |   0 if test case is: 3 1 6 9 and answer is 6 1 1 6 1 9 1 why we can't take gcd(6,9)==3; which is missing in test case. » 5 years ago, # |   0 Don't ever put a contest again ever are you sure that it was for DIV 2 ? really ? • » » 5 years ago, # ^ | ← Rev. 2 →   +3 I partially agree with you since problem B was at the level of task D. But everything else was excellent. It is my opinion. » 5 years ago, # |   +11 I am disqulafied ,why????? My solution was by me . I am disagree with this decision • » » 5 years ago, # ^ | ← Rev. 3 →   0 I think you should create a separate blog about your problem, because there will be more people there and the Administration of the Codeforces will see your appeal. » 5 years ago, # |   -27 really terrible contest :/very easy problems and shity codesvery very easy problem E and because of fast i/o 80% failedworst contest ever with shity problems that sucks » 5 years ago, # |   0 Wow this is the hardest CF Round I have participated. » 5 years ago, # |   +17 Why the ratings are altered again now? • » » 5 years ago, # ^ |   +3 maybe rejudging » 5 years ago, # |   +16 codeforces should become ready for the goodbye 2017 contest which has the most participants :) it did very well this round i hope it remains the same » 5 years ago, # |   +1 Ohhhh!! Hackforces Round!!! » 5 years ago, # |   +1 Really bad time limit for problem E due to big i/o Input TLEios::sync_with_stdio(false); cin.tie(0); cout.tie(0); cin >> ... Input ACinline void Scanf(ll& a) //(https://github.com/mvpossum/eldiego) { char c = 0; while(c<33) c = getc(stdin); a = 0; while(c>33) a = a*10 + c - '0', c = getc(stdin); } ios::sync_with_stdio(false); cin.tie(0); cout.tie(0); Scanf(...) • » » 5 years ago, # ^ |   +3 D as well. Difference between fast IO and scanf/printf is about 1.5s, which is more than half the time limit.You have to be soo careful during implementation to even have a chance at getting AC without fast IO. • » » 5 years ago, # ^ |   0 Author's code uses scanf/printf for D and E. For D,it runs in 1.1 seconds and for E,it runs in 1 second. So maybe cin and cout without ios::sync_with_stdio(false) will get TLE,otherwise,maybe your complexity isn't the intended one,so it gets TLE. For example,the intended time complexity for D is O(nlogn+m(logn)^2) and for E it is O(n+m). » 5 years ago, # |   +17 C was such bait. Made the observation that the largest number in S must be in the sequence, and then tried to construct it from there... red herring.Realised the correct solution with 7 minutes left in the contest. » 5 years ago, # |   -25 What the hell was this? the worst round ever. The worst problems ever! Go home. • » » 5 years ago, # ^ |   +4 If you're not satisfied with the contest,please point out where we didn't do well.Only blaming will make no sense and will never help us improve and prepare better problems.Thank you! • » » » 5 years ago, # ^ |   +10 The problems where very interesting, But the pretest where weak that is why. » 5 years ago, # |   +1 Rip those submissions on C, where people saw the score board, saw high and extremely fast increasing rate of pretest passed, panicked, wrote some garbage greedy and then submitted themselves.And they got pretests passed too ! Too bad, WA. Never ever open score board during the contest is what seems best to me. The problem C is not very difficult and things may have been different with strong pretests.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3468058109283447, "perplexity": 2172.1460753040064}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103943339.53/warc/CC-MAIN-20220701155803-20220701185803-00391.warc.gz"}
https://www.mathdoubts.com/multiplying-matrices-problems-solutions/
# Multiplying the matrices Problems and solutions In mathematics, the matrices are involved in multiplication. It is not actually possible to multiply a matrix by a matrix directly because there is a systematic procedure to multiply the matrices. So, let’s learn how to multiply the matrices mathematically with different cases from the understandable example problems. ### Matrices of the same order List of the $2 \times 2$ matrix multiplication problems with step by step solution to learn how to multiply matrices of the order $2$. List of the $3 \times 3$ matrix multiplication questions with step by step solution to learn how to multiply matrices of the order $3$. Email subscription Math Doubts is a best place to learn mathematics and from basics to advanced scientific level for students, teachers and researchers. Know more
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9584933519363403, "perplexity": 395.4559129978995}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704820894.84/warc/CC-MAIN-20210127024104-20210127054104-00192.warc.gz"}
http://mathoverflow.net/questions/100033/interesting-mathematical-documentaries/100046
# Interesting mathematical documentaries I am looking for mathematical documentaries, both technical and non-technical. They should be "interesting" in that they present either actual mathematics, mathematicians or history of mathematics. I am in charge of nourishing our departmental math library (PUCV) and I would like to add this kind of material in order to attract undergraduates toward mathematics. For this reason, I am not looking for videos of conferences or seminar talks, but rather for introductory or "wide public" material. Here are some good examples. Are there more examples? Thanks, Ricardo. - Not exactly a documentary, but appropriate for an undergrad math library: youtube.com/watch?v=wO61D9x6lNY – Aaron Hoffman Jun 19 '12 at 18:57 Consult this other question mathoverflow.net/questions/1714/best-online-math-videos ... answers include documentaries, perhaps. – Gerald Edgar Jun 19 '12 at 19:10 This question should surely be community wiki. – HJRW Jun 19 '12 at 19:41 Yes, Ricardo: please edit your question and click on the option "communiy wiki" (see the FAQ for more information on what that is). – André Henriques Jun 19 '12 at 19:55 @André Henriques and @HW: thanks, it's done. – Ricardo Menares Jun 19 '12 at 21:38 Leys, Ghys & Alvarez have also made a video series in a similar style about dynamical systems called "Chaos". The nine chapters are all available under a Creative Commons license. - $N$ Is a Number: A Portrait of Paul Erdős. (Official site; Wikipedia link) - Very Interesting and i finally learned the correct pronounciation of Erdos through this documentary, thanks! – Michael Kissner Jun 20 '12 at 17:57 "The Color of Math" is a rather poetic documentary by Katia Eremenko that will have its premiere at the Moscow Film Festival this upcoming Saturday, June 23, 2012. It features Cedric Villani, Anatoly Fomenko, Maxim Kontsevic, Jean-Michel Bismut, Aaditya V. Rangan (and myself). - Awesome! Will it be released internationally? – David Roberts Jun 21 '12 at 2:22 Press conference on the film: youtube.com/watch?v=b9c7zhsK-hM (in Russian) – David Roberts Jul 11 '12 at 7:23 Any relation between Katia Eremenko and MathOverflow user Alexandre Eremenko? – Todd Trimble Dec 16 '12 at 16:28 I really like this short documentary on Fermats Last Theorem and Andrew Wiles struggle and setbacks proving it. - That's the BBC documentary I mentioned in the question. I would like to know how to purchase it, do you have any information about this? Been able to watch it on youtube is of course fine, but the idea is to have it displayed in the library in order to attract potential viewers who wouldn't look for this kind of material on their own initiative. – Ricardo Menares Jun 19 '12 at 21:37 Ahh sorry, missed that part. But, here's a link to a website where you can purchase it for use in classes i believe: bbcactivevideoforlearning.com/1/TitleDetails.aspx?TitleID=664 – Michael Kissner Jun 20 '12 at 7:01 I believe that is the same as the documentary aired on PBS in the US as a Nova episode called "The Proof". It doesnt seem to be available on DVD from PBS, but Amazon has it on VHS as do several other online retailers. – user4535 Jun 21 '12 at 14:14 Programs are often produced jointly for showing on Horizon on BBC and Nova on PBS. – Gerald Edgar Feb 15 '13 at 19:25 Link is now dead. Here is an updated link: dailymotion.com/video/… – Al Jebr Dec 19 '14 at 5:28 The soundtrack is still in progress, so it's not yet fully available but Ester Dalvit made a very interesting movie about braids and knots. At the end it will be distributed under a Creative commons license. A trailer and parts of the movie are available here. Up to now three (old versions of) chapters are available, explaining respectively through computer generated animations: • the group structure of the braid group • "combing" of braids and handle reduction (i.e. solutions to the world problem) • Alexander and Markov theorems (i.e. the relation with knots). - Here the synopsis from IMDB: "When France surrendered in 1940 and German soldiers showed up in the Vosgian village of Housseras, an unknown French foot soldier burned his papers and killed himself in a farmer's barn. Four years later he was identified as "soldat Doblin, Vincent". In fact, he was none other than the mathematician Wolfgang Doeblin, son of the famous German novelist Alfred Döblin ("Berlin Alexanderplatz") who was forced to flee Nazi Germany with his family in 1933. A French citizen since October 1936, Wolfgang Doeblin carried on his research into probability theory during his military service and even during the hardships of the "Phoney War" in the winter of 1939-40. In February 1940, four months before his death at the age of 25, he sent his most important manuscripts ("About the Kolmogoroff Equation") as a "sealed envelope" to the Academy of Science in Paris, where they were kept in safe custody for 60 years. Wolfgang Doeblin's short and dramatic life story, almost forgotten, was finally brought into the limelight when the "sealed envelope" was opened in May 2000. Far ahead of their time, his groundbreaking contributions to the theory of random processes place Wolfgang Doeblin among the major innovators of probability, the "mathematics of randomness". Mathematical models for evaluation of chances and risks went on to gain major importance in numerous domains of modern science, in everyday life and especially in contemporary financial mathematics." - The German title would be: Der versiegelte Brief des Soldaten Döblin. – Jose Capco Jul 4 '14 at 8:58 Something completely different, but I'm sure a lot of people here at MO would enjoy: Beautiful young minds British documentary about International Mathematical Olympiad 2006 with focus on the UK team. Hard problems American documentary about the same event with focus on the US team. Documentaries contain analysis of that year's problems (although the focus is on the competition in general, team selection and contestants - which makes it more interesting to ex-contestants than to professional mathematicians or math enthusiasts). - Marcus du Sautoy's The Story of Maths is a total of four hours attempting to give an overview of the history of mathematics from ancient to modern time, spending 5-10 minutes each on the life and work of some of the most famous mathematicians. While one could quibble with some of the selections, the project is overall a fantastic production. - "Taking the long view: The life of Shiing-Shen Chern" http://takingthelongviewfilm.com/ This very nice documentary was produced recently on the occasion of the Chern Centennial Conference. - Alan Turing: Codebreaker and AI Pioneer (available here). The code-breaking work at Bletchley Park, which helped save Britain from Nazi Germany, qualifies as one of the greatest stories of World War II, and the misunderstood genius, Alan Turing, stands at the center of this tale. Perhaps no one understands Turing's role during this period -- and his larger impact on mathematics and computing -- like B. Jack Copeland. In this lecture, Copeland contends that Turing should be celebrated as the father of artificial intelligence. - Dangerous Knowledge: http://topdocumentaryfilms.com/dangerous-knowledge/ Though this one is more about mathematicians, than mathematics. In this one-off documentary, David Malone looks at four brilliant mathematicians – Georg Cantor, Ludwig Boltzmann, Kurt Gödel and Alan Turing – whose genius has profoundly affected us, but which tragically drove them insane and eventually led to them all committing suicide. - Do we really want to encourage the view that mathematics is done by insane geniuses? – Steven Gubkin Jun 20 '12 at 1:19 Except for Gödel, it is grossly inaccurate to say that they became insane, even because of their genius. For Cantor and Boltzmann, this was because of the bad acception of their works, and for Turing, because of terribly stupid law. – BS. Jun 20 '12 at 9:17 It seems very likely that Boltzmann suffered from bipolar disorder. However, I agree with Steven that this Romantic trope of the "thin line between madness and genius" can be very tiresome indeed. – Todd Trimble Dec 16 '12 at 16:27 Science Lives, made by the Simons Foundation, has nine very long interviews that essentially amount to miniature documentaries about the lives of a number of 20th/21st century mathematicians and physicists. - PBS documentary A Brilliant Madness that looks at the life of Nobel-prize winning mathematician, John Nash - "The Proof", from NOVA, presents Andrew Wiles and the story of Fermat's Last Theorem. http://video.google.com/videoplay?docid=8269328330690408516 - This is an excellent film. I think it may be the film Ricardo refers to as "BBC documentary on Fermat's last theorem", under a different name. – Nik Weaver Jun 1 at 1:14 "Achever l'inachevable" or "Achieving the Unachievable" features many mathematicians discussing M.C. Escher's Print Gallery, most of the material coming from the Notices article here. - A short recent documentary on the life and mathematics of Évariste Galois, produced by the CNRS : http://videotheque.cnrs.fr/index.php?urlaction=doc&id_doc=2839 This documentary can be viewed online here : http://videotheque.cnrs.fr/video.php?urlaction=visualisation&method=QT&action=visu&id=2839&type=grandPublic - Does it manage to avoid 'Men of Mathematics'-style hype? – David Roberts Jun 21 '12 at 2:22 @David : yes, the documentary is accurate and there are interviews of historians of mathematics (Caroline Ehrhardt, Jeremy Gray). – François Brunault Jun 21 '12 at 7:51 The secret life of chaos. A BBC documentary. - I vote for 100年の難問はなぜ解けたのか ~天才数学者 失踪の謎 ("Why the difficult 100-year-old conjecture was proven- The mysterious disappearance of the genius mathematician"), about Grigory Perelman, Geometrization, and the proof of the Poincaré Conjecture, as the best mathematical documentary ever. It's non-technical, but it's really very well-made, and gives a good sense of what research mathematics is like, and why some people spend their lives searching for proofs of mathematical conjectures. The mathematics in it is mostly fairly solid. It inspired fashion design house Issey Miyake to design a fashion collection with Bill Thurston inspired by geometrization. - translate by google: Что такое Hatewa Вселенной? Почему на земле, что Вселенная форме? Совет ткацкий станок большой тайной для многих лет человечество было спрашивал был найден в прошлом году. :):):) – Alexander Chervov Jul 11 '12 at 18:00 Google translate is horrible at translating from Japanese... – Daniel Moskovich Jul 11 '12 at 22:11 There is a new movie that just came out called Comment j'ai détesté les Maths (How I hated math). Cédric Villani, Jim Simons, and others are featured in the film. - That link does not go to the movie. – Byron Schmuland Aug 19 '15 at 0:46 The man who saved Geometry is a documentary on D. Coxeter: https://vimeo.com/120725835 - The only documentary about Grigori Perelman has recently got English subtitles. (Perelman is the guy who proved the Poincare conjecture and then refused a million dollar prize.) It is for laymen, and very dramatic. - - The Geometry Center (formerly of UMN, now apparently defunct) many years ago produced "Not Knot" (about hyperbolic space) and "Outside In" (about sphere eversion). There is apparently a more recent third one "The Shape of Space" I'm not familiar with. Apparently you can still order them here, although copies are now easily found on YouTube. - http://www.math.ru/lib/files/pdf/mehmat/mm3.pdf Interviews and memories by leading Moscow mathematicians I.R. Shafarevich, V.I. Arnold, S.P. Novikov, B.A. Sevostyanov, A.N. Shiryaev, V.V. Golubev (major-general and dean of mexmat MSU at (1933-1935 и 1944-1952)). Some of them are quite free-spoken. PS Sorry it is in Russian. - THE THINGS YOU DON'T UNDERSTAND, 2010, Peter and Petra Lataster Mathematician Hendrik Lenstra has no car, no computer and no television. But he is the proud owner of numerous editions of Homer’s Odyssey. Directors Petra & Peter Lataster draw a portrait of Lenstra, a mathematician who thinks on the highest level. And also in daily life he is not interested in the superfluous and the fashionable. http://www.latasterfilms.nl/en/films/the-things-you-dont-understand/ - That's not that surprising to me. I have no car (and no driver's license) and no television either, and if MO didn't exist, I'm not sure I'd even have a computer. And I collect Dance Music CDs from the 90's as well as editions of "The girl who played with fire" (or should I say "Flickan som lekte med elden"?) in different languages :-) – Sylvain JULIEN Jan 16 '14 at 20:31 The page you linked to has an error on it and the film is not available there. – KConrad Apr 8 '15 at 15:03 A new documentary about Ramanujan just screened at the AMS meeting earlier this evening. It's called, The Genius of Srinivasa Ramanujan. It was claimed that it's already up on YouTube. EDIT: All I was able to find on the web was a very short preview. - - Nova's Mathematical Mystery Tour is a really nice one and is on youtube. RAMANUJAN: Letters from an Indian Clerk again available on youtube.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.290752112865448, "perplexity": 3910.2173901389842}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-26/segments/1466783395613.65/warc/CC-MAIN-20160624154955-00120-ip-10-164-35-72.ec2.internal.warc.gz"}