prompt
listlengths
1
1
compression_prompt
listlengths
1
1
target
stringlengths
1.03k
828k
[ { "content": "Reconstruct the code exactly:\n```python\n#!/usr/bin/env python\n\n# GrovePi LED blink test for the Grove LED Socket (http://www.seeedstudio.com/wiki/Grove_-_LED_Socket_Kit)\n#\n# The GrovePi connects the Raspberry Pi and Grove sensors. You can learn more about GrovePi here: http://www.dexterind...
[ { "content": "Reconstruct the code exactly:\n<|memory_start|>```python\n#!/usr/bin/env python\n\n# GrovePi LED blink test for the Grove LED Socket (http://www.seeedstudio.com/wiki/Grove_-_LED_Socket_Kit)\n#\n# The GrovePi connects the Raspberry Pi and Grove sensors. You can learn more about GrovePi here: http...
```python #!/usr/bin/env python # GrovePi LED blink test for the Grove LED Socket (http://www.seeedstudio.com/wiki/Grove_-_LED_Socket_Kit) # # The GrovePi connects the Raspberry Pi and Grove sensors. You can learn more about GrovePi here: http://www.dexterindustries.com/GrovePi # # Have a question about this example? Ask on the forums here: http://www.dexterindustries.com/forum/?forum=grovepi # ''' ## License The MIT License (MIT) GrovePi for the Raspberry Pi: an open source platform for connecting Grove Sensors to the Raspberry Pi. Copyright (C) 2015 Dexter Industries Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ''' import time from grovepi import * # Connect the Grove LED to digital port D4,D5,D6 led0 = 4 led1 = 5 led2 = 6 pinMode(led0,"OUTPUT") pinMode(led1,"OUTPUT") pinMode(led2,"OUTPUT") while True: try: #Blink the LED digitalWrite(led0,1) # Send HIGH to switch on LED digitalWrite(led1,1) # Send HIGH to switch on LED digitalWrite(led2,1) # Send HIGH to switch on LED print ("LED ON!") time.sleep(1) digitalWrite(led0,0) # Send LOW to switch off LED digitalWrite(led1,0) # Send LOW to switch off LED digitalWrite(led2,0) # Send LOW to switch off LED print ("LED OFF!") time.sleep(1) except IOError: # Print "Error" if communication error encountered print ("Error") ```
[ { "content": "```python\ndef cells():\n '''\n # Linear algebra overview\n '''\n\n '''\n '''\n\n '''\n Linear algebra is the study of **vectors** and **linear transformations**. This notebook introduces concepts form linear algebra in a birds-eye overview. The goal is not to get into the det...
[ { "content": "<|memory_start|>```python\ndef cells():\n '''\n # Linear algebra overview\n '''\n\n '''\n '''\n\n '''\n Linear algebra is the study of **vectors** and **linear transformations**. This notebook introduces concepts form linear algebra in a birds-eye overview. The goal is not to ...
```python def cells(): ''' # Linear algebra overview ''' ''' ''' ''' Linear algebra is the study of **vectors** and **linear transformations**. This notebook introduces concepts form linear algebra in a birds-eye overview. The goal is not to get into the details, but to give the reader a taste of the different types of thinking: computational, geometrical, and theoretical, that are used in linear algebra. ''' ''' ''' ''' ## Chapters overview - 1/ Math fundamentals - 2/ Intro to linear algebra - Vectors - Matrices - Matrix-vector product representation of linear transformations - Linear property: $f(a\mathbf{x} + b\mathbf{y}) = af(\mathbf{x}) + bf(\mathbf{y})$ - 3/ Computational linear algebra - Gauss-Jordan elimination procedure - Augemnted matrix representaiton of systems of linear equations - Reduced row echelon form - Matrix equations - Matrix operations - Matrix product - Determinant - Matrix inverse - 4/ Geometrical linear algebra - Points, lines, and planes - Projection operation - Coordinates - Vector spaces - Vector space techniques - 5/ Linear transformations - Vector functions - Input and output spaces - Matrix representation of linear transformations - Column space and row spaces of matrix representations - Invertible matrix theorem - 6/ Theoretical linear algebra - Eigenvalues and eigenvectors - Special types of matrices - Abstract vectors paces - Abstract inner product spaces - Gram–Schmidt orthogonalization - Matrix decompositions - Linear algebra with complex numbers - 7/ Applications - 8/ Probability theory - 9/ Quantum mechanics - Notation appendix ''' ''' ''' # helper code needed for running in colab if 'google.colab' in str(get_ipython()): print('Downloading plot_helpers.py to util/ (only neded for colab') !mkdir util; wget https://raw.githubusercontent.com/minireference/noBSLAnotebooks/master/util/plot_helpers.py -P util ''' ''' # setup SymPy from sympy import * x, y, z, t = symbols('x y z t') init_printing() # a vector is a special type of matrix (an n-vector is either a nx1 or a 1xn matrix) Vector = Matrix # define alias Vector so I don't have to explain this during video Point = Vector # define alias Point for Vector since they're the same thing # setup plotting %matplotlib inline import matplotlib.pyplot as mpl from util.plot_helpers import plot_vec, plot_vecs, plot_line, plot_plane, autoscale_arrows ''' ''' ''' # 1/ Math fundamentals ''' ''' ''' ''' Linear algebra builds upon high school math concepts like: - Numbers (integers, rationals, reals, complex numbers) - Functions ($f(x)$ takes an input $x$ and produces an output $y$) - Basic rules of algebra - Geometry (lines, curves, areas, triangles) - The cartesian plane ''' ''' ''' ''' ''' ''' # 2/ Intro to linear algebra Linear algebra is the study of vectors and matrices. ''' ''' ''' ''' ## Vectors ''' ''' ''' # define two vectors u = Vector([2,3]) v = Vector([3,0]) u ''' ''' v ''' ''' plot_vecs(u, v) autoscale_arrows() ''' ''' ''' ## Vector operations ''' ''' ''' ''' - Addition (denoted $\vec{u}+\vec{v}$) - Subtraction, the inverse of addition (denoted $\vec{u}-\vec{v}$) - Scaling (denoted $\alpha \vec{u}$) - Dot product (denoted $\vec{u} \cdot \vec{v}$) - Cross product (denoted $\vec{u} \times \vec{v}$) ''' ''' ''' ''' ### Vector addition ''' ''' ''' # algebraic u+v ''' ''' # graphical plot_vecs(u, v) plot_vec(v, at=u, color='b') plot_vec(u+v, color='r') autoscale_arrows() ''' ''' ''' ### Basis When we describe the vector as the coordinate pair $(4,6)$, we're implicitly using the *standard basis* $B_s = \{ \hat{\imath}, \hat{\jmath} \}$. The vector $\hat{\imath} \equiv (1,0)$ is a unit-length vector in the $x$-direciton, and $\hat{\jmath} \equiv (0,1)$ is a unit-length vector in the $y$-direction. To be more precise when referring to vectors, we can indicate the basis as a subscript of every cooridnate vector $\vec{v}=(4,6)_{B_s}$, which tells $\vec{v}= 4\hat{\imath}+6\hat{\jmath}=4(1,0) +6(0,1)$. ''' ''' ''' # the standard basis ihat = Vector([1,0]) jhat = Vector([0,1]) v = 4*ihat + 6*jhat v ''' ''' # geomtrically... plot_vecs(ihat, jhat, 4*ihat, 6*jhat, v) autoscale_arrows() ''' ''' ''' The same vector $\vec{v}$ will correspond to the a different pair of coefficients if a differebt basis is used. For example, if we use the basis $B^\prime = \{ (1,1), (1,-1) \}$, the same vector $\vec{v}$ must be expressed as $\vec{v} = 5\vec{b}_1 +(-1)\vec{b}_2=(5,-1)_{B^\prime}$. ''' ''' ''' # another basis B' = { (1,1), (1,-1) } b1 = Vector([ 1, 1]) b2 = Vector([ 1, -1]) v = 5*b1 + (-1)*b2 v # How did I know 5 and -1 are the coefficients w.r.t basis {b1,b2}? # Matrix([[1,1],[1,-1]]).inv()*Vector([4,6]) ''' ''' # geomtrically... plot_vecs(b1, b2, 5*b1, -1*b2, v) autoscale_arrows() ''' ''' ''' ''' ''' ''' ''' ''' ''' ''' ''' ## Matrix operations ''' ''' ''' ''' - Addition (denoted $A+B$) - Subtraction, the inverse of addition (denoted $A-B$) - Scaling by a constant $\alpha$ (denoted $\alpha A$) - Matrix-vector product (denoted $A\vec{x}$, related to linear transformations) - Matrix product (denoted $AB$) - Matrix inverse (denoted $A^{-1}$) - Trace (denoted $\textrm{Tr}(A)$) - Determinant (denoted $\textrm{det}(A)$ or $|A|$) ''' ''' ''' ''' In linear algebra we'll extend the notion of funttion $f:\mathbb{R}\to \mathbb{R}$, to functions that act on vectors called *linear transformations*. We can understand the properties of linear transformations $T$ in analogy with ordinary functions: \begin{align*} \textrm{function } f:\mathbb{R}\to \mathbb{R} & \ \Leftrightarrow \, \begin{array}{l} \textrm{linear transformation } T:\mathbb{R}^{n}\! \to \mathbb{R}^{m} \end{array} \\ \textrm{input } x\in \mathbb{R} & \ \Leftrightarrow \ \textrm{input } \vec{x} \in \mathbb{R}^n \\ \textrm{output } f(x) \in \mathbb{R} & \ \Leftrightarrow \ \textrm{output } T(\vec{x})\in \mathbb{R}^m \\ g\circ\! f \: (x) = g(f(x)) & \ \Leftrightarrow \ % \textrm{matrix product } S(T(\vec{x})) \\ \textrm{function inverse } f^{-1} & \ \Leftrightarrow \ \textrm{inverse transformation } T^{-1} \\ \textrm{zeros of } f & \ \Leftrightarrow \ \textrm{kernel of } T \\ \textrm{image of } f & \ \Leftrightarrow \ \begin{array}{l} \textrm{image of } T \end{array} \end{align*} ''' ''' ''' ''' ## Linear property $$ T(a\mathbf{x}_1 + b\mathbf{x}_2) = aT(\mathbf{x}_1) + bT(\mathbf{x}_2) $$ ''' ''' ''' ''' ## Matrix-vector product representation of linear transformations ''' ''' ''' ''' Equivalence between linear transformstions $T$ and matrices $M_T$: $$ T : \mathbb{R}^n \to \mathbb{R}^m \qquad \Leftrightarrow \qquad M_T \in \mathbb{R}^{m \times n} $$ $$ \vec{y} = T(\vec{x}) \qquad \Leftrightarrow \qquad \vec{y} = M_T\vec{x} $$ ''' ''' ''' ''' ''' ''' ''' ''' ''' ''' ''' ''' # 3/ Computational linear algebra ''' ''' ''' ''' ## Gauss-Jordan elimination procedure Suppose you're asked to solve for $x_1$ and $x_2$ in the following system of equations \begin{align*} 1x_1 + 2x_2 &= 5 \\ 3x_1 + 9x_2 &= 21. \end{align*} ''' ''' ''' # represent as an augmented matrix AUG = Matrix([ [1, 2, 5], [3, 9, 21]]) AUG ''' ''' # eliminate x_1 in second equation by subtracting 3x times the first equation AUG[1,:] = AUG[1,:] - 3*AUG[0,:] AUG ''' ''' # simplify second equation by dividing by 3 AUG[1,:] = AUG[1,:]/3 AUG ''' ''' # eliminate x_2 from first equation by subtracting 2x times the second equation AUG[0,:] = AUG[0,:] - 2*AUG[1,:] AUG ''' ''' ''' This augmented matrix is in *reduced row echelon form* (RREF), and corresponds to the system of equations: \begin{align*} 1x_1 \ \ \qquad &= 1 \\ 1x_2 &= 2, \end{align*} so the the solution is $x_1=1$ and $x_2=2$. ''' ''' ''' ''' ## Matrix equations ''' ''' ''' ''' See **page 177** in v2.2 of the book. ''' ''' ''' ''' ''' ''' ''' ''' ## Matrix product ''' ''' ''' a,b,c,d,e,f, g,h,i,j = symbols('a b c d e f g h i j') A = Matrix([[a,b], [c,d], [e,f]]) B = Matrix([[g,h], [i,j]]) A, B ''' ''' A*B ''' ''' def mat_prod(A, B): """Compute the matrix product of matrices A and B.""" assert A.cols == B.rows, "Error: matrix dimensions not compatible." m, ell = A.shape # A is a m x ell matrix ell, n = B.shape # B is a ell x n matrix C = zeros(m,n) for i in range(0,m): for j in range(0,n): C[i,j] = A[i,:].dot(B[:,j]) return C mat_prod(A,B) ''' ''' # mat_prod(B,A) ''' ''' ''' ## Determinant ''' ''' ''' a, b, c, d = symbols('a b c d') A = Matrix([[a,b], [c,d]]) A.det() ''' ''' # Consider the parallelogram with sides: u1 = Vector([3,0]) u2 = Vector([2,2]) plot_vecs(u1,u2) plot_vec(u1, at=u2, color='k') plot_vec(u2, at=u1, color='b') autoscale_arrows() # What is the area of this parallelogram? ''' ''' # base = 3, height = 2, so area is 6 ''' ''' # Compute the area of the parallelogram with sides u1 and u2 using the deteminant A = Matrix([[3,0], [2,2]]) A.det() ''' ''' ''' ''' ''' ''' ''' ## Matrix inverse For an invertible matrix $A$, the matrix inverse $A^{-1}$ acts to undo the effects of $A$: $$ A^{-1} A \vec{v} = \vec{v}. $$ The effect applying $A$ followed by $A^{-1}$ (or the other way around) is the identity transformation: $$ A^{-1}A \ = \ \mathbb{1} \ = \ AA^{-1}. $$ ''' ''' ''' A = Matrix([[1, 2], [3, 9]]) A ''' ''' # Compute deteminant to check if inverse matrix exists A.det() ''' ''' ''' The deteminant is non-zero so inverse exists. ''' ''' ''' A.inv() ''' ''' A.inv()*A ''' ''' ''' ### Adjugate-matrix formula The *adjugate matrix* of the matrix $A$ is obtained by replacing each entry of the matrix with a partial determinant calculation (called *minors*). The minor $M_{ij}$ is the determinant of $A$ with its $i$th row and $j$th columns removed. ''' ''' ''' A.adjugate() / A.det() ''' ''' ''' ### Augmented matrix approach $$ \left[ \, A \, | \, \mathbb{1} \, \right] \qquad -\textrm{Gauss-Jordan elimination}\rightarrow \qquad \left[ \, \mathbb{1} \, | \, A^{-1} \, \right] $$ ''' ''' ''' AUG = A.row_join(eye(2)) AUG ''' ''' # perform row operations until left side of AUG is in RREF AUG[1,:] = AUG[1,:] - 3*AUG[0,:] AUG[1,:] = AUG[1,:]/3 AUG[0,:] = AUG[0,:] - 2*AUG[1,:] AUG ''' ''' # the inverse of A is in the right side of RREF(AUG) AUG[:,2:5] # == A-inverse ''' ''' # verify A times A-inverse gives the identity matrix... A*AUG[:,2:5] ''' ''' ''' ### Using elementary matrices Each row operation $\mathcal{R}_i$ can be represented as an elementary matrix $E_i$. The elementary matrix of a given row operation is obtained by performing the row operation on the identity matrix. ''' ''' ''' E1 = eye(2) E1[1,:] = E1[1,:] - 3*E1[0,:] E2 = eye(2) E2[1,:] = E2[1,:]/3 E3 = eye(2) E3[0,:] = E3[0,:] - 2*E3[1,:] E1, E2, E3 ''' ''' # the sequence of three row operations transforms the matrix A into RREF E3*E2*E1*A ''' ''' ''' Recall definition $A^{-1}A=\mathbb{1}$, and we just observed that $E_3E_2E_1 A =\mathbb{1}$, so it must be that $A^{-1}=E_3E_2E_1$. ''' ''' ''' E3*E2*E1 ''' ''' ''' ''' ''' ''' ''' # 4/ Geometrical linear algebra Points, lines, and planes are geometrical objects that are conveniently expressed using the language of vectors. ''' ''' ''' ''' ## Points A point $p=(p_x,p_y,p_z)$ refers to a single location in $\mathbb{R}^3$. ''' ''' ''' p = Point([2,4,5]) p ''' ''' ''' ## Lines A line is a one dimensional infinite subset of $\mathbb{R}^3$ that can be described as $$ \ell: \{ p_o + \alpha \vec{v} \ | \ \forall \alpha \in \mathbb{R} \}. $$ ''' ''' ''' po = Point([1,1,1]) v = Vector([1,1,0]) plot_line(v, po) ''' ''' ''' ## Planes A plane is a two-dimensional infinite subset of $\mathbb{R}^3$ that can be described in one of three ways: The *general equation*: $$ P: \left\{ \, Ax+By+Cz=D \, \right\} $$ The *parametric equation*: $$ P: \{ p_{\textrm{o}}+s\,\vec{v} + t\,\vec{w}, \ \forall s,t \in \mathbb{R} \}, $$ which defines a plane that that contains the point $p_{\textrm{o}}$ and the vectors $\vec{v}$ and $\vec{w}$. Or the *geometric equation*: $$ P: \left\{ \vec{n} \cdot [ (x,y,z) - p_{\textrm{o}} ] = 0 \,\right\}, $$ which defines a plane that contains point $p_{\textrm{o}}$ and has normal vector $\hat{n}$. ''' ''' ''' # plot plane 2x + 1y + 1z = 5 normal = Vector([2, 1, 1]) D = 5 plot_plane(normal, D) ''' ''' ''' ''' ''' ''' ''' ## Projection operation ''' ''' ''' ''' A projection of the vector $\vec{v}$ in the direction $\vec{d}$ is denoted $\Pi_{\vec{d}}(\vec{v})$. The formula for computing the projections uses the dot product operation: $$ \Pi_{\vec{d}}(\vec{v}) \ \equiv \ (\vec{v} \cdot \hat{d}) \hat{d} \ = \ \left(\vec{v} \cdot \frac{\vec{d}}{\|\vec{d}\|} \right) \frac{\vec{d}}{\|\vec{d}\|}. $$ ''' ''' ''' def proj(v, d): """Computes the projection of vector `v` onto direction `d`.""" return v.dot( d/d.norm() )*( d/d.norm() ) ''' ''' v = Vector([2,2]) d = Vector([3,0]) proj_v_on_d = proj(v,d) plot_vecs(d, v, proj_v_on_d) autoscale_arrows() ''' ''' ''' The basic projection operation can be used to compute projection onto planes, and compute distances between geomteirc objects (page 192). ''' ''' ''' ''' ## Bases and coordinate projections ''' ''' ''' ''' See [page 225](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=68) in v2.2 of the book: - Different types of bases - Orthonormal - Orthogonal - Generic - Change of basis operation ''' ''' ''' ''' ## Vector spaces ''' ''' ''' ''' See **page 231** in v2.2 of the book. ''' ''' ''' ''' ## Vector space techniques ''' ''' ''' ''' See **page 244** in the book. ''' ''' ''' ''' ''' ''' ''' ''' # 5/ Linear transformations ''' ''' ''' ''' See [page 257](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=70) in v2.2 of the book. ''' ''' ''' ''' ## Vector functions ''' ''' ''' ''' Functions that take vectors as inputs and produce vectors as outputs: $$ T:\mathbb{R}^{n}\! \to \mathbb{R}^{m} $$ ''' ''' ''' ''' ## Matrix representation of linear transformations ''' ''' ''' ''' $$ T : \mathbb{R}^n \to \mathbb{R}^m \qquad \Leftrightarrow \qquad M_T \in \mathbb{R}^{m \times n} $$ ''' ''' ''' ''' ''' ''' ''' ''' ## Input and output spaces ''' ''' ''' ''' We can understand the properties of linear transformations $T$, and their matrix representations $M_T$ in analogy with ordinary functions: \begin{align*} \textrm{function } f:\mathbb{R}\to \mathbb{R} & \ \Leftrightarrow \, \begin{array}{l} \textrm{linear transformation } T:\mathbb{R}^{n}\! \to \mathbb{R}^{m} \\ \textrm{represented by the matrix } M_T \in \mathbb{R}^{m \times n} \end{array} \\ % \textrm{input } x\in \mathbb{R} & \ \Leftrightarrow \ \textrm{input } \vec{x} \in \mathbb{R}^n \\ %\textrm{compute } \textrm{output } f(x) \in \mathbb{R} & \ \Leftrightarrow \ % \textrm{compute matrix-vector product } \textrm{output } T(\vec{x}) \equiv M_T\vec{x} \in \mathbb{R}^m \\ %\textrm{function composition } g\circ\! f \: (x) = g(f(x)) & \ \Leftrightarrow \ % \textrm{matrix product } S(T(\vec{x})) \equiv M_SM_T \vec{x} \\ \textrm{function inverse } f^{-1} & \ \Leftrightarrow \ \textrm{matrix inverse } M_T^{-1} \\ \textrm{zeros of } f & \ \Leftrightarrow \ \textrm{kernel of } T \equiv \textrm{null space of } M_T \equiv \mathcal{N}(A) \\ \textrm{image of } f & \ \Leftrightarrow \ \begin{array}{l} \textrm{image of } T \equiv \textrm{column space of } M_T \equiv \mathcal{C}(A) \end{array} \end{align*} Observe we refer to the linear transformation $T$ and its matrix representation $M_T$ interchangeably. ''' ''' ''' ''' ## Finding matrix representations ''' ''' ''' ''' See [page 269](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=74) in v2.2 of the book. ''' ''' ''' ''' ''' ''' ''' ''' ## Invertible matrix theorem See [page 288](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=78) in the book. ''' ''' ''' ''' ''' ''' ''' ''' # 6/ Theoretical linear algebra ''' ''' ''' ''' ## Eigenvalues and eigenvectors An eigenvector of the matirx $A$ is a special input vector, for which the matrix $A$ acts as a scaling: $$ A\vec{e}_\lambda = \lambda\vec{e}_\lambda, $$ where $\lambda$ is called the *eigenvalue* and $\vec{e}_\lambda$ is the corresponding eigenvector. ''' ''' ''' A = Matrix([[1, 5], [5, 1]]) A ''' ''' A*Vector([1,0]) ''' ''' A*Vector([1,1]) ''' ''' ''' The *characterisitic polynomial* of the matrix $A$ is defined as $$ p(\lambda) \equiv \det(A-\lambda \mathbb{1}). $$ ''' ''' ''' l = symbols('lambda') (A-l*eye(2)).det() ''' ''' # the roots of the characteristic polynomial are the eigenvalues of A solve( (A-l*eye(2)).det(), l) ''' ''' # or call `eigenvals` method A.eigenvals() ''' ''' A.eigenvects() # can also find eigenvects using (A-6*eye(2)).nullspace() and (A+4*eye(2)).nullspace() ''' ''' Q, Lambda = A.diagonalize() Q, Lambda ''' ''' Q*Lambda*Q.inv() # == eigendecomposition of A ''' ''' ''' ''' ''' ''' ''' ## Special types of matrices ''' ''' ''' ''' See [page 312](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=83) in v2.2 of the book. ''' ''' ''' ''' ''' ''' ''' ''' ## Abstract vectors paces ''' ''' ''' ''' Generalize vector techniques to other vector like quantities. Allow us to talk about basis, dimention, etc. See [page 318](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=84) in the book. ''' ''' ''' ''' ''' ''' ''' ''' ## Abstract inner product spaces ''' ''' ''' ''' Use geometrical notions like length and orthogonaloty for abstract vectors. See **page 322** in the book. ''' ''' ''' ''' ''' ''' ''' ''' ## Gram–Schmidt orthogonalization ''' ''' ''' ''' See **page 328**. ''' ''' ''' ''' ''' ''' ''' ''' ## Matrix decompositions ''' ''' ''' ''' See **page 332**. ''' ''' ''' ''' ''' ''' ''' ''' ## Linear algebra with complex numbers ''' ''' ''' ''' See [page 339](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=88) in v2.2 of the book. ''' ''' ''' ''' ''' ''' ''' ''' # Applications chapters ''' ''' ''' ''' - Chapter 7: Applications - Chapter 8: Probability theory - Chapter 9: Quantum mechanics ''' ''' ''' ''' ''' ''' ''' ''' # Notation appendix ''' ''' ''' ''' Check out [page 571](https://minireference.com/static/excerpts/noBSLA_v2_preview.pdf#page=142) in the book. ''' ''' ''' ''' ''' ''' ''' ```
[ { "content": "```python\n# -*- coding: utf-8 -*-\n# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2....
[ { "content": "<|memory_start|>```python\n# -*- coding: utf-8 -*-\n# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/lic...
```python # -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import proto # type: ignore from google.ads.googleads.v6.common.types import simulation from google.ads.googleads.v6.enums.types import simulation_modification_method from google.ads.googleads.v6.enums.types import simulation_type __protobuf__ = proto.module( package='google.ads.googleads.v6.resources', marshal='google.ads.googleads.v6', manifest={ 'CampaignCriterionSimulation', }, ) class CampaignCriterionSimulation(proto.Message): r"""A campaign criterion simulation. Supported combinations of advertising channel type, criterion ids, simulation type and simulation modification method is detailed below respectively. 1. SEARCH - 30000,30001,30002 - BID_MODIFIER - UNIFORM 2. SHOPPING - 30000,30001,30002 - BID_MODIFIER - UNIFORM 3. DISPLAY - 30001 - BID_MODIFIER - UNIFORM Attributes: resource_name (str): Output only. The resource name of the campaign criterion simulation. Campaign criterion simulation resource names have the form: ``customers/{customer_id}/campaignCriterionSimulations/{campaign_id}~{criterion_id}~{type}~{modification_method}~{start_date}~{end_date}`` campaign_id (int): Output only. Campaign ID of the simulation. criterion_id (int): Output only. Criterion ID of the simulation. type_ (google.ads.googleads.v6.enums.types.SimulationTypeEnum.SimulationType): Output only. The field that the simulation modifies. modification_method (google.ads.googleads.v6.enums.types.SimulationModificationMethodEnum.SimulationModificationMethod): Output only. How the simulation modifies the field. start_date (str): Output only. First day on which the simulation is based, in YYYY-MM-DD format. end_date (str): Output only. Last day on which the simulation is based, in YYYY-MM-DD format. bid_modifier_point_list (google.ads.googleads.v6.common.types.BidModifierSimulationPointList): Output only. Simulation points if the simulation type is BID_MODIFIER. """ resource_name = proto.Field( proto.STRING, number=1, ) campaign_id = proto.Field( proto.INT64, number=9, optional=True, ) criterion_id = proto.Field( proto.INT64, number=10, optional=True, ) type_ = proto.Field( proto.ENUM, number=4, enum=simulation_type.SimulationTypeEnum.SimulationType, ) modification_method = proto.Field( proto.ENUM, number=5, enum=simulation_modification_method.SimulationModificationMethodEnum.SimulationModificationMethod, ) start_date = proto.Field( proto.STRING, number=11, optional=True, ) end_date = proto.Field( proto.STRING, number=12, optional=True, ) bid_modifier_point_list = proto.Field( proto.MESSAGE, number=8, oneof='point_list', message=simulation.BidModifierSimulationPointList, ) __all__ = tuple(sorted(__protobuf__.manifest)) ```
[ { "content": "Write the code verbatim:\n```python\n#!/usr/bin/env python\n#\n# Copyright 2009 Facebook\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apach...
[ { "content": "Write the code verbatim:\n<|memory_start|>```python\n#!/usr/bin/env python\n#\n# Copyright 2009 Facebook\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# ...
```python #!/usr/bin/env python # # Copyright 2009 Facebook # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """An I/O event loop for non-blocking sockets. Typical applications will use a single `IOLoop` object, in the `IOLoop.instance` singleton. The `IOLoop.start` method should usually be called at the end of the ``main()`` function. Atypical applications may use more than one `IOLoop`, such as one `IOLoop` per thread, or per `unittest` case. In addition to I/O events, the `IOLoop` can also schedule time-based events. `IOLoop.add_timeout` is a non-blocking alternative to `time.sleep`. """ from __future__ import absolute_import, division, print_function import collections import datetime import errno import functools import heapq import itertools import logging import numbers import os import select import sys import threading import time import traceback import math from tornado.concurrent import TracebackFuture, is_future from tornado.log import app_log, gen_log from tornado.platform.auto import set_close_exec, Waker from tornado import stack_context from tornado.util import PY3, Configurable, errno_from_exception, timedelta_to_seconds try: import signal except ImportError: signal = None if PY3: import _thread as thread else: import thread try: import asyncio except ImportError: asyncio = None _POLL_TIMEOUT = 3600.0 class TimeoutError(Exception): pass class IOLoop(Configurable): """A level-triggered I/O loop. We use ``epoll`` (Linux) or ``kqueue`` (BSD and Mac OS X) if they are available, or else we fall back on select(). If you are implementing a system that needs to handle thousands of simultaneous connections, you should use a system that supports either ``epoll`` or ``kqueue``. Example usage for a simple TCP server: .. testcode:: import errno import functools import tornado.ioloop import socket def connection_ready(sock, fd, events): while True: try: connection, address = sock.accept() except socket.error as e: if e.args[0] not in (errno.EWOULDBLOCK, errno.EAGAIN): raise return connection.setblocking(0) handle_connection(connection, address) if __name__ == '__main__': sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM, 0) sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) sock.setblocking(0) sock.bind(("", port)) sock.listen(128) io_loop = tornado.ioloop.IOLoop.current() callback = functools.partial(connection_ready, sock) io_loop.add_handler(sock.fileno(), callback, io_loop.READ) io_loop.start() .. testoutput:: :hide: By default, a newly-constructed `IOLoop` becomes the thread's current `IOLoop`, unless there already is a current `IOLoop`. This behavior can be controlled with the ``make_current`` argument to the `IOLoop` constructor: if ``make_current=True``, the new `IOLoop` will always try to become current and it raises an error if there is already a current instance. If ``make_current=False``, the new `IOLoop` will not try to become current. In general, an `IOLoop` cannot survive a fork or be shared across processes in any way. When multiple processes are being used, each process should create its own `IOLoop`, which also implies that any objects which depend on the `IOLoop` (such as `.AsyncHTTPClient`) must also be created in the child processes. As a guideline, anything that starts processes (including the `tornado.process` and `multiprocessing` modules) should do so as early as possible, ideally the first thing the application does after loading its configuration in ``main()``. .. versionchanged:: 4.2 Added the ``make_current`` keyword argument to the `IOLoop` constructor. """ # Constants from the epoll module _EPOLLIN = 0x001 _EPOLLPRI = 0x002 _EPOLLOUT = 0x004 _EPOLLERR = 0x008 _EPOLLHUP = 0x010 _EPOLLRDHUP = 0x2000 _EPOLLONESHOT = (1 << 30) _EPOLLET = (1 << 31) # Our events map exactly to the epoll events NONE = 0 READ = _EPOLLIN WRITE = _EPOLLOUT ERROR = _EPOLLERR | _EPOLLHUP # Global lock for creating global IOLoop instance _instance_lock = threading.Lock() _current = threading.local() @staticmethod def instance(): """Deprecated alias for `IOLoop.current()`. .. versionchanged:: 5.0 Previously, this method returned a global singleton `IOLoop`, in contrast with the per-thread `IOLoop` returned by `current()`. In nearly all cases the two were the same (when they differed, it was generally used from non-Tornado threads to communicate back to the main thread's `IOLoop`). This distinction is not present in `asyncio`, so in order to facilitate integration with that package `instance()` was changed to be an alias to `current()`. Applications using the cross-thread communications aspect of `instance()` should instead set their own global variable to point to the `IOLoop` they want to use. .. deprecated:: 5.0 """ return IOLoop.current() @staticmethod def initialized(): """Returns true if there is a current IOLoop. .. versionchanged:: 5.0 Redefined in terms of `current()` instead of `instance()`. .. deprecated:: 5.0 This method only knows about `IOLoop` objects (and not, for example, `asyncio` event loops), so it is of limited use. """ return IOLoop.current(instance=False) is not None def install(self): """Deprecated alias for `make_current()`. .. versionchanged:: 5.0 Previously, this method would set this `IOLoop` as the global singleton used by `IOLoop.instance()`. Now that `instance()` is an alias for `current()`, `install()` is an alias for `make_current()`. .. deprecated:: 5.0 """ self.make_current() @staticmethod def clear_instance(): """Deprecated alias for `clear_current()`. .. versionchanged:: 5.0 Previously, this method would clear the `IOLoop` used as the global singleton by `IOLoop.instance()`. Now that `instance()` is an alias for `current()`, `clear_instance()` is an alias for `clear_instance()`. .. deprecated:: 5.0 """ IOLoop.clear_current() @staticmethod def current(instance=True): """Returns the current thread's `IOLoop`. If an `IOLoop` is currently running or has been marked as current by `make_current`, returns that instance. If there is no current `IOLoop` and ``instance`` is true, creates one. .. versionchanged:: 4.1 Added ``instance`` argument to control the fallback to `IOLoop.instance()`. .. versionchanged:: 5.0 The ``instance`` argument now controls whether an `IOLoop` is created automatically when there is none, instead of whether we fall back to `IOLoop.instance()` (which is now an alias for this method) """ current = getattr(IOLoop._current, "instance", None) if current is None and instance: current = None if asyncio is not None: from tornado.platform.asyncio import AsyncIOLoop, AsyncIOMainLoop if IOLoop.configured_class() is AsyncIOLoop: current = AsyncIOMainLoop() if current is None: current = IOLoop() if IOLoop._current.instance is not current: raise RuntimeError("new IOLoop did not become current") return current def make_current(self): """Makes this the `IOLoop` for the current thread. An `IOLoop` automatically becomes current for its thread when it is started, but it is sometimes useful to call `make_current` explicitly before starting the `IOLoop`, so that code run at startup time can find the right instance. .. versionchanged:: 4.1 An `IOLoop` created while there is no current `IOLoop` will automatically become current. """ IOLoop._current.instance = self @staticmethod def clear_current(): """Clears the `IOLoop` for the current thread. Intended primarily for use by test frameworks in between tests. """ IOLoop._current.instance = None @classmethod def configurable_base(cls): return IOLoop @classmethod def configurable_default(cls): if asyncio is not None: from tornado.platform.asyncio import AsyncIOLoop return AsyncIOLoop return PollIOLoop def initialize(self, make_current=None): if make_current is None: if IOLoop.current(instance=False) is None: self.make_current() elif make_current: if IOLoop.current(instance=False) is not None: raise RuntimeError("current IOLoop already exists") self.make_current() def close(self, all_fds=False): """Closes the `IOLoop`, freeing any resources used. If ``all_fds`` is true, all file descriptors registered on the IOLoop will be closed (not just the ones created by the `IOLoop` itself). Many applications will only use a single `IOLoop` that runs for the entire lifetime of the process. In that case closing the `IOLoop` is not necessary since everything will be cleaned up when the process exits. `IOLoop.close` is provided mainly for scenarios such as unit tests, which create and destroy a large number of ``IOLoops``. An `IOLoop` must be completely stopped before it can be closed. This means that `IOLoop.stop()` must be called *and* `IOLoop.start()` must be allowed to return before attempting to call `IOLoop.close()`. Therefore the call to `close` will usually appear just after the call to `start` rather than near the call to `stop`. .. versionchanged:: 3.1 If the `IOLoop` implementation supports non-integer objects for "file descriptors", those objects will have their ``close`` method when ``all_fds`` is true. """ raise NotImplementedError() def add_handler(self, fd, handler, events): """Registers the given handler to receive the given events for ``fd``. The ``fd`` argument may either be an integer file descriptor or a file-like object with a ``fileno()`` method (and optionally a ``close()`` method, which may be called when the `IOLoop` is shut down). The ``events`` argument is a bitwise or of the constants ``IOLoop.READ``, ``IOLoop.WRITE``, and ``IOLoop.ERROR``. When an event occurs, ``handler(fd, events)`` will be run. .. versionchanged:: 4.0 Added the ability to pass file-like objects in addition to raw file descriptors. """ raise NotImplementedError() def update_handler(self, fd, events): """Changes the events we listen for ``fd``. .. versionchanged:: 4.0 Added the ability to pass file-like objects in addition to raw file descriptors. """ raise NotImplementedError() def remove_handler(self, fd): """Stop listening for events on ``fd``. .. versionchanged:: 4.0 Added the ability to pass file-like objects in addition to raw file descriptors. """ raise NotImplementedError() def set_blocking_signal_threshold(self, seconds, action): """Sends a signal if the `IOLoop` is blocked for more than ``s`` seconds. Pass ``seconds=None`` to disable. Requires Python 2.6 on a unixy platform. The action parameter is a Python signal handler. Read the documentation for the `signal` module for more information. If ``action`` is None, the process will be killed if it is blocked for too long. """ raise NotImplementedError() def set_blocking_log_threshold(self, seconds): """Logs a stack trace if the `IOLoop` is blocked for more than ``s`` seconds. Equivalent to ``set_blocking_signal_threshold(seconds, self.log_stack)`` """ self.set_blocking_signal_threshold(seconds, self.log_stack) def log_stack(self, signal, frame): """Signal handler to log the stack trace of the current thread. For use with `set_blocking_signal_threshold`. """ gen_log.warning('IOLoop blocked for %f seconds in\n%s', self._blocking_signal_threshold, ''.join(traceback.format_stack(frame))) def start(self): """Starts the I/O loop. The loop will run until one of the callbacks calls `stop()`, which will make the loop stop after the current event iteration completes. """ raise NotImplementedError() def _setup_logging(self): """The IOLoop catches and logs exceptions, so it's important that log output be visible. However, python's default behavior for non-root loggers (prior to python 3.2) is to print an unhelpful "no handlers could be found" message rather than the actual log entry, so we must explicitly configure logging if we've made it this far without anything. This method should be called from start() in subclasses. """ if not any([logging.getLogger().handlers, logging.getLogger('tornado').handlers, logging.getLogger('tornado.application').handlers]): logging.basicConfig() def stop(self): """Stop the I/O loop. If the event loop is not currently running, the next call to `start()` will return immediately. To use asynchronous methods from otherwise-synchronous code (such as unit tests), you can start and stop the event loop like this:: ioloop = IOLoop() async_method(ioloop=ioloop, callback=ioloop.stop) ioloop.start() ``ioloop.start()`` will return after ``async_method`` has run its callback, whether that callback was invoked before or after ``ioloop.start``. Note that even after `stop` has been called, the `IOLoop` is not completely stopped until `IOLoop.start` has also returned. Some work that was scheduled before the call to `stop` may still be run before the `IOLoop` shuts down. """ raise NotImplementedError() def run_sync(self, func, timeout=None): """Starts the `IOLoop`, runs the given function, and stops the loop. The function must return either a yieldable object or ``None``. If the function returns a yieldable object, the `IOLoop` will run until the yieldable is resolved (and `run_sync()` will return the yieldable's result). If it raises an exception, the `IOLoop` will stop and the exception will be re-raised to the caller. The keyword-only argument ``timeout`` may be used to set a maximum duration for the function. If the timeout expires, a `TimeoutError` is raised. This method is useful in conjunction with `tornado.gen.coroutine` to allow asynchronous calls in a ``main()`` function:: @gen.coroutine def main(): # do stuff... if __name__ == '__main__': IOLoop.current().run_sync(main) .. versionchanged:: 4.3 Returning a non-``None``, non-yieldable value is now an error. """ future_cell = [None] def run(): try: result = func() if result is not None: from tornado.gen import convert_yielded result = convert_yielded(result) except Exception: future_cell[0] = TracebackFuture() future_cell[0].set_exc_info(sys.exc_info()) else: if is_future(result): future_cell[0] = result else: future_cell[0] = TracebackFuture() future_cell[0].set_result(result) self.add_future(future_cell[0], lambda future: self.stop()) self.add_callback(run) if timeout is not None: timeout_handle = self.add_timeout(self.time() + timeout, self.stop) self.start() if timeout is not None: self.remove_timeout(timeout_handle) if not future_cell[0].done(): raise TimeoutError('Operation timed out after %s seconds' % timeout) return future_cell[0].result() def time(self): """Returns the current time according to the `IOLoop`'s clock. The return value is a floating-point number relative to an unspecified time in the past. By default, the `IOLoop`'s time function is `time.time`. However, it may be configured to use e.g. `time.monotonic` instead. Calls to `add_timeout` that pass a number instead of a `datetime.timedelta` should use this function to compute the appropriate time, so they can work no matter what time function is chosen. """ return time.time() def add_timeout(self, deadline, callback, *args, **kwargs): """Runs the ``callback`` at the time ``deadline`` from the I/O loop. Returns an opaque handle that may be passed to `remove_timeout` to cancel. ``deadline`` may be a number denoting a time (on the same scale as `IOLoop.time`, normally `time.time`), or a `datetime.timedelta` object for a deadline relative to the current time. Since Tornado 4.0, `call_later` is a more convenient alternative for the relative case since it does not require a timedelta object. Note that it is not safe to call `add_timeout` from other threads. Instead, you must use `add_callback` to transfer control to the `IOLoop`'s thread, and then call `add_timeout` from there. Subclasses of IOLoop must implement either `add_timeout` or `call_at`; the default implementations of each will call the other. `call_at` is usually easier to implement, but subclasses that wish to maintain compatibility with Tornado versions prior to 4.0 must use `add_timeout` instead. .. versionchanged:: 4.0 Now passes through ``*args`` and ``**kwargs`` to the callback. """ if isinstance(deadline, numbers.Real): return self.call_at(deadline, callback, *args, **kwargs) elif isinstance(deadline, datetime.timedelta): return self.call_at(self.time() + timedelta_to_seconds(deadline), callback, *args, **kwargs) else: raise TypeError("Unsupported deadline %r" % deadline) def call_later(self, delay, callback, *args, **kwargs): """Runs the ``callback`` after ``delay`` seconds have passed. Returns an opaque handle that may be passed to `remove_timeout` to cancel. Note that unlike the `asyncio` method of the same name, the returned object does not have a ``cancel()`` method. See `add_timeout` for comments on thread-safety and subclassing. .. versionadded:: 4.0 """ return self.call_at(self.time() + delay, callback, *args, **kwargs) def call_at(self, when, callback, *args, **kwargs): """Runs the ``callback`` at the absolute time designated by ``when``. ``when`` must be a number using the same reference point as `IOLoop.time`. Returns an opaque handle that may be passed to `remove_timeout` to cancel. Note that unlike the `asyncio` method of the same name, the returned object does not have a ``cancel()`` method. See `add_timeout` for comments on thread-safety and subclassing. .. versionadded:: 4.0 """ return self.add_timeout(when, callback, *args, **kwargs) def remove_timeout(self, timeout): """Cancels a pending timeout. The argument is a handle as returned by `add_timeout`. It is safe to call `remove_timeout` even if the callback has already been run. """ raise NotImplementedError() def add_callback(self, callback, *args, **kwargs): """Calls the given callback on the next I/O loop iteration. It is safe to call this method from any thread at any time, except from a signal handler. Note that this is the **only** method in `IOLoop` that makes this thread-safety guarantee; all other interaction with the `IOLoop` must be done from that `IOLoop`'s thread. `add_callback()` may be used to transfer control from other threads to the `IOLoop`'s thread. To add a callback from a signal handler, see `add_callback_from_signal`. """ raise NotImplementedError() def add_callback_from_signal(self, callback, *args, **kwargs): """Calls the given callback on the next I/O loop iteration. Safe for use from a Python signal handler; should not be used otherwise. Callbacks added with this method will be run without any `.stack_context`, to avoid picking up the context of the function that was interrupted by the signal. """ raise NotImplementedError() def spawn_callback(self, callback, *args, **kwargs): """Calls the given callback on the next IOLoop iteration. Unlike all other callback-related methods on IOLoop, ``spawn_callback`` does not associate the callback with its caller's ``stack_context``, so it is suitable for fire-and-forget callbacks that should not interfere with the caller. .. versionadded:: 4.0 """ with stack_context.NullContext(): self.add_callback(callback, *args, **kwargs) def add_future(self, future, callback): """Schedules a callback on the ``IOLoop`` when the given `.Future` is finished. The callback is invoked with one argument, the `.Future`. """ assert is_future(future) callback = stack_context.wrap(callback) future.add_done_callback( lambda future: self.add_callback(callback, future)) def _run_callback(self, callback): """Runs a callback with error handling. For use in subclasses. """ try: ret = callback() if ret is not None: from tornado import gen # Functions that return Futures typically swallow all # exceptions and store them in the Future. If a Future # makes it out to the IOLoop, ensure its exception (if any) # gets logged too. try: ret = gen.convert_yielded(ret) except gen.BadYieldError: # It's not unusual for add_callback to be used with # methods returning a non-None and non-yieldable # result, which should just be ignored. pass else: self.add_future(ret, self._discard_future_result) except Exception: self.handle_callback_exception(callback) def _discard_future_result(self, future): """Avoid unhandled-exception warnings from spawned coroutines.""" future.result() def handle_callback_exception(self, callback): """This method is called whenever a callback run by the `IOLoop` throws an exception. By default simply logs the exception as an error. Subclasses may override this method to customize reporting of exceptions. The exception itself is not passed explicitly, but is available in `sys.exc_info`. """ app_log.error("Exception in callback %r", callback, exc_info=True) def split_fd(self, fd): """Returns an (fd, obj) pair from an ``fd`` parameter. We accept both raw file descriptors and file-like objects as input to `add_handler` and related methods. When a file-like object is passed, we must retain the object itself so we can close it correctly when the `IOLoop` shuts down, but the poller interfaces favor file descriptors (they will accept file-like objects and call ``fileno()`` for you, but they always return the descriptor itself). This method is provided for use by `IOLoop` subclasses and should not generally be used by application code. .. versionadded:: 4.0 """ try: return fd.fileno(), fd except AttributeError: return fd, fd def close_fd(self, fd): """Utility method to close an ``fd``. If ``fd`` is a file-like object, we close it directly; otherwise we use `os.close`. This method is provided for use by `IOLoop` subclasses (in implementations of ``IOLoop.close(all_fds=True)`` and should not generally be used by application code. .. versionadded:: 4.0 """ try: try: fd.close() except AttributeError: os.close(fd) except OSError: pass class PollIOLoop(IOLoop): """Base class for IOLoops built around a select-like function. For concrete implementations, see `tornado.platform.epoll.EPollIOLoop` (Linux), `tornado.platform.kqueue.KQueueIOLoop` (BSD and Mac), or `tornado.platform.select.SelectIOLoop` (all platforms). """ def initialize(self, impl, time_func=None, **kwargs): super(PollIOLoop, self).initialize(**kwargs) self._impl = impl if hasattr(self._impl, 'fileno'): set_close_exec(self._impl.fileno()) self.time_func = time_func or time.time self._handlers = {} self._events = {} self._callbacks = collections.deque() self._timeouts = [] self._cancellations = 0 self._running = False self._stopped = False self._closing = False self._thread_ident = None self._pid = os.getpid() self._blocking_signal_threshold = None self._timeout_counter = itertools.count() # Create a pipe that we send bogus data to when we want to wake # the I/O loop when it is idle self._waker = Waker() self.add_handler(self._waker.fileno(), lambda fd, events: self._waker.consume(), self.READ) @classmethod def configurable_base(cls): return PollIOLoop @classmethod def configurable_default(cls): if hasattr(select, "epoll"): from tornado.platform.epoll import EPollIOLoop return EPollIOLoop if hasattr(select, "kqueue"): # Python 2.6+ on BSD or Mac from tornado.platform.kqueue import KQueueIOLoop return KQueueIOLoop from tornado.platform.select import SelectIOLoop return SelectIOLoop def close(self, all_fds=False): self._closing = True self.remove_handler(self._waker.fileno()) if all_fds: for fd, handler in list(self._handlers.values()): self.close_fd(fd) self._waker.close() self._impl.close() self._callbacks = None self._timeouts = None def add_handler(self, fd, handler, events): fd, obj = self.split_fd(fd) self._handlers[fd] = (obj, stack_context.wrap(handler)) self._impl.register(fd, events | self.ERROR) def update_handler(self, fd, events): fd, obj = self.split_fd(fd) self._impl.modify(fd, events | self.ERROR) def remove_handler(self, fd): fd, obj = self.split_fd(fd) self._handlers.pop(fd, None) self._events.pop(fd, None) try: self._impl.unregister(fd) except Exception: gen_log.debug("Error deleting fd from IOLoop", exc_info=True) def set_blocking_signal_threshold(self, seconds, action): if not hasattr(signal, "setitimer"): gen_log.error("set_blocking_signal_threshold requires a signal module " "with the setitimer method") return self._blocking_signal_threshold = seconds if seconds is not None: signal.signal(signal.SIGALRM, action if action is not None else signal.SIG_DFL) def start(self): if self._running: raise RuntimeError("IOLoop is already running") if os.getpid() != self._pid: raise RuntimeError("Cannot share PollIOLoops across processes") self._setup_logging() if self._stopped: self._stopped = False return old_current = getattr(IOLoop._current, "instance", None) IOLoop._current.instance = self self._thread_ident = thread.get_ident() self._running = True # signal.set_wakeup_fd closes a race condition in event loops: # a signal may arrive at the beginning of select/poll/etc # before it goes into its interruptible sleep, so the signal # will be consumed without waking the select. The solution is # for the (C, synchronous) signal handler to write to a pipe, # which will then be seen by select. # # In python's signal handling semantics, this only matters on the # main thread (fortunately, set_wakeup_fd only works on the main # thread and will raise a ValueError otherwise). # # If.0.00to # disturb it. This is an issue for twisted, which does its # SIGCHLD processing in response to its own wakeup fd being # written to. As long as the wakeup fd is registered on the IOLoop, # the loop will still wake up and everything should work. old_wakeup_fd = None if hasattr(signal, 'set_wakeup_fd') and os.name == 'posix': # requires python 2.6+, unix. set_wakeup_fd exists but crashes # the python process on windows. try: old_wakeup_fd = signal.set_wakeup_fd(self._waker.write_fileno()) if old_wakeup_fd != -1: # Already set, restore previous value. This is a little racy, # but there's no clean get_wakeup_fd and in real use the # IOLoop is just started once at the beginning. signal.set_wakeup_fd(old_wakeup_fd) old_wakeup_fd = None except ValueError: # Non-main thread, or the previous value of wakeup_fd # is no longer valid. old_wakeup_fd = None try: while True: # Prevent IO event starvation by delaying new callbacks # to the next iteration of the event loop. ncallbacks = len(self._callbacks) # Add any timeouts that have come due to the callback list. # Do not run anything until we have determined which ones # are ready, so timeouts that call add_timeout cannot # schedule anything in this iteration. due_timeouts = [] if self._timeouts: now = self.time() while self._timeouts: if self._timeouts[0].callback is None: # The timeout was cancelled. Note that the # cancellation check is repeated below for timeouts # that are cancelled by another timeout or callback. heapq.heappop(self._timeouts) self._cancellations -= 1 elif self._timeouts[0].deadline <= now: due_timeouts.append(heapq.heappop(self._timeouts)) else: break if (self._cancellations > 512 and self._cancellations > (len(self._timeouts) >> 1)): # Clean up the timeout queue when it gets large and it's # more than half cancellations. self._cancellations = 0 self._timeouts = [x for x in self._timeouts if x.callback is not None] heapq.heapify(self._timeouts) for i in range(ncallbacks): self._run_callback(self._callbacks.popleft()) for timeout in due_timeouts: if timeout.callback is not None: self._run_callback(timeout.callback) # Closures may be holding on to a lot of memory, so allow # them to be freed before we go into our poll wait. due_timeouts = timeout = None if self._callbacks: # If any callbacks or timeouts called add_callback, # we don't want to wait in poll() before we run them. poll_timeout = 0.0 elif self._timeouts: # If there are any timeouts, schedule the first one. # Use self.time() instead of 'now' to account for time # spent running callbacks. poll_timeout = self._timeouts[0].deadline - self.time() poll_timeout = max(0, min(poll_timeout, _POLL_TIMEOUT)) else: # No timeouts and no callbacks, so use the default. poll_timeout = _POLL_TIMEOUT if not self._running: break if self._blocking_signal_threshold is not None: # clear alarm so it doesn't fire while poll is waiting for # events. signal.setitimer(signal.ITIMER_REAL, 0, 0) try: event_pairs = self._impl.poll(poll_timeout) except Exception as e: # Depending on python version and IOLoop implementation, # different exception types may be thrown and there are # two ways EINTR might be signaled: # * e.errno == errno.EINTR # * e.args is like (errno.EINTR, 'Interrupted system call') if errno_from_exception(e) == errno.EINTR: continue else: raise if self._blocking_signal_threshold is not None: signal.setitimer(signal.ITIMER_REAL, self._blocking_signal_threshold, 0) # Pop one fd at a time from the set of pending fds and run # its handler. Since that handler may perform actions on # other file descriptors, there may be reentrant calls to # this IOLoop that modify self._events self._events.update(event_pairs) while self._events: fd, events = self._events.popitem() try: fd_obj, handler_func = self._handlers[fd] handler_func(fd_obj, events) except (OSError, IOError) as e: if errno_from_exception(e) == errno.EPIPE: # Happens when the client closes the connection pass else: self.handle_callback_exception(self._handlers.get(fd)) except Exception: self.handle_callback_exception(self._handlers.get(fd)) fd_obj = handler_func = None finally: # reset the stopped flag so another start/stop pair can be issued self._stopped = False if self._blocking_signal_threshold is not None: signal.setitimer(signal.ITIMER_REAL, 0, 0) IOLoop._current.instance = old_current if old_wakeup_fd is not None: signal.set_wakeup_fd(old_wakeup_fd) def stop(self): self._running = False self._stopped = True self._waker.wake() def time(self): return self.time_func() def call_at(self, deadline, callback, *args, **kwargs): timeout = _Timeout( deadline, functools.partial(stack_context.wrap(callback), *args, **kwargs), self) heapq.heappush(self._timeouts, timeout) return timeout def remove_timeout(self, timeout): # Removing from a heap is complicated, so just leave the defunct # timeout object in the queue (see discussion in # http://docs.python.org/library/heapq.html). # If this turns out to be a problem, we could add a garbage # collection pass whenever there are too many dead timeouts. timeout.callback = None self._cancellations += 1 def add_callback(self, callback, *args, **kwargs): if self._closing: return # Blindly insert into self._callbacks. This is safe even # from signal handlers because deque.append is atomic. self._callbacks.append(functools.partial( stack_context.wrap(callback), *args, **kwargs)) if thread.get_ident() != self._thread_ident: # This will write one byte but Waker.consume() reads many # at once, so it's ok to write even when not strictly # necessary. self._waker.wake() else: # If we're on the IOLoop's thread, we don't need to wake anyone. pass def add_callback_from_signal(self, callback, *args, **kwargs): with stack_context.NullContext(): self.add_callback(callback, *args, **kwargs) class _Timeout(object): """An IOLoop timeout, a UNIX timestamp and a callback""" # Reduce memory overhead when there are lots of pending callbacks __slots__ = ['deadline', 'callback', 'tdeadline'] def __init__(self, deadline, callback, io_loop): if not isinstance(deadline, numbers.Real): raise TypeError("Unsupported deadline %r" % deadline) self.deadline = deadline self.callback = callback self.tdeadline = (deadline, next(io_loop._timeout_counter)) # Comparison methods to sort by deadline, with object id as a tiebreaker # to guarantee a consistent ordering. The heapq module uses __le__ # in python2.5, and __lt__ in 2.6+ (sort() and most other comparisons # use __lt__). def __lt__(self, other): return self.tdeadline < other.tdeadline def __le__(self, other): return self.tdeadline <= other.tdeadline class PeriodicCallback(object): """Schedules the given callback to be called periodically. The callback is called every ``callback_time`` milliseconds. Note that the timeout is given in milliseconds, while most other time-related functions in Tornado use seconds. If the callback runs for longer than ``callback_time`` milliseconds, subsequent invocations will be skipped to get back on schedule. `start` must be called after the `PeriodicCallback` is created. .. versionchanged:: 5.0 The ``io_loop`` argument (deprecated since version 4.1) has been removed. """ def __init__(self, callback, callback_time): self.callback = callback if callback_time <= 0: raise ValueError("Periodic callback must have a positive callback_time") self.callback_time = callback_time self.io_loop = IOLoop.current() self._running = False self._timeout = None def start(self): """Starts the timer.""" self._running = True self._next_timeout = self.io_loop.time() self._schedule_next() def stop(self): """Stops the timer.""" self._running = False if self._timeout is not None: self.io_loop.remove_timeout(self._timeout) self._timeout = None def is_running(self): """Return True if this `.PeriodicCallback` has been started. .. versionadded:: 4.1 """ return self._running def _run(self): if not self._running: return try: return self.callback() except Exception: self.io_loop.handle_callback_exception(self.callback) finally: self._schedule_next() def _schedule_next(self): if self._running: current_time = self.io_loop.time() if self._next_timeout <= current_time: callback_time_sec = self.callback_time / 1000.0 self._next_timeout += (math.floor((current_time - self._next_timeout) / callback_time_sec) + 1) * callback_time_sec self._timeout = self.io_loop.add_timeout(self._next_timeout, self._run) ```
[ { "content": "Here is the script:\n```python\n# -*- coding: utf-8 -*-\n#\n# This file is part of Glances.\n#\n# Copyright (C) 2015 Nicolargo <nicolas@nicolargo.com>\n#\n# Glances is free software; you can redistribute it and/or modify\n# it under the terms of the GNU Lesser General Public License as published b...
[ { "content": "Here is the script:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n#\n# This file is part of Glances.\n#\n# Copyright (C) 2015 Nicolargo <nicolas@nicolargo.com>\n#\n# Glances is free software; you can redistribute it and/or modify\n# it under the terms of the GNU Lesser General Public Licens...
```python # -*- coding: utf-8 -*- # # This file is part of Glances. # # Copyright (C) 2015 Nicolargo <nicolas@nicolargo.com> # # Glances is free software; you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Glances is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. """Manage the Glances client.""" # Import system libs import json import socket import sys try: from xmlrpc.client import Transport, ServerProxy, ProtocolError, Fault except ImportError: # Python 2 from xmlrpclib import Transport, ServerProxy, ProtocolError, Fault try: import http.client as httplib except ImportError: # Python 2 import httplib # Import Glances libs from glances.core.glances_globals import version from glances.core.glances_logging import logger from glances.core.glances_stats import GlancesStatsClient from glances.outputs.glances_curses import GlancesCursesClient class GlancesClientTransport(Transport): """This class overwrite the default XML-RPC transport and manage timeout""" def set_timeout(self, timeout): self.timeout = timeout class GlancesClient(object): """This class creates and manages the TCP client.""" def __init__(self, config=None, args=None, timeout=7, return_to_browser=False): # Store the arg/config self.args = args self.config = config # Client mode: self.set_mode() # Return to browser or exit self.return_to_browser = return_to_browser # Build the URI if args.password != "": uri = 'http://{0}:{1}@{2}:{3}'.format(args.username, args.password, args.client, args.port) else: uri = 'http://{0}:{1}'.format(args.client, args.port) logger.debug("Try to connect to {0}".format(uri)) # Try to connect to the URI transport = GlancesClientTransport() # Configure the server timeout transport.set_timeout(timeout) try: self.client = ServerProxy(uri, transport=transport) except Exception as e: self.log_and_exit("Client couldn't create socket {0}: {1}".format(uri, e)) def log_and_exit(self, msg=''): """Log and (exit)""" if not self.return_to_browser: logger.critical(msg) sys.exit(2) else: logger.error(msg) def set_mode(self, mode='glances'): """Set the client mode. - 'glances' = Glances server (default) - 'snmp' = SNMP (fallback) """ self.mode = mode return self.mode def get_mode(self): """Get the client mode. - 'glances' = Glances server (default) - 'snmp' = SNMP (fallback) """ return self.mode def login(self): """Logon to the server.""" ret = True if not self.args.snmp_force: # First of all, trying to connect to a Glances server self.set_mode('glances') client_version = None try: client_version = self.client.init() except socket.error as err: # Fallback to SNMP logger.error("Connection to Glances server failed (%s)" % err) self.set_mode('snmp') fallbackmsg = _("Trying fallback to SNMP...") if not self.return_to_browser: print(fallbackmsg) else: logger.info(fallbackmsg) except ProtocolError as err: # Others errors if str(err).find(" 401 ") > 0: msg = "Connection to server failed (bad password)" else: msg = "Connection to server failed ({0})".format(err) self.log_and_exit(msg) return False if self.get_mode() == 'glances' and version.split('.')[0] == client_version.split('.')[0]: # Init stats self.stats = GlancesStatsClient(config=self.config, args=self.args) self.stats.set_plugins(json.loads(self.client.getAllPlugins())) logger.debug( "Client version: %s / Server version: %s" % (version, client_version)) elif self.get_mode() == 'glances': self.log_and_exit("Client and server not compatible: Client version: %s / Server version: %s" % (version, client_version)) return False else: self.set_mode('snmp') if self.get_mode() == 'snmp': logger.info("Trying to grab stats by SNMP...") # Fallback to SNMP if needed from glances.core.glances_stats import GlancesStatsClientSNMP # Init stats self.stats = GlancesStatsClientSNMP(config=self.config, args=self.args) if not self.stats.check_snmp(): self.log_and_exit("Connection to SNMP server failed") return False if ret: # Load limits from the configuration file # Each client can choose its owns limits self.stats.load_limits(self.config) # Init screen self.screen = GlancesCursesClient(args=self.args) # Return result return ret def update(self): """Update stats from Glances/SNMP server.""" if self.get_mode() == 'glances': return self.update_glances() elif self.get_mode() == 'snmp': return self.update_snmp() else: self.end() logger.critical("Unknown server mode: {0}".format(self.get_mode())) sys.exit(2) def update_glances(self): """Get stats from Glances server. Return the client/server connection status: - Connected: Connection OK - Disconnected: Connection NOK """ # Update the stats try: server_stats = json.loads(self.client.getAll()) server_stats['monitor'] = json.loads(self.client.getAllMonitored()) except socket.error: # Client cannot get server stats return "Disconnected" except Fault: # Client cannot get server stats (issue #375) return "Disconnected" else: # Put it in the internal dict self.stats.update(server_stats) return "Connected" def update_snmp(self): """Get stats from SNMP server. Return the client/server connection status: - SNMP: Connection with SNMP server OK - Disconnected: Connection NOK """ # Update the stats try: self.stats.update() except Exception: # Client cannot get SNMP server stats return "Disconnected" else: # Grab success return "SNMP" def serve_forever(self): """Main client loop.""" exitkey = False while True and not exitkey: # Update the stats cs_status = self.update() # Update the screen exitkey = self.screen.update(self.stats, cs_status=cs_status, return_to_browser=self.return_to_browser) # Export stats using export modules self.stats.export(self.stats) return self.get_mode() def end(self): """End of the client session.""" self.screen.end() ```
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n```python\n# Script name : tune_scale.py\n# Semantics version: semantics-4\n# Description : Tento skript umoznuje ladeni scaling faktoru\n# SCALE_CONCEPT12 a SCALE_PUSHPOP. Vychozi hodnoty jsou\n# ...
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n<|memory_start|>```python\n# Script name : tune_scale.py\n# Semantics version: semantics-4\n# Description : Tento skript umoznuje ladeni scaling faktoru\n# SCALE_CONCEPT12 a SCALE_PUSHPOP. Vychozi hodnoty jsou\n# ...
```python # Script name : tune_scale.py # Semantics version: semantics-4 # Description : Tento skript umoznuje ladeni scaling faktoru # SCALE_CONCEPT12 a SCALE_PUSHPOP. Vychozi hodnoty jsou # brany ze souboru settings, vychozi rozsah je +-0.6 a krok # 0.2. Pro otestovani predat na prikazove radce retezec # "test", pro postupne zjemnovani mrizky za kazde zjemneni # pridat "refine", neni-li treba spoustet proceduru all, # predejte "noall". Skript zapise do souboru # 'tune_scale.csv' hodnoty kriterii pro dany beh. import os from svc.utils import linrange, linspace if not 'noall' in argv and 'test' not in argv: all(moveResults=False) def tune_scale(**env): eps = 1e-6 if env['SCALE_CONCEPT12'] < 1.0-eps or env['SCALE_PUSHPOP'] < 1.0-eps: logger.info("Scaling factor is less than 1.0") return 0 if 'test' not in argv: scale(env=env) res = decodeHldt() # return res['cAcc'], res['uCorr'] return res['sActAcc'], res['iF'] else: # V pripade testovani je ztratova funkce maximalni (rovna 1) v bodech # 1.83, 1.97 global SCALE_CONCEPT12, SCALE_PUSHPOP return 1 - (env['SCALE_CONCEPT12']-1.83)**2 \ - (env['SCALE_PUSHPOP']-1.97)**2 n_iters = argv.count('refine')+1 SCALE_PUSHPOP = float(env['SCALE_PUSHPOP']) SCALE_PUSHPOP_RANGE = +-0.6 SCALE_PUSHPOP_STEP = 0.2 SCALE_CONCEPT12 = float(env['SCALE_CONCEPT12']) SCALE_CONCEPT12_RANGE = +-0.6 SCALE_CONCEPT12_STEP = 0.2 for i in range(n_iters): logger.info("_" * 80) logger.info('') logger.info("Setting tuning steps:") logger.info("=====================") logger.info(" SCALE_CONCEPT12_STEP: %.2f" % SCALE_CONCEPT12_STEP) logger.info(" SCALE_PUSHPOP_STEP : %.2f" % SCALE_PUSHPOP_STEP) logger.info("_" * 80) logger.info('') logger.info('') params = { 'SCALE_PUSHPOP': linrange(SCALE_PUSHPOP, SCALE_PUSHPOP_RANGE, SCALE_PUSHPOP_STEP), 'SCALE_CONCEPT12': linrange(SCALE_CONCEPT12, SCALE_CONCEPT12_RANGE, SCALE_CONCEPT12_STEP), } params = Grid.cartezianGrid(params) value, tuned_params = params.tune(tune_scale, logger=logger) if i == 0: fn = 'tune_cued_scale.csv' else: fn = 'tune_cued_scale%d.csv' % (i+1, ) params.writeCSV(os.path.join(env['BUILD_DIR'], fn)) SCALE_CONCEPT12 = tuned_params['SCALE_CONCEPT12'] SCALE_CONCEPT12_RANGE = +-SCALE_CONCEPT12_STEP SCALE_CONCEPT12_STEP /= 2 SCALE_PUSHPOP = tuned_params['SCALE_PUSHPOP'] SCALE_PUSHPOP_RANGE = +-SCALE_PUSHPOP_STEP SCALE_PUSHPOP_STEP /= 2 env.update(tuned_params) if 'test' not in argv: scale() decodeHldt() decodeTst() moveResults() ```
[ { "content": "```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n save mail to file for ablog\n\"\"\"\nfrom __future__ import print_function # for pylint\nimport sys\nimport os\nimport os.path\nimport shutil\nimport datetime\n\nimport imaplib\nimport getpass\nimport email\nimport smtplib\n\n...
[ { "content": "<|memory_start|>```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n save mail to file for ablog\n\"\"\"\nfrom __future__ import print_function # for pylint\nimport sys\nimport os\nimport os.path\nimport shutil\nimport datetime\n\nimport imaplib\nimport getpass\nimport email\nim...
```python #!/usr/bin/env python # -*- coding: utf-8 -*- """ save mail to file for ablog """ from __future__ import print_function # for pylint import sys import os import os.path import shutil import datetime import imaplib import getpass import email import smtplib try: from email.MIMEMultipart import MIMEMultipart from email.MIMEText import MIMEText except: # for python3 from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText import myterm import myterm.log from myterm.parser import OptionParser, read_conf, find_confdir import mail2blog __version__ = mail2blog.__version__ class MyError(Exception): """ manage Error specific """ def __init__(self, value): Exception.__init__(self) self.value = value def __str__(self): return repr(self.value) def create_logger(level='DEBUG', file='log.txt', format="%(asctime)s - %(levelname)s - %(message)s"): """add logger""" # create logger logger = myterm.Logger(level=myterm.log.LEVELLOG[level]) logger.add_stream() # Add the log message handler to the logger if len(file): logger.add_rotating_file(path=file, maxBytes=5012, form=format) return logger def send_mail(server, port, login, password, mfrom, mto, subject, body="", ssl=True): """ send mail by smtp """ msg = MIMEMultipart() msg['From'] = mfrom msg['To'] = mto msg['Subject'] = subject msg.attach(MIMEText(body, 'plain')) server = smtplib.SMTP(server, port) if ssl: server.starttls() server.login(login, password) server.sendmail(mfrom, mto, msg.as_string()) server.quit() def main(): """ run mail2blog """ parser = OptionParser( version="%prog " + __version__, usage="usage: %prog [options] args") parser.description = "mail to blog: save attachement in blog path and del mail" parser.epilog = "by Frederic Aoustin" parser.add_option("-l", "--level-log", dest="level", choices=[key for key in myterm.log.LEVELLOG.keys()], help="level of log", default="INFO", type="choice") parser.add_option("-d", "--path-log", dest="logfile", help="directory of log file", default="", type="string") parser.add_option("-c", "--path-conf", dest="conf", help="path of conf.py", default="", type="string") (options, args) = parser.parse_args() try: logger = create_logger(options.level, options.logfile) if len(options.conf): if os.path.isdir(options.conf): options.conf = os.path.join(options.conf, "conf.py") if os.path.isfile(options.conf): logger.debug("read conf %s" % options.conf) conf = read_conf(os.path.dirname(options.conf), os.path.basename( os.path.splitext(options.conf)[0])) else: logger.error("%s does not a file" % options.conf) raise MyError("not found %s" % options.conf) else: dirn, name, conf = find_confdir("conf") options.conf = os.path.join(dirn, name) logger.debug("find and read conf %s" % os.path.join(dirn, name)) # manage conf by conf imap = getattr(conf, 'mail2blog_imap', 'imap.gmail.com') mailbox = getattr(conf, 'mail2blog_mailbox', None) mailboxpassword = getattr(conf, 'mail2blog_mailbox_password', None) authorized = getattr(conf, 'mail2blog_user_authorized', []) blogpath = os.path.dirname(options.conf) bckpath = getattr(conf, 'mail2blog_bck_path', None) forkpath = getattr(conf, 'mail2blog_fork_path', {}) smtp = getattr(conf, 'mail2blog_smtp', 'smtp.gmail.com') smtpport = getattr(conf, 'mail2blog_smtp_port', 587) smtpnotssl = getattr(conf, 'mail2blog_smtp_not_ssl', False) smtplogin = getattr(conf, 'mail2blog_smtp_login', None) smtppassword = getattr(conf, 'mail2blog_smtp_password', None) smtpfrom = getattr(conf, 'mail2blog_smtp_from', None) build = getattr(conf, 'mail2blog_build', False) if not len(mailbox) or not len(imap): logger.error("mailbox, imap are mandatory ") sys.exit(1) if not len(mailboxpassword): mailboxpassword = getpass.getpass("password for mailbox:") try: runbuild = False logger.info("connect to imap server") try: mailinbox = imaplib.IMAP4_SSL(imap) logger.info(mailinbox.login(mailbox, mailboxpassword)[1]) except Exception as exp: logger.critical(exp) sys.exit(1) mailinbox.select() typ, data = mailinbox.uid('SEARCH', 'ALL') msgs = data[0].split() logger.info("Found {0} msgs".format(len(msgs))) for uid in msgs: typ, content = mailinbox.uid('FETCH', uid, '(RFC822)') mail = email.message_from_string(content[0][1]) logger.info("From: {0}, Subject: {1}, Date: {2}\n".format( mail["From"], mail["Subject"], mail["Date"])) mailfrom = mail["From"] if '<' in mailfrom: mailfrom = mailfrom.split('<')[-1].split('>')[0] if mailfrom in authorized: logger.debug("From %s authorized" % mailfrom) if mail.is_multipart(): for part in mail.walk(): if part.get_filename(): ext = os.path.splitext( part.get_filename())[1].lower() logger.debug( "treat %s, extension : %s" % (part.get_filename(), ext)) if ext in forkpath.keys(): logger.debug( "save %s" % part.get_filename()) if not os.path.exists(os.path.join(blogpath, forkpath[ext])): os.makedirs( os.path.join(blogpath, forkpath[ext])) pathfile = os.path.join( blogpath, forkpath[ext], part.get_filename()) if os.path.isfile(pathfile): if len(bckpath): logger.debug( "save bkp %s" % part.get_filename()) if not os.path.exists(bckpath): os.makedirs(bckpath) qfile = os.path.join( bckpath, '%s_%s' % (datetime.datetime.now(), part.get_filename())) shutil.move(pathfile, qfile) else: os.remove(pathfile) open(pathfile, 'wb').write( part.get_payload(decode=True)) runbuild = True else: logger.debug("not save %s" % part.get_filename()) if len(smtplogin) and len(smtp): logger.info("send response by mail") send_mail(smtp, smtpport, smtplogin, smtppassword, smtpfrom, mailfrom, 'Add content in blog', "Your mail %s add in blog" % mail[ "Subject"], not smtpnotssl) else: logger.warning("From %s not authorized" % mailfrom) logger.info("delete mail") mailinbox.uid('STORE', uid, '+FLAGS', '(\Deleted)') logger.info("disconnect to imap server") mailinbox.close() mailinbox.logout() if runbuild and build: import ablog.commands ablog.commands.ablog_build() except Exception as exp: logger.critical(parser.error(exp)) raise exp except Exception as exp: print(parser.error(exp)) parser.print_help() sys.exit(1) if __name__ == '__main__': main() ```
[ { "content": "Return the code unaltered:\n```python\n# -*- coding: utf-8 -*-\n# Created by apple on 2017/1/31.\n\nimport re\nimport aiofiles\nimport plistlib\nfrom platform import system\nfrom .regex import Regex\nfrom ..log import log\nfrom zipfile import ZipFile\nfrom subprocess import Popen, PIPE\nfrom ..db ...
[ { "content": "Return the code unaltered:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n# Created by apple on 2017/1/31.\n\nimport re\nimport aiofiles\nimport plistlib\nfrom platform import system\nfrom .regex import Regex\nfrom ..log import log\nfrom zipfile import ZipFile\nfrom subprocess import Popen, ...
```python # -*- coding: utf-8 -*- # Created by apple on 2017/1/31. import re import aiofiles import plistlib from platform import system from .regex import Regex from ..log import log from zipfile import ZipFile from subprocess import Popen, PIPE from ..db import AppType aapt = './aapt_mac' if system() == 'Darwin' else './aapt_centos' class PackageParse: __apk_manifest_path = 'AndroidManifest.xml' @classmethod async def parse(cls, package_path: str): """ 解析安装包(ipa、apk) :param package_path: 安装包路径 :return: 成功解析返回 PackageParse, 失败返回 None """ if package_path.endswith('.ipa'): return await cls.__ipa_parse(package_path) elif package_path.endswith('.apk'): return await cls.__apk_parse(package_path) else: return None @staticmethod async def __apk_info(file_path: str) -> str: """ 获取apk信息 :param file_path: apk路径 :return: string """ popen = Popen('{} dump badging {}'.format(aapt, file_path), stdout=PIPE, shell=True) result = popen.communicate()[0] return result.decode() @classmethod async def __apk_parse(cls, file_path: str): """ 解析apk包 :param file_path: apk路径 :return: PackageParse """ apk_file = ZipFile(file_path) info = await cls.__apk_info(file_path) match_info = Regex.APKInfo.match(info) package_name = match_info.group(1) version_name = match_info.group(3) version_code = match_info.group(2) names = Regex.APKName.findall(info) app_name = names[0] if names else package_name icon_path = Regex.APKIcon.findall(info)[0] log.debug('app: {}, V{} build {}, package: {}'.format(app_name, version_name, version_code, package_name)) return PackageParse(apk_file, AppType.android, package_name, app_name, icon_path, version_name, version_code) @staticmethod async def __ipa_parse(file_path: str): """ 解析ipa包 :param file_path: ipa路径 :return: PackageParse """ ipa_file = ZipFile(file_path) # 解析info.plist路径 ns = [n for n in ipa_file.namelist() if Regex.IPAInfoPlistPath.match(n)] if not ns: log.warning('parse info.plist failure: {}'.format(file_path)) return plist_path = ns[-1] # 解析plist plist_data = ipa_file.read(plist_path) plist_file = plistlib.loads(plist_data) # 解析icon'CFBundleIconFiles' (4400546488) if plist_file.get('CFBundleIconFiles'): icon_name = plist_file['CFBundleIconFiles'][-1] else: if plist_file.get('CFBundleIcons'): icon_dict = plist_file['CFBundleIcons'] elif plist_file.get('CFBundleIcons'): icon_dict = plist_file['CFBundleIcons~ipad'] else: log.warning('parse icon failure: {}'.format(file_path)) return icon_name = icon_dict['CFBundlePrimaryIcon']['CFBundleIconFiles'][-1] log.debug('parse icon name: {}'.format(icon_name)) # 获取icon路径 re_icon_name_end = '(@\dx)\.png' if not icon_name.endswith('.png') else '' re_icon_name = re.compile('([^/]+/){{2}}{}{}'.format(icon_name, re_icon_name_end)) ns = [n for n in ipa_file.namelist() if re_icon_name.match(n)] if not ns: log.warning('read icon failure: {}'.format(file_path)) return icon_path = ns[-1] log.debug('parse icon path: {}'.format(icon_path)) # 版本号 version_number = plist_file['CFBundleShortVersionString'] # build号 build_number = plist_file['CFBundleVersion'] # 包名 package_name = plist_file['CFBundleIdentifier'] # app名称 app_name = plist_file['CFBundleDisplayName'] if plist_file.get('CFBundleDisplayName') else plist_file[ 'CFBundleName'] log.debug( 'app: {}, V{} build {}, package: {}'.format(app_name, version_number, build_number, package_name)) return PackageParse(ipa_file, AppType.iOS, package_name, app_name, icon_path, version_number, build_number) def __init__(self, zip_file: ZipFile, app_type: AppType, package_name: str, app_name: str, icon_path: str, version_name: str, version_code: str): """ 安装包信息 :param zip_file: zip_file :param app_type: 安装包类型 iOS Android :param package_name: 包名 :param app_name: 应用名 :param icon_path: logo路径 :param version_name: 版本名 :param version_code: 版本号 """ self.zip_file = zip_file self.app_type = app_type self.package_name = package_name self.app_name = app_name self.icon_path = icon_path self.version_name = version_name self.version_code = version_code async def save_icon(self, save_path): async with aiofiles.open(save_path, 'wb+') as f: await f.write(self.zip_file.read(self.icon_path)) if self.app_type == AppType.iOS: dirs = save_path.split('/') if len(dirs) > 2: save_dir = '/'.join(dirs[:-1]) else: save_dir = './' popen = Popen('./pngdefry -o {} {}'.format(save_dir, save_path), stdout=PIPE, shell=True) popen.wait() ```
[ { "content": "Provide an exact copy of the source code:\n```python\n# ###################################################\n# Copyright (C) 2008-2017 The Unknown Horizons Team\n# team@unknown-horizons.org\n# This file is part of Unknown Horizons.\n#\n# Unknown Horizons is free software; you can redistribute it a...
[ { "content": "Provide an exact copy of the source code:\n<|memory_start|>```python\n# ###################################################\n# Copyright (C) 2008-2017 The Unknown Horizons Team\n# team@unknown-horizons.org\n# This file is part of Unknown Horizons.\n#\n# Unknown Horizons is free software; you can r...
```python # ################################################### # Copyright (C) 2008-2017 The Unknown Horizons Team # team@unknown-horizons.org # This file is part of Unknown Horizons. # # Unknown Horizons is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the # Free Software Foundation, Inc., # 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA # ################################################### from operator import itemgetter import horizons.globals from horizons.constants import AI, COLORS from horizons.util.color import Color from horizons.util.difficultysettings import DifficultySettings class StartGameOptions: def __init__(self, game_identifier): super().__init__() # TODO: check if this call is needed self.game_identifier = game_identifier self._player_list = None self.trader_enabled = True self.pirate_enabled = True self.natural_resource_multiplier = 1 self.disasters_enabled = True self.force_player_id = None self.is_map = False self.is_multiplayer = False self.is_scenario = False self.player_name = 'Player' self.player_color = None self.ai_players = 0 self.human_ai = AI.HUMAN_AI # this is used by the map editor to pass along the new map's size self.map_padding = None self.is_editor = False def init_new_world(self, session): # NOTE: this must be sorted before iteration, cause there is no defined order for # iterating a dict, and it must happen in the same order for mp games. for i in sorted(self._get_player_list(), key=itemgetter('id')): session.world.setup_player(i['id'], i['name'], i['color'], i['clientid'] if self.is_multiplayer else None, i['local'], i['ai'], i['difficulty']) session.world.set_forced_player(self.force_player_id) center = session.world.init_new_world(self.trader_enabled, self.pirate_enabled, self.natural_resource_multiplier) session.view.center(center[0], center[1]) def set_human_data(self, player_name, player_color): self.player_name = player_name self.player_color = player_color def _get_player_list(self): if self._player_list is not None: return self._player_list # for now just make it a bit easier for the AI difficulty_level = {False: DifficultySettings.DEFAULT_LEVEL, True: DifficultySettings.EASY_LEVEL} players = [] players.append({ 'id': 1, 'name': self.player_name, 'color': Color.get(1) if self.player_color is None else self.player_color, 'local': True, 'ai': self.human_ai, 'difficulty': difficulty_level[bool(self.human_ai)], }) cur_locale = horizons.globals.fife.get_locale() # add AI players with a distinct color; if none can be found then use black for num in range(self.ai_players): color = Color.get(COLORS.BLACK) # if none can be found then be black for possible_color in Color.get_defaults(): if possible_color == Color.get(COLORS.BLACK): continue # black is used by the trader and the pirate used = any(possible_color == player['color'] for player in players) if not used: color = possible_color break name = horizons.globals.db.get_random_ai_name(cur_locale, [p['name'] for p in players]) # out of pre-defined names? if name is None: name = 'AI' + str(num + 1) players.append({ 'id': num + 2, 'name': name, 'color': color, 'local': False, 'ai': True, 'difficulty': difficulty_level[True], }) return players @classmethod def create_start_multiplayer(cls, game_file, player_list, is_map): options = StartGameOptions(game_file) options._player_list = player_list options.is_map = is_map options.is_multiplayer = True return options @classmethod def create_start_singleplayer(cls, game_identifier, is_scenario, ai_players, trader_enabled, pirate_enabled, force_player_id, is_map): options = StartGameOptions(game_identifier) options.is_scenario = is_scenario options.ai_players = ai_players options.trader_enabled = trader_enabled options.pirate_enabled = pirate_enabled options.force_player_id = force_player_id options.is_map = is_map return options @classmethod def create_start_random_map(cls, ai_players, seed, force_player_id): from horizons.util.random_map import generate_map_from_seed options = StartGameOptions(generate_map_from_seed(seed)) options.ai_players = ai_players options.force_player_id = force_player_id options.is_map = True return options @classmethod def create_editor_load(cls, map_name): options = StartGameOptions(map_name) options.player_name = 'Editor' options.trader_enabled = False options.pirate_enabled = False options.natural_resource_multiplier = 0 options.disasters_enabled = False options.is_map = True options.is_editor = True return options @classmethod def create_start_scenario(cls, scenario_file): options = StartGameOptions(scenario_file) options.is_scenario = True return options @classmethod def create_start_map(cls, map_name): options = StartGameOptions(map_name) options.is_map = True return options @classmethod def create_load_game(cls, saved_game, force_player_id): options = StartGameOptions(saved_game) options.force_player_id = force_player_id return options @classmethod def create_game_test(cls, game_identifier, player_list): options = StartGameOptions(game_identifier) options._player_list = player_list options.trader_enabled = False options.pirate_enabled = False options.natural_resource_multiplier = 0 return options @classmethod def create_ai_test(cls, game_identifier, player_list): options = StartGameOptions(game_identifier) options._player_list = player_list options.is_map = True return options ```
[ { "content": "Output the full code verbatim (no extra comments):\n```python\n#! /usr/bin/env python3\n# coding: utf-8\n\n'''\nFonctions de manipulation et vérifications du fichier de configuration\n'''\n\nimport json\nimport utils.misc as misc\n\nCONF_FILE_NAME = \"conf/conf.json\"\n\n\n'''\ncheck_conf_valid: v...
[ { "content": "Output the full code verbatim (no extra comments):\n<|memory_start|>```python\n#! /usr/bin/env python3\n# coding: utf-8\n\n'''\nFonctions de manipulation et vérifications du fichier de configuration\n'''\n\nimport json\nimport utils.misc as misc\n\nCONF_FILE_NAME = \"conf/conf.json\"\n\n\n'''\nche...
```python #! /usr/bin/env python3 # coding: utf-8 ''' Fonctions de manipulation et vérifications du fichier de configuration ''' import json import utils.misc as misc CONF_FILE_NAME = "conf/conf.json" ''' check_conf_valid: vérifie que le fichier de conf est bien dans un format json valide entrée: pas d'argument (nom du fichier dépend de la variable globale CONF_FILE_NAME) sortie: retourne 0 si le fichier est valide, -1 sinon ''' def check_conf_valid(): try: with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) return 0 except: return -1 ''' get_list_equipment_from_conf: renvoie la liste des équipements contenus dans le fichier conf.json entrée: pas d'argument sortie: liste de tuples (nom de l'equipement, ip de l'equipement) ''' def get_list_equipment_from_conf(): with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) list_eq = [] for dat in data["EQUIPEMENTS"]: var_nom = str(data["EQUIPEMENTS"][dat]["NOM"]) var_ip = str(data["EQUIPEMENTS"][dat]["IP"]) tuple_eq = (var_nom, var_ip) list_eq.append(tuple_eq) return list_eq ''' get_list_equipment_from_conf: renvoie la liste des équipements contenus dans le fichier conf.json entrée: pas d'argument sortie: liste de tuples (nom de l'equipement, "", False) ''' def get_list_equipment_from_conf_for_checklist(): with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) list_eq = [] for dat in data["EQUIPEMENTS"]: var_nom = str(data["EQUIPEMENTS"][dat]["NOM"]) tuple_eq = (var_nom, "", False) list_eq.append(tuple_eq) return list_eq ''' get_list_files_from_conf: renvoie la liste des fichiers contenus dans le fichier conf.json entrée: pas d'argument sortie: liste de tuples (nom de l'equipement, ip de l'equipement) ''' def get_list_files_from_conf(): with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) list_fic = [] for dat in data["FICHIERS"]: var_nom = str(data["FICHIERS"][dat]["NOM"]) var_path = str(data["FICHIERS"][dat]["PATH"]) tuple_eq = (var_nom, var_path) list_fic.append(tuple_eq) return list_fic ''' delete_file_from_conf: supprime un fichier du fichier de configuration entrée: nom du fichier à supprimer sortie: 0 si OK, -1 autrement ''' def delete_file_from_conf(file_name): try: with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) for element in data["FICHIERS"]: if file_name == data["FICHIERS"][element]["NOM"]: data["FICHIERS"].pop(element) break with open(CONF_FILE_NAME, 'w') as data_file: data = json.dump(data, data_file) return 0 except: return -1 ''' delete_equipment_from_conf: supprime un équipement du fichier de configuration entrée: nom de l'équipement à supprimer sortie: 0 si OK, -1 autrement ''' def delete_equipment_from_conf(equipment_name): try: with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) for element in data["EQUIPEMENTS"]: if equipment_name == data["EQUIPEMENTS"][element]["NOM"]: data["EQUIPEMENTS"].pop(element) break with open(CONF_FILE_NAME, 'w') as data_file: data = json.dump(data, data_file) return 0 except: return -1 ''' add_file_to_conf: ajoute un fichier dans le fichier de configuration entrée: liste avec les paramètres du fichier [nom, path, type, equipement] sortie: 0 si OK, -1 si le nom existe déjà, -2 si autre erreur ''' def add_file_to_conf(list_params_file): file_name = list_params_file[0] file_path = list_params_file[1] file_type = list_params_file[2] equipment_name = list_params_file[3] try: with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) #vérification de l'unicité du nom for element in data["FICHIERS"]: if file_name == data["FICHIERS"][element]["NOM"]: return -1 #on formate les paramètres du fichier en JSON data["FICHIERS"][file_name] = {} data["FICHIERS"][file_name]["NOM"] = file_name data["FICHIERS"][file_name]["TYPE"] = file_type data["FICHIERS"][file_name]["EQUIPEMENT"] = equipment_name data["FICHIERS"][file_name]["PATH"] = file_path #On modifie le fichier de configuration with open(CONF_FILE_NAME, 'w') as data_file: data = json.dump(data, data_file) return 0 except: return -1 ''' add_equipment_to_conf: ajoute un équipement dans le fichier de configuration entrée: liste avec les paramètres de l'équipement [nom, IP, type, login, MDP] sortie: 0 si OK, -1 si le nom existe déjà, -2 si autre erreur ''' def add_equipment_to_conf(list_params_equipment): equipment_name = list_params_equipment[0] equipment_ip = list_params_equipment[1] equipment_type = list_params_equipment[2] equipment_login = list_params_equipment[3] equipment_mdp = list_params_equipment[4] try: with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) #vérification de l'unicité du nom for element in data["EQUIPEMENTS"]: if equipment_name == data["EQUIPEMENTS"][element]["NOM"]: return -1 #on formate les paramètres du fichier en JSON data["EQUIPEMENTS"][equipment_name] = {} data["EQUIPEMENTS"][equipment_name]["NOM"] = equipment_name data["EQUIPEMENTS"][equipment_name]["IP"] = equipment_ip data["EQUIPEMENTS"][equipment_name]["TYPE"] = equipment_type data["EQUIPEMENTS"][equipment_name]["LOGIN"] = equipment_login data["EQUIPEMENTS"][equipment_name]["MDP"] = equipment_mdp #On modifie le fichier de configuration with open(CONF_FILE_NAME, 'w') as data_file: data = json.dump(data, data_file) return 0 except: return -1 ''' check_list_equipment_valid: vérifie que la demande de création d'ajout d'un équipement est valide entrée: liste de paramètres concernant l'équipement [nom, IP, type, login, MDP] sortie: retourne 0 si l'équipement peut être ajouté -1 si le nom de l'équipement n'est pas unique -2 si l'IP fournie n'est pas valable -3 si l'IP n'est pas unique -4 si le type n'est pas "DB" (base de données), "S" (serveur), "R" (équipement réseau) -5 si tous les champs ne sont pas remplis ''' def check_list_equipment_valid(list_params_equipment): equipment_name = list_params_equipment[0] equipment_ip = list_params_equipment[1] equipment_type = list_params_equipment[2] equipment_login = list_params_equipment[3] equipment_mdp = list_params_equipment[4] #Vérification que tous les champs sont remplis if equipment_name == "" or equipment_ip == "" or equipment_type == "" or equipment_login == "" or equipment_mdp == "": return -5 #Ouverture du fichier de conf with open(CONF_FILE_NAME) as data_file: data = json.load(data_file) #Vérification de l'unicité du nom if equipment_name in data["EQUIPEMENTS"]: return -1 #Vérification de la validité de l'IP if misc.is_valid_ipv4_address(equipment_ip) == False: return -2 #Vérification de l'unicité de l'IP dans le fichier de conf for element in data["EQUIPEMENTS"]: if equipment_ip in data["EQUIPEMENTS"][element]["IP"]: return -3 #Vérification du type d'équipement if equipment_type != "DB" and equipment_type != "S" and equipment_type != "R": return -4 return 0 ```
[ { "content": "Reconstruct the code exactly:\n```python\n#!/usr/bin/env python\n\nimport os, sys\nimport filecmp\nimport re\nimport shutil\nholderlist=[]\n\ndef compareme(dir1, dir2):\n dircomp=filecmp.dircmp(dir1,dir2)\n only_in_one=dircomp.left_only\n diff_in_one=dircomp.diff_files\n dirpath=os.pat...
[ { "content": "Reconstruct the code exactly:\n<|memory_start|>```python\n#!/usr/bin/env python\n\nimport os, sys\nimport filecmp\nimport re\nimport shutil\nholderlist=[]\n\ndef compareme(dir1, dir2):\n dircomp=filecmp.dircmp(dir1,dir2)\n only_in_one=dircomp.left_only\n diff_in_one=dircomp.diff_files\n ...
```python #!/usr/bin/env python import os, sys import filecmp import re import shutil holderlist=[] def compareme(dir1, dir2): dircomp=filecmp.dircmp(dir1,dir2) only_in_one=dircomp.left_only diff_in_one=dircomp.diff_files dirpath=os.path.abspath(dir1) [holderlist.append(os.path.abspath( os.path.join(dir1,x) )) for x in only_in_one] [holderlist.append(os.path.abspath( os.path.join(dir1,x) )) for x in diff_in_one] if len(dircomp.common_dirs) > 0: for item in dircomp.common_dirs: compareme(os.path.abspath(os.path.join(dir1,item)), \ os.path.abspath(os.path.join(dir2,item))) return holderlist def main(): if len(sys.argv) > 2: dir1=sys.argv[1] dir2=sys.argv[2] else: print "Usage: ", sys.argv[0], "datadir backupdir" sys.exit() source_files=compareme(dir1,dir2) dir1=os.path.abspath(dir1) if not dir2.endswith('/'): dir2=dir2+'/' dir2=os.path.abspath(dir2) destination_files=[] createdir_bool=False for item in source_files: destination_dir=re.sub(dir1, dir2, item) destination_files.append(destination_dir) if os.path.isdir(item): if not os.path.exists(destination_dir): os.makedirs(destination_dir) createdir_bool=True if createdir_bool: destination_files=[] source_files=[] source_files=compareme(dir1,dir2) for item in source_files: destination_dir=re.sub(dir1, dir2, item) destination_files.append(destination_dir) print "update item:" print source_files copy_pair=zip(source_files,destination_files) for item in copy_pair: if os.path.isfile(item[0]): shutil.copyfile(item[0], item[1]) if __name__ == '__main__': main() ```
[ { "content": "Produce an exact reconstruction of the code:\n```python\n#!/usr/bin/env python\n\nimport os\nfrom argh import arg, ArghParser\nfrom argh.exceptions import CommandError\nfrom functools import wraps\n\n\nCONFIG = os.environ.get('ALFRED_CONFIG')\n\n\ndef with_app(func):\n @wraps(func)\n @arg('-...
[ { "content": "Produce an exact reconstruction of the code:\n<|memory_start|>```python\n#!/usr/bin/env python\n\nimport os\nfrom argh import arg, ArghParser\nfrom argh.exceptions import CommandError\nfrom functools import wraps\n\n\nCONFIG = os.environ.get('ALFRED_CONFIG')\n\n\ndef with_app(func):\n @wraps(fu...
```python #!/usr/bin/env python import os from argh import arg, ArghParser from argh.exceptions import CommandError from functools import wraps CONFIG = os.environ.get('ALFRED_CONFIG') def with_app(func): @wraps(func) @arg('--config', help='path to config') def wrapper(args): from alfred import create_app if not CONFIG and not args.config: raise CommandError('There is no config file specified') app = create_app(args.config or CONFIG) return func(app, args) return wrapper @arg('--host', default='127.0.0.1', help='the host') @arg('--port', default=5000, help='the port') @arg('--noreload', action='store_true', help='disable code reloader') @with_app def runserver(app, args): app.run(args.host, args.port, use_reloader=not args.noreload) @with_app def shell(app, args): from alfred.helpers import get_shell with app.test_request_context(): sh = get_shell() sh(app=app) @with_app def collectassets(app, args): from alfred.assets import gears gears.get_environment(app).save() def main(): parser = ArghParser() parser.add_commands([runserver, shell, collectassets]) parser.dispatch() if __name__ == '__main__': main() ```
[ { "content": "Provide an exact copy of the source code:\n```python\nimport logging\nfrom pyvisdk.exceptions import InvalidArgumentError\n\n########################################\n# Automatically generated, do not edit.\n########################################\n\nlog = logging.getLogger(__name__)\n\ndef HostA...
[ { "content": "Provide an exact copy of the source code:\n<|memory_start|>```python\nimport logging\nfrom pyvisdk.exceptions import InvalidArgumentError\n\n########################################\n# Automatically generated, do not edit.\n########################################\n\nlog = logging.getLogger(__name...
```python import logging from pyvisdk.exceptions import InvalidArgumentError ######################################## # Automatically generated, do not edit. ######################################## log = logging.getLogger(__name__) def HostAdminEnableEvent(vim, *args, **kwargs): '''This event records that the administrator permission has been restored.''' obj = vim.client.factory.create('ns0:HostAdminEnableEvent') # do some validation checking... if (len(args) + len(kwargs)) < 4: raise IndexError('Expected at least 5 arguments got: %d' % len(args)) required = [ 'chainId', 'createdTime', 'key', 'userName' ] optional = [ 'changeTag', 'computeResource', 'datacenter', 'ds', 'dvs', 'fullFormattedMessage', 'host', 'net', 'vm', 'dynamicProperty', 'dynamicType' ] for name, arg in zip(required+optional, args): setattr(obj, name, arg) for name, value in kwargs.items(): if name in required + optional: setattr(obj, name, value) else: raise InvalidArgumentError("Invalid argument: %s. Expected one of %s" % (name, ", ".join(required + optional))) return obj ```
[ { "content": "Here is the source code:\n```python\n#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\n Blockstack\n ~~~~~\n copyright: (c) 2014-2015 by Halfmoon Labs, Inc.\n copyright: (c) 2016 by Blockstack.org\n\n This file is part of Blockstack\n\n Blockstack is free software: you can...
[ { "content": "Here is the source code:\n<|memory_start|>```python\n#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\n Blockstack\n ~~~~~\n copyright: (c) 2014-2015 by Halfmoon Labs, Inc.\n copyright: (c) 2016 by Blockstack.org\n\n This file is part of Blockstack\n\n Blockstack is free s...
```python #!/usr/bin/env python2 # -*- coding: utf-8 -*- """ Blockstack ~~~~~ copyright: (c) 2014-2015 by Halfmoon Labs, Inc. copyright: (c) 2016 by Blockstack.org This file is part of Blockstack Blockstack is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. Blockstack is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with Blockstack. If not, see <http://www.gnu.org/licenses/>. """ # activate F-day 2017 """ TEST ENV BLOCKSTACK_EPOCH_1_END_BLOCK 682 TEST ENV BLOCKSTACK_EPOCH_2_END_BLOCK 683 TEST ENV BLOCKSTACK_EPOCH_2_NAMESPACE_LIFETIME_MULTIPLIER 1 TEST ENV BLOCKSTACK_EPOCH_3_NAMESPACE_LIFETIME_MULTIPLIER 1 TEST ENV BLOCKSTACK_EPOCH_3_NAMESPACE_LIFETIME_GRACE_PERIOD 0 TEST ENV BLOCKSTACK_EPOCH_3_NAMESPACE_RECEIVE_FEES_PERIOD 22 """ import testlib import virtualchain import blockstack wallets = [ testlib.Wallet( "5JesPiN68qt44Hc2nT8qmyZ1JDwHebfoh9KQ52Lazb1m1LaKNj9", 100000000000 ), testlib.Wallet( "5KHqsiU9qa77frZb6hQy9ocV7Sus9RWJcQGYYBJJBb2Efj1o77e", 100000000000 ), testlib.Wallet( "5Kg5kJbQHvk1B64rJniEmgbD83FpZpbw2RjdAZEzTefs9ihN3Bz", 100000000000 ), testlib.Wallet( "5JuVsoS9NauksSkqEjbUZxWwgGDQbMwPsEfoRBSpLpgDX1RtLX7", 100000000000 ), testlib.Wallet( "5KEpiSRr1BrT8vRD7LKGCEmudokTh1iMHbiThMQpLdwBwhDJB1T", 100000000000 ) ] consensus = "17ac43c1d8549c3181b200f1bf97eb7d" def scenario( wallets, **kw ): testlib.blockstack_namespace_preorder( "test", wallets[1].addr, wallets[0].privkey ) testlib.next_block( **kw ) testlib.blockstack_namespace_reveal( "test", wallets[1].addr, 3, 250, 4, [6,5,4,3,2,1,0,0,0,0,0,0,0,0,0,0], 10, 10, wallets[0].privkey, version_bits=2) testlib.next_block( **kw ) resp = testlib.blockstack_name_import( "foo.test", wallets[3].addr, "11" * 20, wallets[1].privkey ) if 'error' in resp: print json.dumps( resp, indent=4 ) return False testlib.next_block( **kw ) testlib.blockstack_namespace_ready( "test", wallets[1].privkey ) testlib.next_block( **kw ) namespace_rec = testlib.blockstack_cli_get_namespace_blockchain_record("test") if 'error' in namespace_rec: print namespace_rec return False namespace_balance = testlib.get_balance(namespace_rec['address']) burn_balance = testlib.get_balance(blockstack.lib.config.BLOCKSTACK_BURN_ADDRESS) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) # expired res = testlib.blockstack_name_preorder( "foo.test", wallets[2].privkey, wallets[3].addr ) # +name_cost if 'error' in res: print res return False testlib.next_block( **kw ) res = testlib.blockstack_name_register( "foo.test", wallets[2].privkey, wallets[3].addr ) if 'error' in res: print res return False testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) # expired res = testlib.blockstack_name_preorder( "foo.test", wallets[3].privkey, wallets[4].addr ) # +name_cost if 'error' in res: print res return False testlib.next_block( **kw ) res = testlib.blockstack_name_register( "foo.test", wallets[3].privkey, wallets[4].addr ) if 'error' in res: print res return False testlib.next_block( **kw ) testlib.next_block( **kw ) res = testlib.blockstack_name_renew("foo.test", wallets[4].privkey) # +name_cost if 'error' in res: print res return False testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) testlib.next_block( **kw ) # expired res = testlib.blockstack_name_preorder( "foo.test", wallets[2].privkey, wallets[3].addr ) # +name_cost if 'error' in res: print res return False testlib.next_block( **kw ) res = testlib.blockstack_name_register( "foo.test", wallets[2].privkey, wallets[3].addr ) if 'error' in res: print res return False testlib.next_block( **kw ) new_namespace_balance = testlib.get_balance(namespace_rec['address']) name_rec = testlib.get_name_blockchain_record('foo.test') name_cost = name_rec['op_fee'] testlib.next_block( **kw ) testlib.next_block( **kw ) # stop fee collection testlib.next_block( **kw ) testlib.next_block( **kw ) # expired if new_namespace_balance - namespace_balance != 4*name_cost: print 'address {} did not get credited'.format(namespace_rec['address']) print '{} != {} + 4*{}'.format(new_namespace_balance, namespace_balance, name_cost) return False # preorder should send to the null burn address now. res = testlib.blockstack_name_preorder( "foo2.test", wallets[4].privkey, wallets[0].addr ) # does not pay to namespace if 'error' in res: print res return False # try forcing it to the namespace burn address, to verify that it fails res = testlib.blockstack_name_preorder( "foo_fail.test", wallets[4].privkey, wallets[0].addr, burn_addr=namespace_rec['address'], expect_fail=True ) # does not pay to namespace (command fails) if 'error' not in res: print res return False res = testlib.blockstack_name_preorder( "foo_fail.test", wallets[4].privkey, wallets[0].addr, burn_addr=namespace_rec['address'], price={'units': 'BTC', 'amount': name_cost}, safety_checks=False, tx_fee=10000*5 ) # +name_cost if 'error' in res: print res return False testlib.next_block( **kw ) testlib.expect_snv_fail_at('foo_fail.test', testlib.get_current_block(**kw)) # should be accepted res = testlib.blockstack_name_register( "foo2.test", wallets[4].privkey, wallets[0].addr ) if 'error' in res: print res return False # should be rejected res = testlib.blockstack_name_register( "foo_fail.test", wallets[4].privkey, wallets[0].addr, safety_checks=False ) if 'error' in res: print res return False testlib.next_block( **kw ) testlib.expect_snv_fail_at('foo_fail.test', testlib.get_current_block(**kw)) # should have been rejected due to wrong burn address whois = testlib.blockstack_cli_whois('foo_fail.test') if 'error' not in whois: print whois return False new_burn_balance = testlib.get_balance(blockstack.lib.config.BLOCKSTACK_BURN_ADDRESS) new_namespace_balance = testlib.get_balance(namespace_rec['address']) name_rec_2 = testlib.get_name_blockchain_record('foo2.test') name_cost_2 = name_rec_2['op_fee'] # namespace should NOT have gotten the fee for foo_fail. It should only have gotten it for foo.test if new_namespace_balance - namespace_balance < 5*name_cost or new_namespace_balance - namespace_balance > 6*name_cost: print 'address {} got credited after fee capture period'.format(namespace_rec['address']) print '{} != {} + 5*{}'.format(new_namespace_balance, namespace_balance, name_cost) return False # burn address should have received the fee for the second name if new_burn_balance - name_cost_2 != burn_balance: print 'null burn address did not get credited' print '{} != {} + {}'.format(new_burn_balance, burn_balance, name_cost_2) return False def check( state_engine ): # not revealed, but ready ns = state_engine.get_namespace_reveal( "test" ) if ns is not None: print "namespace reveal exists" return False ns = state_engine.get_namespace( "test" ) if ns is None: print "no namespace" return False if ns['namespace_id'] != 'test': print "wrong namespace" return False for name in ['foo2.test']: # not preordered preorder = state_engine.get_name_preorder( name, virtualchain.make_payment_script(wallets[4].addr), wallets[0].addr ) if preorder is not None: print "preorder exists" return False # registered name_rec = state_engine.get_name( name ) if name_rec is None: print "name does not exist" return False # owned by if name_rec['address'] != wallets[0].addr or name_rec['sender'] != virtualchain.make_payment_script(wallets[0].addr): print "sender is wrong" return False return True ```
[ { "content": "```python\nimport json\n\nfrom django.db import models\nfrom django.utils import timezone\nfrom django.conf import settings\n\n\nclass Client(models.Model):\n\n RESPONSE_TYPE_CHOICES = [\n ('code', 'code (Authorization Code Flow)'),\n ('id_token', 'id_token (Implicit Flow)'),\n ...
[ { "content": "<|memory_start|>```python\nimport json\n\nfrom django.db import models\nfrom django.utils import timezone\nfrom django.conf import settings\n\n\nclass Client(models.Model):\n\n RESPONSE_TYPE_CHOICES = [\n ('code', 'code (Authorization Code Flow)'),\n ('id_token', 'id_token (Implic...
```python import json from django.db import models from django.utils import timezone from django.conf import settings class Client(models.Model): RESPONSE_TYPE_CHOICES = [ ('code', 'code (Authorization Code Flow)'), ('id_token', 'id_token (Implicit Flow)'), ('id_token token', 'id_token token (Implicit Flow)'), ] name = models.CharField(max_length=100, default='') client_id = models.CharField(max_length=255, unique=True) client_secret = models.CharField(max_length=255, unique=True) response_type = models.CharField(max_length=30, choices=RESPONSE_TYPE_CHOICES) _redirect_uris = models.TextField(default='') def __str__(self): return u'%s'.format(self.name) def __unicode__(self): return self.__str__() def redirect_uris(): def fget(self): return self._redirect_uris.splitlines() def fset(self, value): self._redirect_uris = '\n'.join(value) return locals() redirect_uris = property(**redirect_uris()) @property def default_redirect_uri(self): return self.redirect_uris[0] if self.redirect_uris else '' class BaseCodeTokenModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL) client = models.ForeignKey(Client) expires_at = models.DateTimeField() _scope = models.TextField(default='') def scope(): def fget(self): return self._scope.split() def fset(self, value): self._scope = ' '.join(value) return locals() scope = property(**scope()) def has_expired(self): return timezone.now() >= self.expires_at def __str__(self): return u'%s - %s (%s)'.format(self.client, self.user.email, self.expires_at) def __unicode__(self): return self.__str__() class Meta: abstract = True class Code(BaseCodeTokenModel): code = models.CharField(max_length=255, unique=True) nonce = models.CharField(max_length=255, blank=True, default='') class Token(BaseCodeTokenModel): access_token = models.CharField(max_length=255, unique=True) refresh_token = models.CharField(max_length=255, unique=True, null=True) _id_token = models.TextField() def id_token(): def fget(self): return json.loads(self._id_token) def fset(self, value): self._id_token = json.dumps(value) return locals() id_token = property(**id_token()) class UserConsent(BaseCodeTokenModel): pass ```
[ { "content": "Here is a code file:\n```python\n# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport copy\nimport re\nfrom importlib import import_module\n\nfrom django import forms\nfrom django.utils import six\nfrom django.utils.decorators import classonlymethod\nfrom django.utils.encodin...
[ { "content": "Here is a code file:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport copy\nimport re\nfrom importlib import import_module\n\nfrom django import forms\nfrom django.utils import six\nfrom django.utils.decorators import classonlymethod\nfrom djan...
```python # -*- coding: utf-8 -*- from __future__ import unicode_literals import copy import re from importlib import import_module from django import forms from django.utils import six from django.utils.decorators import classonlymethod from django.utils.encoding import python_2_unicode_compatible from django.utils.translation import ugettext_lazy as _ def format_display_label(cls_name): if cls_name.endswith('Field'): cls_name = cls_name[:-5] # Strip trailing 'Field' # Precedes each group of capital letters by a whitespace except first return re.sub(r'([A-Z]+)', r' \1', cls_name).lstrip() def load_class_from_string(cls_string): mod, cls = cls_string.rsplit('.', 1) module = import_module(mod) return getattr(module, cls) class DynamicFormFieldRegistry(object): def __init__(self): self._fields = {} def get(self, key): return self._fields.get(key) def get_as_choices(self): for k, c in sorted(six.iteritems(self._fields)): yield k, c.get_display_label() def register(self, cls): if not issubclass(cls, BaseDynamicFormField): raise ValueError('%r must inherit from %r' % ( cls, BaseDynamicFormField)) key = '%s.%s' % (cls.__module__, cls.__name__) self._fields[key] = cls def unregister(self, key): if key in self._fields: del self._fields[key] formfield_registry = DynamicFormFieldRegistry() dynamic_form_field_registry = formfield_registry def dynamic_form_field(cls): """ A class decorator to register the class as a dynamic form field in the :class:`DynamicFormFieldRegistry`. """ formfield_registry.register(cls) return cls class DFFMetaclass(type): def __new__(cls, name, bases, attrs): meta = attrs.pop('Meta', None) new_class = super(DFFMetaclass, cls).__new__(cls, name, bases, attrs) opts = {} super_opts = getattr(new_class, '_meta', {}) if meta: excludes = getattr(meta, '_exclude', ()) # Copy all attributes from super's options not excluded here. No # need to check for leading _ as this is already sorted out on the # super class for k, v in six.iteritems(super_opts): if k in excludes: continue opts[k] = v # Copy all attributes not starting with a '_' from this Meta class for k, v in six.iteritems(meta.__dict__): if k.startswith('_') or k in excludes: continue opts[k] = v else: opts = copy.deepcopy(super_opts) setattr(new_class, '_meta', opts) return new_class @python_2_unicode_compatible class BaseDynamicFormField(six.with_metaclass(DFFMetaclass)): cls = None display_label = None widget = None class Meta: help_text = [six.string_types, '', (forms.CharField, forms.Textarea)] required = [bool, True, forms.NullBooleanField] def __new__(cls, *args, **kwargs): self = super(BaseDynamicFormField, cls).__new__(cls) self._meta = copy.deepcopy(self.__class__._meta) return self def __init__(self, name, label, widget_attrs=None, **kwargs): self.name = name self.label = label self.widget_attrs = widget_attrs or {} self.set_options(**kwargs) def __str__(self): if isinstance(self.cls, six.string_types): clsname = self.cls else: clsname = '%s.%s' % (self.cls.__module__, self.cls.__name__) return '<%(class)s, name=%(name)s, label=%(label)s>' % { 'class': clsname, 'name': self.name, 'label': self.label, } def construct(self, **kwargs): if isinstance(self.cls, six.string_types): cls_type = load_class_from_string(self.cls) else: cls_type = self.cls f_kwargs = {} for key, val in six.iteritems(self.options): f_kwargs[key] = val[1] f_kwargs['label'] = self.label if self.widget is not None: if isinstance(self.widget, six.string_types): widget_type = load_class_from_string(self.widget) else: widget_type = self.widget f_kwargs['widget'] = widget_type(**self.get_widget_attrs()) f_kwargs.update(kwargs) # Update the field kwargs by those given return cls_type(**f_kwargs) def contribute_to_form(self, form): form.fields[self.name] = self.construct() @classonlymethod def get_display_label(cls): if cls.display_label: return cls.display_label return format_display_label(cls.__name__) @property def options(self): return self._meta def get_widget_attrs(self): return self.widget_attrs def set_options(self, **kwargs): for key, value in six.iteritems(kwargs): if key not in self.options: raise KeyError('%s is not a valid option.' % key) expected_type = self.options[key][0] if not isinstance(value, expected_type) and value is not None: raise TypeError('Neither of type %r nor None' % expected_type) self.options[key][1] = value self.options_valid() def options_valid(self): return True @classonlymethod def do_display_data(cls): return True @dynamic_form_field class BooleanField(BaseDynamicFormField): cls = 'django.forms.BooleanField' display_label = _('Boolean') @dynamic_form_field class ChoiceField(BaseDynamicFormField): cls = 'django.forms.ChoiceField' display_label = _('Choices') class Meta: choices = [six.string_types, '', (forms.CharField, forms.Textarea)] def construct(self, **kwargs): value = self.options.get('choices')[1] choices = [(row, row) for row in value.splitlines() if row] return super(ChoiceField, self).construct(choices=choices) def options_valid(self): if not self.options['choices'] or not self.options['choices'][1]: raise ValueError('choices must not be defined for %r' % self) return True @dynamic_form_field class DateField(BaseDynamicFormField): cls = 'django.forms.DateField' display_label = _('Date') class Meta: localize = [bool, True, forms.NullBooleanField] @dynamic_form_field class DateTimeField(BaseDynamicFormField): cls = 'django.forms.DateTimeField' display_label = _('Date and Time') class Meta: localize = [bool, True, forms.NullBooleanField] @dynamic_form_field class EmailField(BaseDynamicFormField): cls = 'django.forms.EmailField' display_label = _('Email') @dynamic_form_field class IntegerField(BaseDynamicFormField): cls = 'django.forms.IntegerField' display_label = _('Integer') class Meta: localize = [bool, True, forms.NullBooleanField] max_value = [int, None, forms.IntegerField] min_value = [int, None, forms.IntegerField] @dynamic_form_field class MultiLineTextField(BaseDynamicFormField): cls = 'django.forms.CharField' display_label = _('Multi Line Text') widget = 'django.forms.widgets.Textarea' @dynamic_form_field class SingleLineTextField(BaseDynamicFormField): cls = 'django.forms.CharField' display_label = _('Single Line Text') class Meta: max_length = [int, None, forms.IntegerField] min_length = [int, None, forms.IntegerField] @dynamic_form_field class TimeField(BaseDynamicFormField): cls = 'django.forms.TimeField' display_label = _('Time') class Meta: localize = [bool, True, forms.NullBooleanField] ```
[ { "content": "Here is the script:\n```python\nclass ProxyDict(object):\r\n def __init__(self, parent, collection_name, childclass, keyname):\r\n self.parent = parent\r\n self.collection_name = collection_name\r\n self.childclass = childclass\r\n self.keyname = keyname\r\n\r\n @...
[ { "content": "Here is the script:\n<|memory_start|>```python\nclass ProxyDict(object):\r\n def __init__(self, parent, collection_name, childclass, keyname):\r\n self.parent = parent\r\n self.collection_name = collection_name\r\n self.childclass = childclass\r\n self.keyname = keyn...
```python class ProxyDict(object): def __init__(self, parent, collection_name, childclass, keyname): self.parent = parent self.collection_name = collection_name self.childclass = childclass self.keyname = keyname @property def collection(self): return getattr(self.parent, self.collection_name) def keys(self): descriptor = getattr(self.childclass, self.keyname) return [x[0] for x in self.collection.values(descriptor)] def __getitem__(self, key): x = self.collection.filter_by(**{self.keyname:key}).first() if x: return x else: raise KeyError(key) def __setitem__(self, key, value): try: existing = self[key] self.collection.remove(existing) except KeyError: pass self.collection.append(value) from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import create_engine, Column, Integer, String, ForeignKey from sqlalchemy.orm import sessionmaker, relationship engine = create_engine('sqlite://', echo=True) Base = declarative_base(engine) class Parent(Base): __tablename__ = 'parent' id = Column(Integer, primary_key=True) name = Column(String(50)) _collection = relationship("Child", lazy="dynamic", cascade="all, delete-orphan") @property def child_map(self): return ProxyDict(self, '_collection', Child, 'key') class Child(Base): __tablename__ = 'child' id = Column(Integer, primary_key=True) key = Column(String(50)) parent_id = Column(Integer, ForeignKey('parent.id')) def __repr__(self): return "Child(key=%r)" % self.key Base.metadata.create_all() sess = sessionmaker()() p1 = Parent(name='p1') sess.add(p1) print "\n---------begin setting nodes, autoflush occurs\n" p1.child_map['k1'] = Child(key='k1') p1.child_map['k2'] = Child(key='k2') # this will autoflush the current map. # ['k1', 'k2'] print "\n---------print keys - flushes first\n" print p1.child_map.keys() # k1 print "\n---------print 'k1' node\n" print p1.child_map['k1'] print "\n---------update 'k2' node - must find existing, and replace\n" p1.child_map['k2'] = Child(key='k2') print "\n---------print 'k2' key - flushes first\n" # k2 print p1.child_map['k2'] print "\n---------print all child nodes\n" # [k1, k2b] print sess.query(Child).all() ```
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n```python\n\"\"\"\nDescriptive HTTP status codes, for code readability.\n\nSee RFC 2616 - Sec 10: http://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html\nAlso see django.core.handlers.wsgi.STATUS_CODE_TEXT\n\"\"\"\n\nHTTP_100_CONTINUE = 100\nHT...
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n<|memory_start|>```python\n\"\"\"\nDescriptive HTTP status codes, for code readability.\n\nSee RFC 2616 - Sec 10: http://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html\nAlso see django.core.handlers.wsgi.STATUS_CODE_TEXT\n\"\"\"\n\nHTTP_100_CO...
```python """ Descriptive HTTP status codes, for code readability. See RFC 2616 - Sec 10: http://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html Also see django.core.handlers.wsgi.STATUS_CODE_TEXT """ HTTP_100_CONTINUE = 100 HTTP_101_SWITCHING_PROTOCOLS = 101 HTTP_200_OK = 200 HTTP_201_CREATED = 201 HTTP_202_ACCEPTED = 202 HTTP_203_NON_AUTHORITATIVE_INFORMATION = 203 HTTP_204_NO_CONTENT = 204 HTTP_205_RESET_CONTENT = 205 HTTP_206_PARTIAL_CONTENT = 206 HTTP_300_MULTIPLE_CHOICES = 300 HTTP_301_MOVED_PERMANENTLY = 301 HTTP_302_FOUND = 302 HTTP_303_SEE_OTHER = 303 HTTP_304_NOT_MODIFIED = 304 HTTP_305_USE_PROXY = 305 HTTP_306_RESERVED = 306 HTTP_307_TEMPORARY_REDIRECT = 307 HTTP_400_BAD_REQUEST = 400 HTTP_401_UNAUTHORIZED = 401 HTTP_402_PAYMENT_REQUIRED = 402 HTTP_403_FORBIDDEN = 403 HTTP_404_NOT_FOUND = 404 HTTP_405_METHOD_NOT_ALLOWED = 405 HTTP_406_NOT_ACCEPTABLE = 406 HTTP_407_PROXY_AUTHENTICATION_REQUIRED = 407 HTTP_408_REQUEST_TIMEOUT = 408 HTTP_409_CONFLICT = 409 HTTP_410_GONE = 410 HTTP_411_LENGTH_REQUIRED = 411 HTTP_412_PRECONDITION_FAILED = 412 HTTP_413_REQUEST_ENTITY_TOO_LARGE = 413 HTTP_414_REQUEST_URI_TOO_LONG = 414 HTTP_415_UNSUPPORTED_MEDIA_TYPE = 415 HTTP_416_REQUESTED_RANGE_NOT_SATISFIABLE = 416 HTTP_417_EXPECTATION_FAILED = 417 HTTP_500_INTERNAL_SERVER_ERROR = 500 HTTP_501_NOT_IMPLEMENTED = 501 HTTP_502_BAD_GATEWAY = 502 HTTP_503_SERVICE_UNAVAILABLE = 503 HTTP_504_GATEWAY_TIMEOUT = 504 HTTP_505_HTTP_VERSION_NOT_SUPPORTED = 505 ```
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n```python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import unicode_literals\nfrom django.apps import apps\nfrom django.contrib import messages\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.urlresolvers import re...
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import unicode_literals\nfrom django.apps import apps\nfrom django.contrib import messages\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.urlres...
```python # -*- coding: utf-8 -*- from __future__ import unicode_literals from django.apps import apps from django.contrib import messages from django.contrib.auth.decorators import login_required from django.core.urlresolvers import reverse_lazy from django.db.models import Q from django.shortcuts import redirect from django.views.generic import CreateView, ListView, DetailView, UpdateView, DeleteView from djspikeval.forms import AlgorithmForm __all__ = [ "AlgorithmBaseView", "AlgorithmList", "AlgorithmCreate", "AlgorithmDetail", "AlgorithmUpdate", "AlgorithmDelete"] __author__ = "pmeier82" Algorithm = apps.get_model("djspikeval", "algorithm") class AlgorithmBaseView(object): model = Algorithm class AlgorithmList(AlgorithmBaseView, ListView): template_name = "djspikeval/algorithm/list.html" paginate_by = 10 def get_context_data(self, **kwargs): cntx = super(AlgorithmList, self).get_context_data(**kwargs) cntx.update(scope=self.request.GET.get("scope")) return cntx def get_queryset(self): if self.request.GET.get("scope"): scope = self.request.GET.get("scope") return Algorithm.objects.filter( Q(name__icontains=scope) | Q(kind__name__icontains=scope)) return Algorithm.objects.all() class AlgorithmCreate(AlgorithmBaseView, CreateView): template_name = "djspikeval/algorithm/create.html" form_class = AlgorithmForm def get_form_kwargs(self): kwargs = super(AlgorithmCreate, self).get_form_kwargs() kwargs["user"] = self.request.user return kwargs class AlgorithmDetail(AlgorithmBaseView, DetailView): template_name = "djspikeval/algorithm/detail.html" class AlgorithmUpdate(AlgorithmBaseView, UpdateView): template_name = "djspikeval/algorithm/update.html" form_class = AlgorithmForm def get_form_kwargs(self): kwargs = super(AlgorithmUpdate, self).get_form_kwargs() kwargs["user"] = self.request.user return kwargs class AlgorithmDelete(AlgorithmBaseView, DeleteView): template_name = "djspikeval/algorithm/delete.html" success_url = reverse_lazy("algorithm:list") if __name__ == "__main__": pass ```
[ { "content": "Provide an exact copy of the source code:\n```python\nimport datetime\nimport logging\nimport Mollie\n\nfrom django.conf import settings\nfrom django.contrib.auth.decorators import login_required\nfrom django.template.context_processors import csrf\nfrom django.core.urlresolvers import reverse\nfr...
[ { "content": "Provide an exact copy of the source code:\n<|memory_start|>```python\nimport datetime\nimport logging\nimport Mollie\n\nfrom django.conf import settings\nfrom django.contrib.auth.decorators import login_required\nfrom django.template.context_processors import csrf\nfrom django.core.urlresolvers im...
```python import datetime import logging import Mollie from django.conf import settings from django.contrib.auth.decorators import login_required from django.template.context_processors import csrf from django.core.urlresolvers import reverse from django.http import HttpResponse, HttpResponseRedirect from django.shortcuts import render_to_response, get_object_or_404 from django.utils.html import escape from django.utils.safestring import mark_safe from django.utils.translation import ugettext as _ from django.views.decorators.csrf import csrf_exempt from django.views.generic.list import ListView from subscribe.models import Event, EventQuestion from subscribe.forms import Registration, SubscribeForm, fill_subscription def event_message(request, event, message): c = {"event": event, "request": request, "message": message} return render_to_response("subscribe/event_message.html", c) def register(request, slug): logger = logging.getLogger(__name__) # Get the event event = get_object_or_404(Event, slug=slug) # If not staff, check if allowed if not request.user.is_staff: now = datetime.datetime.now() if event.start_registration > now or event.end_registration < now: return event_message(request, event, _("Inschrijving gesloten.")) if event.is_full(): return event_message(request, event, _("Helaas is het maximum aantal inschrijvingen bereikt.")) # If this is not a POST request if request.method != "POST": # then just create the form... form = SubscribeForm(event) c = {"event": event, "request": request, "form": form, "user_is_staff": request.user.is_staff} c.update(csrf(request)) return render_to_response("subscribe/form.html", c) # It is a POST request, check if the form is valid... form = SubscribeForm(event, request.POST) if not form.is_valid(): c = {"event": event, "request": request, "form": form, "user_is_staff": request.user.is_staff} c.update(csrf(request)) return render_to_response("subscribe/form.html", c) # It is a POST request, and the form is valid, check if this is the confirmation page if 'registration_preview' not in request.POST and not form.is_free_event: form.confirm_page() c = {"event": event, "request": request, "form": form, "user_is_staff": request.user.is_staff} c.update(csrf(request)) return render_to_response("subscribe/form.html", c) # Maybe this is just a test of the confirmation email? if 'testsubmit' in request.POST: subscription = fill_subscription(form, event) msg = subscription.send_confirmation_email() subscription.delete() msg = '<br/>'.join(escape(msg).split('\n')) return event_message(request, event, mark_safe("De volgende email is verstuurd:<br/><br/>{}".format(msg))) # We confirmed. Create the subscription in the database... subscription = fill_subscription(form, event) # Check if the subscription form could be saved if not subscription: # Error Filling subscription error_str = "Error in saving form." return HttpResponse(_(error_str)) # Check (again) if maybe the number of registrations is over the limit... if subscription in event.get_registrations_over_limit(): subscription.delete() if event.is_full(): error_str = "De inschrijving kan niet worden voltooid, omdat het maximum aantal inschrijvingen is bereikt." else: error_str = "De inschrijving kan niet worden voltooid, omdat een van de gekozen opties het maximum aantal inschrijvingen heeft bereikt." return event_message(request, event, _(error_str)) # Check if we need to pay or not... if subscription.price <= 0: subscription.paid = True subscription.send_confirmation_email() subscription.save() return event_message(request, event, _("Inschrijving geslaagd. Ter bevestiging is een e-mail verstuurd.")) # Payment required... try: mollie = Mollie.API.Client() mollie.setApiKey(settings.MOLLIE_KEY) # METADATA TOEVOEGEN webhookUrl = request.build_absolute_uri(reverse("webhook", args=[subscription.id])) redirectUrl = request.build_absolute_uri(reverse("return_page", args=[subscription.id])) payment = mollie.payments.create({ 'amount': float(subscription.price) / 100.0, 'description': subscription.event.name, 'webhookUrl': webhookUrl, 'redirectUrl': redirectUrl, }) subscription.trxid = payment["id"] subscription.save() return HttpResponseRedirect(payment.getPaymentUrl()) except Mollie.API.Error as e: error_str = "register: Technische fout, probeer later opnieuw.\n\n" + str(e) logger.error(error_str) return event_message(request, event, _(error_str)) def check_transaction(subscription): logger = logging.getLogger(__name__) logger.info('check_transaction: Checking transaction %d with id %s' % (subscription.id, subscription.trxid)) mollie = Mollie.API.Client() mollie.setApiKey(settings.MOLLIE_KEY) payment = mollie.payments.get(subscription.trxid) logger.info("check_transaction: Transaction %s has status %s" % (subscription.id, payment['status'])) subscription.status = payment['status'] subscription.paid = payment.isPaid() subscription.save() if subscription.paid: subscription.send_confirmation_email() # called when the user returns from Mollie def return_page(request, id): logger = logging.getLogger(__name__) logger.info('views::return_page() - registration id: ' + str(id)) # Retrieve the registration try: subscription = Registration.objects.get(id=id) except: return HttpResponse(_("iDEAL error (onbekende inschrijving): Neem contact op met ict@jongedemocraten.nl. Controleer of uw betaling is afgeschreven alvorens de betaling opnieuw uit te voeren.")) # If status unknown, then check it... if subscription.status == "": try: check_transaction(subscription) except Mollie.API.Error as e: error_str = "return_page: Technische fout, probeer later opnieuw." + "\n\n%s" % (str(e),) logger.error(error_str) return event_message(request, subscription.event, _(error_str)) if subscription.status == "paid": return event_message(request, subscription.event, _("Betaling geslaagd. Ter bevestiging is een e-mail verstuurd.")) elif subscription.status == "cancelled" or subscription.status == "expired": return event_message(request, subscription.event, _("Je betaling is geannuleerd.")) elif subscription.status == "open" or subscription.status == "pending": return event_message(request, subscription.event, _("Je betaling staat geregistreerd in ons systeem, maar wordt nog verwerkt door onze bank. Als je binnen een uur geen bevestigingsmail ontvangt, is er mogelijk iets fout gegaan met de betaling. Neem in dat geval contact op met ict@jongedemocraten.nl.")) else: return event_message(request, subscription.event, _("Er is een fout opgetreden bij het verwerken van je iDEAL transactie. Neem contact op met ict@jongedemocraten.nl of probeer het later nogmaals. Controleer of je betaling is afgeschreven alvorens de betaling opnieuw uit te voeren.")) @csrf_exempt def webhook(request, id): # trigger checking if request.method == "POST": transaction_id = request.POST['id'] else: transaction_id = request.GET['id'] logger = logging.getLogger(__name__) logger.info('views::check() - id: %s, transaction id: %s' % (id, transaction_id)) try: subscription = Registration.objects.get(id=id, trxid=transaction_id) except: logger.error("views::check() - cannot find matching subscription") return HttpResponse(_("NOT OK")) try: check_transaction(subscription) except Mollie.API.Error as e: logger.error("webhook: error %s" % (str(e),)) return HttpResponse(_("OK")) @login_required def delete_event_question(request): questionId = request.GET['questionId'] warning = int(request.GET['warning']) if warning == 0: eventQuestion = EventQuestion.objects.get(pk=questionId) eventQuestion.delete() return HttpResponse(_('Vraag verwijderd. <br /> <a href="/admin/">Terug naar admin.</a>')) else: return HttpResponse(_("""Weet je zeker dat je deze vraag wilt verwijderen? <br /> <a href="/deleteEventQuestion/?questionId=%d&warning=%d">Ja</a> <a href="/admin/">Nee</a>""" % (int(questionId), 0))) class HomeView(ListView): model = Event queryset = Event.objects.order_by('-end_registration', '-start_registration') template_name = "subscribe/index.html" context_object_name = "events" def get(self, request, *args, **kwargs): if not request.user.is_staff: now = datetime.datetime.now() self.queryset = self.queryset.filter(start_registration__lte=now, end_registration__gte=now) return super().get(self, request, *args, **kwargs) ```
[ { "content": "Here is the snippet:\n```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# dbload.py\n#\n# Copyright 2015 Hartland PC LLC\n#\n# This file is part of the of the database loader for CCE 4.0 (open source version).\n#\n# This package is free software: you can redistribute it and/or modify\...
[ { "content": "Here is the snippet:\n<|memory_start|>```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# dbload.py\n#\n# Copyright 2015 Hartland PC LLC\n#\n# This file is part of the of the database loader for CCE 4.0 (open source version).\n#\n# This package is free software: you can redistribute i...
```python #!/usr/bin/env python # -*- coding: utf-8 -*- # # dbload.py # # Copyright 2015 Hartland PC LLC # # This file is part of the of the database loader for CCE 4.0 (open source version). # # This package is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This package is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this package. If not, see <http://www.gnu.org/licenses/>. import sys import stats from comm import * from decimal import * def startcheck(lockdir, recheckdir): # Check to see if the loader is already running and set lock file if os.access(os.path.expanduser(lockdir), os.F_OK): pidfile = open(os.path.expanduser(lockdir), "r") pidfile.seek(0) old_pid = pidfile.readline() if os.path.exists("/proc/%s" % old_pid): sys.exit(0) else: os.remove(os.path.expanduser(lockdir)) pidfile = open(os.path.expanduser(lockdir), "w") pidfile.write("%s" % os.getpid()) pidfile.close() # Check for recheck file ret = 'normal' if os.access(os.path.expanduser(recheckdir), os.F_OK): ret = 'recheck' else: checkfile = open(os.path.expanduser(recheckdir), "w") checkfile.close() return ret # Error Logging def loader_error_log(msg, function_name='No function name provided'): currtime = time.strftime('%m-%d %H:%M:%S', time.gmtime()) logging.basicConfig(filename=str(os.getcwd() + "/loader.log"), level=logging.ERROR) logging.error(currtime + ' ' + str(msg) + ' : ' + str(function_name)) # Address accounting. If credit is true, amount is added to address, else amount is subtracted. # count_tx determines if the number of transactions on an account is incremented, decremented or unchanged. def accounting(address, amount, credit, count_tx): try: ret = query_single('SELECT balance FROM address WHERE address = %s', address) if ret is None: ret = query_noreturn('INSERT INTO address (address,balance) VALUES(%s,%s)', address, amount) conn.commit() else: if credit: balance = Decimal(ret[0] + amount) else: balance = Decimal(ret[0] - amount) if balance < 0: balance = Decimal(0) ret = query_noreturn('UPDATE address SET balance = %s WHERE address = %s', balance, address) if count_tx == 'add': ret = query_noreturn('UPDATE address SET n_tx = n_tx + 1 WHERE address = %s', address) elif count_tx == 'subtract': ret = query_noreturn('UPDATE address SET n_tx = abs(n_tx - 1) WHERE address = %s', address) conn.commit() except Exception as e: loader_error_log(e, "Accounting loop error") # Place data in table rows def add_row(table, row_data): cur = conn.cursor() cur.execute("describe %s" % table) allowed_keys = set(row[0] for row in cur.fetchall()) keys = allowed_keys.intersection(row_data) columns = ", ".join(keys) data_template = ", ".join(["%s"] * len(keys)) sql = "insert into %s (%s) values (%s)" % (table, columns, data_template) data_tuple = tuple(row_data[key] for key in keys) cur.execute(sql, data_tuple) cur.close() # Parse Transaction def process_tx(tx_hash, blk_height): rawtx = jsonrpc("getrawtransaction", tx_hash) if rawtx['Status'] == 'error': loader_error_log(rawtx['Data'], str('Raw tx on block:' + blk_height)) return rawtx decode = jsonrpc("decoderawtransaction", rawtx['Data']) if decode['Status'] == 'error': loader_error_log(decode['Data'], str('Decode tx on block:' + blk_height)) return decode jsn_decode = json.dumps(decode['Data']) ret = query_noreturn('INSERT INTO tx_raw (tx_hash,raw,decoded,height) VALUES(%s,%s,%s,%s)', tx_hash, rawtx['Data'], jsn_decode, blk_height) total_out = Decimal(0) # Transaction addresses are stored in tx_address to determine duplicate addresses in tx_in / tx_out. # POS chains use the same address in both tx_in and tx_out for the generation transaction. # If a duplicate address is found, the tx count for address will only be incremented once. tx_address = [] for key in decode['Data']['vout']: try: key['address'] = key['scriptPubKey']['addresses'][0] tx_address.append(key['address']) # KeyError is not fatal, as generation transactions have no tx_in address except KeyError: key['address'] = "Unknown" key['asm'] = key['scriptPubKey']['asm'] key['type'] = key['scriptPubKey']['type'] key['height'] = blk_height key['tx_hash'] = tx_hash key['raw'] = rawtx['Data'] key['value'] = Decimal(str(key['value'])) add_row('tx_out', key) if key['address'] != 'Unknown': accounting(key['address'], key['value'], True, 'add') conn.commit() total_out = Decimal(total_out + key['value']) # If the transaction total out is larger then the lowest entry on the large tx table, # replace the lowest transaction with this transaction try: low = query_single('SELECT * FROM large_tx ORDER BY amount ASC LIMIT 1') if total_out > low[1]: ret = query_noreturn('UPDATE large_tx SET tx = %s,amount = %s WHERE tx = %s', tx_hash, total_out,low[0]) # Exceptions in this block are non-fatal as the information value of the transaction itself far exceeds the value of large_tx except: pass for key in decode['Data']['vin']: try: key['asm'] = key['scriptSig']['asm'] key['hex'] = key['scriptSig']['hex'] key['prev_out_hash'] = key['txid'] ret = query_single('SELECT * FROM tx_out WHERE tx_hash = %s AND n = %s', key['prev_out_hash'], key['vout']) if not ret: key['address'] = 'Not Available' key['value_in'] = Decimal(total_out) else: count_tx = 'add' key['address'] = str(ret[4]) key['value_in'] = ret[2] if key['address'] in tx_address: count_tx = 'no' accounting(key['address'],key['value_in'],False,count_tx) # Exceptions occur in this loop due to POW generation transactions. # The value of tx_in and tx_out are always the same in these types of transactions except Exception: key['value_in'] = total_out key['tx_hash'] = tx_hash key['height'] = blk_height add_row('tx_in', key) return {'Status': 'ok', 'Data': {'out': total_out}} # Parse block def process_block(blk_height): try: if blk_height == -1: raise Exception('Bad block height (-1)') counter = 0 total_sent = Decimal(0) b_hash = jsonrpc("getblockhash", blk_height)['Data'] block = jsonrpc("getblock", b_hash)['Data'] # In POS chains, nonce is used to determine if a block is POS. # The 'flags' field in the daemon output is unreliable due to different verbiage and multiple flags. # Merged mine chains also use 0 in the nonce field. This system will not work with POS merged mined chains. # POS merged mined compatibility will be added in the future if CONFIG["chain"]["pos"] == 'true' and block['nonce'] == 0: counter = 1 for key in block['tx']: if counter == 1: counter = 2 elif counter == 2: block['pos'] = key counter = 0 prostx = process_tx(key, blk_height) if prostx['Status'] == 'error': raise Exception(prostx['Data']) total_sent = Decimal(total_sent + prostx['Data']['out']) block['raw'] = json.dumps(block, sort_keys=False, indent=1) add_row('block', block) conn.commit() ret = query_noreturn('UPDATE block SET total_sent = %s, n_tx = %s WHERE height = %s', total_sent, len(block['tx']), blk_height) conn.commit() except Exception as e: return {'Status':'error','Data':e} return {'Status':'ok'} # Orphan correction. Copy to orphan tables,delete block/tx information, and re-parse block. # If recheck is true, block/tx information is not copied to orphan tables. def orphan(blk_height, recheck=False): try: if not recheck: loader_error_log("Orphan routine called", blk_height) ret = query_noreturn('INSERT INTO orph_block SELECT * FROM block WHERE height = %s', blk_height) ret = query_noreturn('INSERT INTO orph_tx_raw SELECT * FROM tx_raw WHERE height = %s', blk_height) ret = query_noreturn('DELETE FROM block WHERE height = %s', blk_height) ret = query_noreturn('DELETE FROM tx_raw WHERE height = %s', blk_height) txin = query_multi('SELECT * FROM tx_in WHERE height = %s', blk_height) for key in txin: if key[7] != '0': accounting(str(key[7]),key[6], True,'subtract') txout = query_multi('SELECT * FROM tx_out WHERE height = %s', blk_height) for key in txout: accounting(str(key[4]),key[2], False,'subtract') if not recheck: ret = query_noreturn('INSERT INTO orph_tx_in SELECT * FROM tx_in WHERE height = %s', blk_height) ret = query_noreturn('INSERT INTO orph_tx_out SELECT * FROM tx_out WHERE height = %s', blk_height) ret = query_noreturn('INSERT INTO orph_tx_raw SELECT * FROM tx_raw WHERE height = %s', blk_height) ret = query_noreturn('DELETE FROM tx_in WHERE height = %s', blk_height) ret = query_noreturn('DELETE FROM tx_out WHERE height = %s', blk_height) ret = query_noreturn('DELETE FROM tx_raw WHERE height = %s', blk_height) ret = process_block(blk_height) if ret['status'] == 'error': raise Exception(ret['Data']) conn.commit() except Exception as e: loader_error_log(e, "Orphan loop error") conn.rollback() if not recheck: loader_error_log('Successful orphan recovery: ', str(blk_height)) def main(argv): lockdir = str(os.getcwd() + "/" + "dataload.lock") recheckdir = str(os.getcwd() + "/" + "recheck") startmode = startcheck(lockdir, recheckdir) verbose = False # Set cowtime (loader timeout) to 5 minutes cowtime = 60 * 5 try: for opt in argv: # Set new database mode and cowtime to 24 hours if -n flag if opt == '-n': startmode = 'newdb' cowtime = 60 * 60 * 24 # Run recheck if -r flag elif opt == '-r' and startmode != 'newdb': startmode = 'recheck' # Send verbose messages to stderr if -v flag elif opt == '-v': verbose = True # Set cowtime to 24 hours if -l flag elif opt == '-l': cowtime = 60 * 60 * 24 except: pass try: with timeout(cowtime, exception=Exception('DBLoader Timeout')): # Get block heights daemon = jsonrpc("getblockcount") if daemon['Status'] != 'error': top_height = daemon['Data'] blk_height = query_single('SELECT height FROM block ORDER BY height DESC LIMIT 1') if not blk_height: blk_height = 1 else: blk_height = int(blk_height[0] + 1) else: loader_error_log(daemon['Data'], 'Get Block Height') raise Exception(daemon['Data']) # Sleep is needed to allow the daemon time to catch orphans if startmode != 'newdb': time.sleep(15) # Recheck mode, re-parse the last 5 blocks in the database if startmode == 'recheck' and blk_height > 5: if verbose: print >> sys.stderr, "Recheck Called" for blk in range(blk_height - 5, blk_height): orphan(blk, True) # Check last (blockcheck) blocks for orphans and fix if needed blockcheck = int(CONFIG["loader"]["blockcheck"]) if blk_height > blockcheck: for blk in range(blk_height - blockcheck, blk_height): d_hash = jsonrpc('getblockhash', blk) db_hash = query_single('SELECT hash FROM block where height = %s', blk)[0] if d_hash['Data'] != db_hash: orphan(blk) # Genesis block TX needs to be entered manually. Process block information only if startmode == 'newdb': b_hash = jsonrpc("getblockhash", 0)['Data'] block = jsonrpc("getblock", b_hash)['Data'] block['raw'] = json.dumps(block, sort_keys=False, indent=1) add_row('block', block) # Set up top_address table for i in range(int(CONFIG['stat']['richlistlen'])): ret = query_noreturn('INSERT INTO top_address (rank) VALUES(%s)', i + 1) # Set up stats table ret = query_noreturn('INSERT INTO stats (peer_txt) VALUES("None")') blk_height = 1 # Process blocks loop while blk_height <= top_height: ret = process_block(blk_height) if ret['Status'] == 'error': raise Exception(ret['Data']) if startmode == 'newdb' and blk_height == 101: ret = query_noreturn('TRUNCATE large_tx') time.sleep(5) ret = query_noreturn('INSERT INTO large_tx SELECT tx_hash,SUM(value) FROM tx_out GROUP BY tx_hash ORDER BY SUM(value) DESC LIMIT 100') blk_height += 1 if verbose: print >> sys.stderr, 'Processing Block: ', blk_height, ' of ', top_height, '\r', # Call Statistics module if CONFIG['loader']['stats'] == 'true': if verbose: print >> sys.stderr, '\nCalling Statistics Module' stats.main() except Exception as e: loader_error_log(str(e), 'Main loop') conn.close() os.remove(os.path.expanduser(lockdir)) if verbose: print >> sys.stderr, '\nMain Loop', str(e) sys.exit(0) # Clean up conn.close() if verbose: print >> sys.stderr, "Database load complete" os.remove(os.path.expanduser(recheckdir)) os.remove(os.path.expanduser(lockdir)) if __name__ == '__main__': main(sys.argv[1:]) ```
[ { "content": "```python\nfrom .adaptive import *\nfrom .base import build, Optimizer\nfrom .dataset import Dataset\nfrom .first_order import *\n\n__version__ = '0.2.2'\n\n\ndef minimize(loss, train, valid=None, params=None, inputs=None, algo='rmsprop',\n updates=(), monitors=(), monitor_gradients=Fa...
[ { "content": "<|memory_start|>```python\nfrom .adaptive import *\nfrom .base import build, Optimizer\nfrom .dataset import Dataset\nfrom .first_order import *\n\n__version__ = '0.2.2'\n\n\ndef minimize(loss, train, valid=None, params=None, inputs=None, algo='rmsprop',\n updates=(), monitors=(), moni...
```python from .adaptive import * from .base import build, Optimizer from .dataset import Dataset from .first_order import * __version__ = '0.2.2' def minimize(loss, train, valid=None, params=None, inputs=None, algo='rmsprop', updates=(), monitors=(), monitor_gradients=False, batch_size=32, train_batches=None, valid_batches=None, **kwargs): '''Minimize a loss function with respect to some symbolic parameters. Additional keyword arguments are passed to the underlying :class:`Optimizer <downhill.base.Optimizer>` instance. Parameters ---------- loss : Theano expression Loss function to minimize. This must be a scalar-valued expression. train : :class:`Dataset <downhill.dataset.Dataset>`, ndarray, or callable Dataset to use for computing gradient updates. valid : :class:`Dataset <downhill.dataset.Dataset>`, ndarray, or callable, optional Dataset to use for validating the minimization process. The training dataset is used if this is not provided. params : list of Theano variables, optional Symbolic variables to adjust to minimize the loss. If not given, these will be computed automatically by walking the computation graph. inputs : list of Theano variables, optional Symbolic variables required to compute the loss. If not given, these will be computed automatically by walking the computation graph. algo : str, optional Name of the minimization algorithm to use. Must be one of the strings that can be passed to :func:`build`. Defaults to ``'rmsprop'``. updates : list of update pairs, optional A list of pairs providing updates for the internal of the loss computation. Normally this is empty, but it can be provided if the loss, for example, requires an update to an internal random number generator. monitors : dict or sequence of (str, Theano expression) tuples, optional Additional values to monitor during optimization. These must be provided as either a sequence of (name, expression) tuples, or as a dictionary mapping string names to Theano expressions. monitor_gradients : bool, optional If True, add monitors to log the norms of the parameter gradients during optimization. Defaults to False. batch_size : int, optional Size of batches provided by datasets. Defaults to 32. train_batches : int, optional Number of batches of training data to iterate over during one pass of optimization. Defaults to None, which uses the entire training dataset. valid_batches : int, optional Number of batches of validation data to iterate over during one pass of validation. Defaults to None, which uses the entire validation dataset. Returns ------- train_monitors : dict A dictionary mapping monitor names to monitor values. This dictionary will always contain the ``'loss'`` key, giving the value of the loss evaluated on the training dataset. valid_monitors : dict A dictionary mapping monitor names to monitor values, evaluated on the validation dataset. This dictionary will always contain the ``'loss'`` key, giving the value of the loss function. Because validation is not always computed after every optimization update, these monitor values may be "stale"; however, they will always contain the most recently computed values. ''' if not isinstance(train, Dataset): train = Dataset( train, name='train', batch_size=batch_size, iteration_size=train_batches, ) if valid is not None and not isinstance(valid, Dataset): valid = Dataset( valid, name='valid', batch_size=batch_size, iteration_size=valid_batches, ) return build( algo, loss=loss, params=params, inputs=inputs, updates=updates, monitors=monitors, monitor_gradients=monitor_gradients, ).minimize(train, valid, **kwargs) ```
[ { "content": "Here is the script:\n```python\nfrom os.path import exists, abspath, dirname, join\nimport misc\n\n\nTHIS_DIR = dirname(abspath(__file__))\n\n# this is a personal access token used by chaosbot to perform merges and other\n# api requests. it is a secret, and lives on the server, but since chaosbot...
[ { "content": "Here is the script:\n<|memory_start|>```python\nfrom os.path import exists, abspath, dirname, join\nimport misc\n\n\nTHIS_DIR = dirname(abspath(__file__))\n\n# this is a personal access token used by chaosbot to perform merges and other\n# api requests. it is a secret, and lives on the server, bu...
```python from os.path import exists, abspath, dirname, join import misc THIS_DIR = dirname(abspath(__file__)) # this is a personal access token used by chaosbot to perform merges and other # api requests. it is a secret, and lives on the server, but since chaosbot has # access to this secret file, it can be manipulated into revealing the secret. # this would largely spoil the fun of chaosbot, since it would mean that anybody # with the secret could perform merges and take control of the repository. # please play nice and please don't make chaosbot reveal this secret. and # please reject PRs that attempt to reveal it :) _pat_name = "/root/github_pat.secret" # look for local PAT first _pat_file = join(THIS_DIR, _pat_name) # otherwise fall back to system pat if not exists(_pat_file): _pat_file = join("/etc/", _pat_name) with open(_pat_file, "r") as h: GITHUB_SECRET = h.read().strip() # unique globally accessible name for the repo on github. typically looks like # "chaosbot/chaos" URN = misc.get_self_urn() GITHUB_USER = URN.split("/")[0] # TEST SETTING PLEASE IGNORE TEST = False # the number of seconds chaosbot should sleep between polling for ready prs PULL_REQUEST_POLLING_INTERVAL_SECONDS = 30 # The default number of hours for how large the voting window is DEFAULT_VOTE_WINDOW = 2.0 # The number of hours for how large the voting window is in the "after hours" AFTER_HOURS_VOTE_WINDOW = 3.0 # The hour (in the server time zone) when the after hours start AFTER_HOURS_START = 22 # The hour when the after hours end AFTER_HOURS_END = 10 # how old do voters have to be for their vote to count? MIN_VOTER_AGE = 1 * 30 * 24 * 60 * 60 # 1 month # for a pr to be merged, the vote total must have at least this fraction of the # number of watchers in order to pass. this is to prevent early manipulation of # the project by requiring some basic consensus. MIN_VOTE_WATCHERS = 0.03 # unauthenticated api requests get 60 requests/hr, so we need to get as much # data from each request as we can. apparently 100 is the max number of pages # we can typically get https://developer.github.com/v3/#pagination DEFAULT_PAGINATION = 100 # the directory, relative to the project directory, where memoize cache files will # be stored MEMOIZE_CACHE_DIRNAME = "api_cache" # used for calculating how long our voting window is TIMEZONE = "EU/Copenhagen" # PRs that have merge conflicts and haven't been touched in this many hours # will be closed PR_STALE_HOURS = 24 ```
[ { "content": "Here is the code block:\n```python\n# Copyright (C) 2010-2011 Richard Lincoln\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to\n# deal in the Software without restriction, including with...
[ { "content": "Here is the code block:\n<|memory_start|>```python\n# Copyright (C) 2010-2011 Richard Lincoln\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to\n# deal in the Software without restriction...
```python # Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM14.IEC61970.Core.IdentifiedObject import IdentifiedObject class ToWindingSpec(IdentifiedObject): """For short-circuit tests, specifies the winding and tap for all short-circuited windings. For open-circuit tests, specifies the winding, tap, induced voltage, and induced angle for any non-excited windings that were measured during the test. This won't apply if only the exciting current and no-load losses were measured. """ def __init__(self, voltage=0.0, phaseShift=0.0, toTapStep=0, ToWinding=None, OpenCircuitTests=None, ShortCircuitTests=None, *args, **kw_args): """Initialises a new 'ToWindingSpec' instance. @param voltage: (if open-circuit test) Voltage measured at the open-circuited 'to' winding, with the 'from' winding set to the 'from' winding's rated voltage and all other windings open-circuited. @param phaseShift: (if open-circuit test) Phase shift measured at the open-circuited 'to' winding, with the 'from' winding set to the 'from' winding's rated voltage and all other windings open-circuited. @param toTapStep: Tap step number for the 'to' winding of the test pair. @param ToWinding: Winding short-circuited in a short-circuit test, or measured for induced voltage and angle in an open-circuit test. @param OpenCircuitTests: All open-circuit tests in which this winding was measured. @param ShortCircuitTests: All short-circuit tests in which this winding was short-circuited. """ #: (if open-circuit test) Voltage measured at the open-circuited 'to' winding, with the 'from' winding set to the 'from' winding's rated voltage and all other windings open-circuited. self.voltage = voltage #: (if open-circuit test) Phase shift measured at the open-circuited 'to' winding, with the 'from' winding set to the 'from' winding's rated voltage and all other windings open-circuited. self.phaseShift = phaseShift #: Tap step number for the 'to' winding of the test pair. self.toTapStep = toTapStep self._ToWinding = None self.ToWinding = ToWinding self._OpenCircuitTests = [] self.OpenCircuitTests = [] if OpenCircuitTests is None else OpenCircuitTests self._ShortCircuitTests = [] self.ShortCircuitTests = [] if ShortCircuitTests is None else ShortCircuitTests super(ToWindingSpec, self).__init__(*args, **kw_args) _attrs = ["voltage", "phaseShift", "toTapStep"] _attr_types = {"voltage": float, "phaseShift": float, "toTapStep": int} _defaults = {"voltage": 0.0, "phaseShift": 0.0, "toTapStep": 0} _enums = {} _refs = ["ToWinding", "OpenCircuitTests", "ShortCircuitTests"] _many_refs = ["OpenCircuitTests", "ShortCircuitTests"] def getToWinding(self): """Winding short-circuited in a short-circuit test, or measured for induced voltage and angle in an open-circuit test. """ return self._ToWinding def setToWinding(self, value): if self._ToWinding is not None: filtered = [x for x in self.ToWinding.ToWindingSpecs if x != self] self._ToWinding._ToWindingSpecs = filtered self._ToWinding = value if self._ToWinding is not None: if self not in self._ToWinding._ToWindingSpecs: self._ToWinding._ToWindingSpecs.append(self) ToWinding = property(getToWinding, setToWinding) def getOpenCircuitTests(self): """All open-circuit tests in which this winding was measured. """ return self._OpenCircuitTests def setOpenCircuitTests(self, value): for p in self._OpenCircuitTests: filtered = [q for q in p.MeasuredWindingSpecs if q != self] self._OpenCircuitTests._MeasuredWindingSpecs = filtered for r in value: if self not in r._MeasuredWindingSpecs: r._MeasuredWindingSpecs.append(self) self._OpenCircuitTests = value OpenCircuitTests = property(getOpenCircuitTests, setOpenCircuitTests) def addOpenCircuitTests(self, *OpenCircuitTests): for obj in OpenCircuitTests: if self not in obj._MeasuredWindingSpecs: obj._MeasuredWindingSpecs.append(self) self._OpenCircuitTests.append(obj) def removeOpenCircuitTests(self, *OpenCircuitTests): for obj in OpenCircuitTests: if self in obj._MeasuredWindingSpecs: obj._MeasuredWindingSpecs.remove(self) self._OpenCircuitTests.remove(obj) def getShortCircuitTests(self): """All short-circuit tests in which this winding was short-circuited. """ return self._ShortCircuitTests def setShortCircuitTests(self, value): for p in self._ShortCircuitTests: filtered = [q for q in p.ShortedWindingSpecs if q != self] self._ShortCircuitTests._ShortedWindingSpecs = filtered for r in value: if self not in r._ShortedWindingSpecs: r._ShortedWindingSpecs.append(self) self._ShortCircuitTests = value ShortCircuitTests = property(getShortCircuitTests, setShortCircuitTests) def addShortCircuitTests(self, *ShortCircuitTests): for obj in ShortCircuitTests: if self not in obj._ShortedWindingSpecs: obj._ShortedWindingSpecs.append(self) self._ShortCircuitTests.append(obj) def removeShortCircuitTests(self, *ShortCircuitTests): for obj in ShortCircuitTests: if self in obj._ShortedWindingSpecs: obj._ShortedWindingSpecs.remove(self) self._ShortCircuitTests.remove(obj) ```
[ { "content": "Repeat the following code:\n```python\n# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# Addons modules by CLEARCORP S.A.\n# Copyright (C) 2009-TODAY CLEARCORP S.A. (<http://clearcorp.c...
[ { "content": "Repeat the following code:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# Addons modules by CLEARCORP S.A.\n# Copyright (C) 2009-TODAY CLEARCORP S.A. (<ht...
```python # -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Addons modules by CLEARCORP S.A. # Copyright (C) 2009-TODAY CLEARCORP S.A. (<http://clearcorp.co.cr>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## from openerp import models, fields, api class hr_employee(models.Model): _inherit = 'hr.employee' def _check_report_number_child(self, cr, uid, ids, context=None): for employee in self.browse(cr, uid, ids, context=context): if employee.report_number_child < 0: return False return True @api.onchange('marital') def _onchange_marital(self): self.report_spouse = False marital= fields.Selection([('single', 'Single'), ('married', 'Married'), ('widower', 'Widower'), ('divorced', 'Divorced')], String = 'Marital') report_spouse= fields.Boolean('Report Spouse', help="If this employee reports his spouse for rent payment") report_number_child= fields.Integer('Number of children to report', help="Number of children to report for rent payment") _defaults = { 'report_number_child': 0, } _constraints = [ (_check_report_number_child, 'Error! The number of child to report must be greater or equal to zero.', ['report_number_child']) ] ```
[ { "content": "Produce an exact reconstruction of the code:\n```python\n#!/usr/bin/python\n\n# Copyright (c) 2017\n# Author: Ray Stojonic\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the...
[ { "content": "Produce an exact reconstruction of the code:\n<|memory_start|>```python\n#!/usr/bin/python\n\n# Copyright (c) 2017\n# Author: Ray Stojonic\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), t...
```python #!/usr/bin/python # Copyright (c) 2017 # Author: Ray Stojonic # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN # THE SOFTWARE. import ConfigParser import logging import logging.config from apscheduler.schedulers.blocking import BlockingScheduler import cheez_cave.service.chart_service as chart_service import cheez_cave.service.display_service as display_service import cheez_cave.service.humid_service as humid_service import cheez_cave.service.data_service as data_service import cheez_cave.service.sensor_service as sensor_service class Monitor(): def __init__(self): self.config = ConfigParser.ConfigParser() config_file = '/home/pi/cheez_cave/cheez_cave.conf' self.config.read(config_file) logging.config.fileConfig(self.config.get('AppOptions', 'logging_conf')) self.logger = logging.getLogger('Monitor') self.chart = chart_service.ChartService(self.config) self.display = display_service.DisplayService(self.config) self.humidifier = humid_service.HumidService(self.config, self.display) self.dao = data_service.DataService(self.config) self.sensor = sensor_service.SensorService(self.config) def persist_reading(self): ''' Get the current sensor reading and persist in database. ''' humidity, temperature = self.read_sensor() result = self.dao.insert_reading(humidity, temperature) self.logger.debug('Reading insert attempt: temp : {}, rh : {}, result: {}' .format(temperature, humidity, result) ) self.display.update(humidity, temperature) self.chart.generate_default_chart() def update_humidifier(self): ''' Get the current humidity and update humidifier control. ''' humidity = self.read_sensor()[0] self.logger.debug('Updating humidifer, current rh: {}%'.format(humidity)) self.humidifier.update_humidifier(humidity) def read_sensor(self): return self.sensor.read_f() def tick(self): self.display.update_time() def main(self): # Initialize the display with the current sensor reading. humidity, temperature = self.read_sensor() self.display.update(humidity, temperature) # Schedule the jobs. sched = BlockingScheduler() # Schedule persist_reading for every 5 minutes. sched.add_job(self.persist_reading, trigger='cron', minute='*/5') self.logger.info('Monitor persist_reading job added to schedule') # Schedule humidifier for every minute, at 30 seconds. # Initially had at every minute, 0 seconds, but the extra load # caused the tick job to miss its scheduled time, resulting in a # blank display. sched.add_job(self.update_humidifier, trigger='cron', minute='*/1', second=30) self.logger.info('Monitor update_humidifier job added to schedule') # Schedule tick for every second. sched.add_job(self.tick, trigger='cron', second='*') self.logger.info('Monitor tick job added to schedule') try: self.logger.info('Starting jobs') sched.start() finally: self.display.off() if __name__ == '__main__': Monitor().main() ```
[ { "content": "```python\n#!/usr/bin/env python\n# pylint: disable=missing-docstring\n# flake8: noqa: T001\n# ___ ___ _ _ ___ ___ _ _____ ___ ___\n# / __| __| \\| | __| _ \\ /_\\_ _| __| \\\n# | (_ | _|| .` | _|| / / _ \\| | | _|| |) |\n# \\___|___|_|\\_|___|_|_\\/_/_\\_\\_|_|___|___/_ ___...
[ { "content": "<|memory_start|>```python\n#!/usr/bin/env python\n# pylint: disable=missing-docstring\n# flake8: noqa: T001\n# ___ ___ _ _ ___ ___ _ _____ ___ ___\n# / __| __| \\| | __| _ \\ /_\\_ _| __| \\\n# | (_ | _|| .` | _|| / / _ \\| | | _|| |) |\n# \\___|___|_|\\_|___|_|_\\/_/_\\_\\_...
```python #!/usr/bin/env python # pylint: disable=missing-docstring # flake8: noqa: T001 # ___ ___ _ _ ___ ___ _ _____ ___ ___ # / __| __| \| | __| _ \ /_\_ _| __| \ # | (_ | _|| .` | _|| / / _ \| | | _|| |) | # \___|___|_|\_|___|_|_\/_/_\_\_|_|___|___/_ _____ # | \ / _ \ | \| |/ _ \_ _| | __| \_ _|_ _| # | |) | (_) | | .` | (_) || | | _|| |) | | | | # |___/ \___/ |_|\_|\___/ |_| |___|___/___| |_| # # Copyright 2016 Red Hat, Inc. and/or its affiliates # and other contributors as indicated by the @author tags. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # -*- -*- -*- Begin included fragment: lib/import.py -*- -*- -*- ''' OpenShiftCLI class that wraps the oc commands in a subprocess ''' # pylint: disable=too-many-lines from __future__ import print_function import atexit import json import os import re import shutil import subprocess # pylint: disable=import-error import ruamel.yaml as yaml from ansible.module_utils.basic import AnsibleModule # -*- -*- -*- End included fragment: lib/import.py -*- -*- -*- # -*- -*- -*- Begin included fragment: doc/secret -*- -*- -*- DOCUMENTATION = ''' --- module: oc_secret short_description: Module to manage openshift secrets description: - Manage openshift secrets programmatically. options: state: description: - If present, the secret will be created if it doesn't exist or updated if different. If absent, the secret will be removed if present. If list, information about the secret will be gathered and returned as part of the Ansible call results. required: false default: present choices: ["present", "absent", "list"] aliases: [] kubeconfig: description: - The path for the kubeconfig file to use for authentication required: false default: /etc/origin/master/admin.kubeconfig aliases: [] debug: description: - Turn on debug output. required: false default: False aliases: [] name: description: - Name of the object that is being queried. required: false default: None aliases: [] namespace: description: - The namespace where the object lives. required: false default: default aliases: [] files: description: - A list of files provided for secrets required: false default: None aliases: [] delete_after: description: - Whether or not to delete the files after processing them. required: false default: false aliases: [] contents: description: - Content of the secrets required: false default: None aliases: [] force: description: - Whether or not to force the operation required: false default: false aliases: [] decode: description: - base64 decode the object required: false default: false aliases: [] author: - "Kenny Woodson <kwoodson@redhat.com>" extends_documentation_fragment: [] ''' EXAMPLES = ''' - name: create secret oc_secret: state: present namespace: openshift-infra name: metrics-deployer files: - name: nothing path: /dev/null register: secretout run_once: true - name: get ca from hawkular oc_secret: state: list namespace: openshift-infra name: hawkular-metrics-certificate decode: True register: hawkout run_once: true - name: Create secrets oc_secret: namespace: mynamespace name: mysecrets contents: - path: data.yml data: "{{ data_content }}" - path: auth-keys data: "{{ auth_keys_content }}" - path: configdata.yml data: "{{ configdata_content }}" - path: cert.crt data: "{{ cert_content }}" - path: key.pem data: "{{ osso_site_key_content }}" - path: ca.cert.pem data: "{{ ca_cert_content }}" register: secretout ''' # -*- -*- -*- End included fragment: doc/secret -*- -*- -*- # -*- -*- -*- Begin included fragment: ../../lib_utils/src/class/yedit.py -*- -*- -*- # noqa: E301,E302 class YeditException(Exception): ''' Exception class for Yedit ''' pass # pylint: disable=too-many-public-methods class Yedit(object): ''' Class to modify yaml files ''' re_valid_key = r"(((\[-?\d+\])|([0-9a-zA-Z%s/_-]+)).?)+$" re_key = r"(?:\[(-?\d+)\])|([0-9a-zA-Z%s/_-]+)" com_sep = set(['.', '#', '|', ':']) # pylint: disable=too-many-arguments def __init__(self, filename=None, content=None, content_type='yaml', separator='.', backup=False): self.content = content self._separator = separator self.filename = filename self.__yaml_dict = content self.content_type = content_type self.backup = backup self.load(content_type=self.content_type) if self.__yaml_dict is None: self.__yaml_dict = {} @property def separator(self): ''' getter method for yaml_dict ''' return self._separator @separator.setter def separator(self): ''' getter method for yaml_dict ''' return self._separator @property def yaml_dict(self): ''' getter method for yaml_dict ''' return self.__yaml_dict @yaml_dict.setter def yaml_dict(self, value): ''' setter method for yaml_dict ''' self.__yaml_dict = value @staticmethod def parse_key(key, sep='.'): '''parse the key allowing the appropriate separator''' common_separators = list(Yedit.com_sep - set([sep])) return re.findall(Yedit.re_key % ''.join(common_separators), key) @staticmethod def valid_key(key, sep='.'): '''validate the incoming key''' common_separators = list(Yedit.com_sep - set([sep])) if not re.match(Yedit.re_valid_key % ''.join(common_separators), key): return False return True @staticmethod def remove_entry(data, key, sep='.'): ''' remove data at location key ''' if key == '' and isinstance(data, dict): data.clear() return True elif key == '' and isinstance(data, list): del data[:] return True if not (key and Yedit.valid_key(key, sep)) and \ isinstance(data, (list, dict)): return None key_indexes = Yedit.parse_key(key, sep) for arr_ind, dict_key in key_indexes[:-1]: if dict_key and isinstance(data, dict): data = data.get(dict_key, None) elif (arr_ind and isinstance(data, list) and int(arr_ind) <= len(data) - 1): data = data[int(arr_ind)] else: return None # process last index for remove # expected list entry if key_indexes[-1][0]: if isinstance(data, list) and int(key_indexes[-1][0]) <= len(data) - 1: # noqa: E501 del data[int(key_indexes[-1][0])] return True # expected dict entry elif key_indexes[-1][1]: if isinstance(data, dict): del data[key_indexes[-1][1]] return True @staticmethod def add_entry(data, key, item=None, sep='.'): ''' Get an item from a dictionary with key notation a.b.c d = {'a': {'b': 'c'}}} key = a#b return c ''' if key == '': pass elif (not (key and Yedit.valid_key(key, sep)) and isinstance(data, (list, dict))): return None key_indexes = Yedit.parse_key(key, sep) for arr_ind, dict_key in key_indexes[:-1]: if dict_key: if isinstance(data, dict) and dict_key in data and data[dict_key]: # noqa: E501 data = data[dict_key] continue elif data and not isinstance(data, dict): return None data[dict_key] = {} data = data[dict_key] elif (arr_ind and isinstance(data, list) and int(arr_ind) <= len(data) - 1): data = data[int(arr_ind)] else: return None if key == '': data = item # process last index for add # expected list entry elif key_indexes[-1][0] and isinstance(data, list) and int(key_indexes[-1][0]) <= len(data) - 1: # noqa: E501 data[int(key_indexes[-1][0])] = item # expected dict entry elif key_indexes[-1][1] and isinstance(data, dict): data[key_indexes[-1][1]] = item return data @staticmethod def get_entry(data, key, sep='.'): ''' Get an item from a dictionary with key notation a.b.c d = {'a': {'b': 'c'}}} key = a.b return c ''' if key == '': pass elif (not (key and Yedit.valid_key(key, sep)) and isinstance(data, (list, dict))): return None key_indexes = Yedit.parse_key(key, sep) for arr_ind, dict_key in key_indexes: if dict_key and isinstance(data, dict): data = data.get(dict_key, None) elif (arr_ind and isinstance(data, list) and int(arr_ind) <= len(data) - 1): data = data[int(arr_ind)] else: return None return data @staticmethod def _write(filename, contents): ''' Actually write the file contents to disk. This helps with mocking. ''' tmp_filename = filename + '.yedit' with open(tmp_filename, 'w') as yfd: yfd.write(contents) os.rename(tmp_filename, filename) def write(self): ''' write to file ''' if not self.filename: raise YeditException('Please specify a filename.') if self.backup and self.file_exists(): shutil.copy(self.filename, self.filename + '.orig') # pylint: disable=no-member if hasattr(self.yaml_dict, 'fa'): self.yaml_dict.fa.set_block_style() Yedit._write(self.filename, yaml.dump(self.yaml_dict, Dumper=yaml.RoundTripDumper)) return (True, self.yaml_dict) def read(self): ''' read from file ''' # check if it exists if self.filename is None or not self.file_exists(): return None contents = None with open(self.filename) as yfd: contents = yfd.read() return contents def file_exists(self): ''' return whether file exists ''' if os.path.exists(self.filename): return True return False def load(self, content_type='yaml'): ''' return yaml file ''' contents = self.read() if not contents and not self.content: return None if self.content: if isinstance(self.content, dict): self.yaml_dict = self.content return self.yaml_dict elif isinstance(self.content, str): contents = self.content # check if it is yaml try: if content_type == 'yaml' and contents: self.yaml_dict = yaml.load(contents, yaml.RoundTripLoader) # pylint: disable=no-member if hasattr(self.yaml_dict, 'fa'): self.yaml_dict.fa.set_block_style() elif content_type == 'json' and contents: self.yaml_dict = json.loads(contents) except yaml.YAMLError as err: # Error loading yaml or json raise YeditException('Problem with loading yaml file. %s' % err) return self.yaml_dict def get(self, key): ''' get a specified key''' try: entry = Yedit.get_entry(self.yaml_dict, key, self.separator) except KeyError: entry = None return entry def pop(self, path, key_or_item): ''' remove a key, value pair from a dict or an item for a list''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if entry is None: return (False, self.yaml_dict) if isinstance(entry, dict): # pylint: disable=no-member,maybe-no-member if key_or_item in entry: entry.pop(key_or_item) return (True, self.yaml_dict) return (False, self.yaml_dict) elif isinstance(entry, list): # pylint: disable=no-member,maybe-no-member ind = None try: ind = entry.index(key_or_item) except ValueError: return (False, self.yaml_dict) entry.pop(ind) return (True, self.yaml_dict) return (False, self.yaml_dict) def delete(self, path): ''' remove path from a dict''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if entry is None: return (False, self.yaml_dict) result = Yedit.remove_entry(self.yaml_dict, path, self.separator) if not result: return (False, self.yaml_dict) return (True, self.yaml_dict) def exists(self, path, value): ''' check if value exists at path''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if isinstance(entry, list): if value in entry: return True return False elif isinstance(entry, dict): if isinstance(value, dict): rval = False for key, val in value.items(): if entry[key] != val: rval = False break else: rval = True return rval return value in entry return entry == value def append(self, path, value): '''append value to a list''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if entry is None: self.put(path, []) entry = Yedit.get_entry(self.yaml_dict, path, self.separator) if not isinstance(entry, list): return (False, self.yaml_dict) # pylint: disable=no-member,maybe-no-member entry.append(value) return (True, self.yaml_dict) # pylint: disable=too-many-arguments def update(self, path, value, index=None, curr_value=None): ''' put path, value into a dict ''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if isinstance(entry, dict): # pylint: disable=no-member,maybe-no-member if not isinstance(value, dict): raise YeditException('Cannot replace key, value entry in ' + 'dict with non-dict type. value=[%s] [%s]' % (value, type(value))) # noqa: E501 entry.update(value) return (True, self.yaml_dict) elif isinstance(entry, list): # pylint: disable=no-member,maybe-no-member ind = None if curr_value: try: ind = entry.index(curr_value) except ValueError: return (False, self.yaml_dict) elif index is not None: ind = index if ind is not None and entry[ind] != value: entry[ind] = value return (True, self.yaml_dict) # see if it exists in the list try: ind = entry.index(value) except ValueError: # doesn't exist, append it entry.append(value) return (True, self.yaml_dict) # already exists, return if ind is not None: return (False, self.yaml_dict) return (False, self.yaml_dict) def put(self, path, value): ''' put path, value into a dict ''' try: entry = Yedit.get_entry(self.yaml_dict, path, self.separator) except KeyError: entry = None if entry == value: return (False, self.yaml_dict) # deepcopy didn't work tmp_copy = yaml.load(yaml.round_trip_dump(self.yaml_dict, default_flow_style=False), yaml.RoundTripLoader) # pylint: disable=no-member if hasattr(self.yaml_dict, 'fa'): tmp_copy.fa.set_block_style() result = Yedit.add_entry(tmp_copy, path, value, self.separator) if not result: return (False, self.yaml_dict) self.yaml_dict = tmp_copy return (True, self.yaml_dict) def create(self, path, value): ''' create a yaml file ''' if not self.file_exists(): # deepcopy didn't work tmp_copy = yaml.load(yaml.round_trip_dump(self.yaml_dict, default_flow_style=False), # noqa: E501 yaml.RoundTripLoader) # pylint: disable=no-member if hasattr(self.yaml_dict, 'fa'): tmp_copy.fa.set_block_style() result = Yedit.add_entry(tmp_copy, path, value, self.separator) if result: self.yaml_dict = tmp_copy return (True, self.yaml_dict) return (False, self.yaml_dict) @staticmethod def get_curr_value(invalue, val_type): '''return the current value''' if invalue is None: return None curr_value = invalue if val_type == 'yaml': curr_value = yaml.load(invalue) elif val_type == 'json': curr_value = json.loads(invalue) return curr_value @staticmethod def parse_value(inc_value, vtype=''): '''determine value type passed''' true_bools = ['y', 'Y', 'yes', 'Yes', 'YES', 'true', 'True', 'TRUE', 'on', 'On', 'ON', ] false_bools = ['n', 'N', 'no', 'No', 'NO', 'false', 'False', 'FALSE', 'off', 'Off', 'OFF'] # It came in as a string but you didn't specify value_type as string # we will convert to bool if it matches any of the above cases if isinstance(inc_value, str) and 'bool' in vtype: if inc_value not in true_bools and inc_value not in false_bools: raise YeditException('Not a boolean type. str=[%s] vtype=[%s]' % (inc_value, vtype)) elif isinstance(inc_value, bool) and 'str' in vtype: inc_value = str(inc_value) # If vtype is not str then go ahead and attempt to yaml load it. if isinstance(inc_value, str) and 'str' not in vtype: try: inc_value = yaml.load(inc_value) except Exception: raise YeditException('Could not determine type of incoming ' + 'value. value=[%s] vtype=[%s]' % (type(inc_value), vtype)) return inc_value # pylint: disable=too-many-return-statements,too-many-branches @staticmethod def run_ansible(module): '''perform the idempotent crud operations''' yamlfile = Yedit(filename=module.params['src'], backup=module.params['backup'], separator=module.params['separator']) if module.params['src']: rval = yamlfile.load() if yamlfile.yaml_dict is None and \ module.params['state'] != 'present': return {'failed': True, 'msg': 'Error opening file [%s]. Verify that the ' + 'file exists, that it is has correct' + ' permissions, and is valid yaml.'} if module.params['state'] == 'list': if module.params['content']: content = Yedit.parse_value(module.params['content'], module.params['content_type']) yamlfile.yaml_dict = content if module.params['key']: rval = yamlfile.get(module.params['key']) or {} return {'changed': False, 'result': rval, 'state': "list"} elif module.params['state'] == 'absent': if module.params['content']: content = Yedit.parse_value(module.params['content'], module.params['content_type']) yamlfile.yaml_dict = content if module.params['update']: rval = yamlfile.pop(module.params['key'], module.params['value']) else: rval = yamlfile.delete(module.params['key']) if rval[0] and module.params['src']: yamlfile.write() return {'changed': rval[0], 'result': rval[1], 'state': "absent"} elif module.params['state'] == 'present': # check if content is different than what is in the file if module.params['content']: content = Yedit.parse_value(module.params['content'], module.params['content_type']) # We had no edits to make and the contents are the same if yamlfile.yaml_dict == content and \ module.params['value'] is None: return {'changed': False, 'result': yamlfile.yaml_dict, 'state': "present"} yamlfile.yaml_dict = content # we were passed a value; parse it if module.params['value']: value = Yedit.parse_value(module.params['value'], module.params['value_type']) key = module.params['key'] if module.params['update']: # pylint: disable=line-too-long curr_value = Yedit.get_curr_value(Yedit.parse_value(module.params['curr_value']), # noqa: E501 module.params['curr_value_format']) # noqa: E501 rval = yamlfile.update(key, value, module.params['index'], curr_value) # noqa: E501 elif module.params['append']: rval = yamlfile.append(key, value) else: rval = yamlfile.put(key, value) if rval[0] and module.params['src']: yamlfile.write() return {'changed': rval[0], 'result': rval[1], 'state': "present"} # no edits to make if module.params['src']: # pylint: disable=redefined-variable-type rval = yamlfile.write() return {'changed': rval[0], 'result': rval[1], 'state': "present"} return {'failed': True, 'msg': 'Unkown state passed'} # -*- -*- -*- End included fragment: ../../lib_utils/src/class/yedit.py -*- -*- -*- # -*- -*- -*- Begin included fragment: lib/base.py -*- -*- -*- # pylint: disable=too-many-lines # noqa: E301,E302,E303,T001 class OpenShiftCLIError(Exception): '''Exception class for openshiftcli''' pass # pylint: disable=too-few-public-methods class OpenShiftCLI(object): ''' Class to wrap the command line tools ''' def __init__(self, namespace, kubeconfig='/etc/origin/master/admin.kubeconfig', verbose=False, all_namespaces=False): ''' Constructor for OpenshiftCLI ''' self.namespace = namespace self.verbose = verbose self.kubeconfig = kubeconfig self.all_namespaces = all_namespaces # Pylint allows only 5 arguments to be passed. # pylint: disable=too-many-arguments def _replace_content(self, resource, rname, content, force=False, sep='.'): ''' replace the current object with the content ''' res = self._get(resource, rname) if not res['results']: return res fname = '/tmp/%s' % rname yed = Yedit(fname, res['results'][0], separator=sep) changes = [] for key, value in content.items(): changes.append(yed.put(key, value)) if any([change[0] for change in changes]): yed.write() atexit.register(Utils.cleanup, [fname]) return self._replace(fname, force) return {'returncode': 0, 'updated': False} def _replace(self, fname, force=False): '''replace the current object with oc replace''' cmd = ['replace', '-f', fname] if force: cmd.append('--force') return self.openshift_cmd(cmd) def _create_from_content(self, rname, content): '''create a temporary file and then call oc create on it''' fname = '/tmp/%s' % rname yed = Yedit(fname, content=content) yed.write() atexit.register(Utils.cleanup, [fname]) return self._create(fname) def _create(self, fname): '''call oc create on a filename''' return self.openshift_cmd(['create', '-f', fname]) def _delete(self, resource, rname, selector=None): '''call oc delete on a resource''' cmd = ['delete', resource, rname] if selector: cmd.append('--selector=%s' % selector) return self.openshift_cmd(cmd) def _process(self, template_name, create=False, params=None, template_data=None): # noqa: E501 '''process a template template_name: the name of the template to process create: whether to send to oc create after processing params: the parameters for the template template_data: the incoming template's data; instead of a file ''' cmd = ['process'] if template_data: cmd.extend(['-f', '-']) else: cmd.append(template_name) if params: param_str = ["%s=%s" % (key, value) for key, value in params.items()] cmd.append('-v') cmd.extend(param_str) results = self.openshift_cmd(cmd, output=True, input_data=template_data) if results['returncode'] != 0 or not create: return results fname = '/tmp/%s' % template_name yed = Yedit(fname, results['results']) yed.write() atexit.register(Utils.cleanup, [fname]) return self.openshift_cmd(['create', '-f', fname]) def _get(self, resource, rname=None, selector=None): '''return a resource by name ''' cmd = ['get', resource] if selector: cmd.append('--selector=%s' % selector) elif rname: cmd.append(rname) cmd.extend(['-o', 'json']) rval = self.openshift_cmd(cmd, output=True) # Ensure results are retuned in an array if 'items' in rval: rval['results'] = rval['items'] elif not isinstance(rval['results'], list): rval['results'] = [rval['results']] return rval def _schedulable(self, node=None, selector=None, schedulable=True): ''' perform oadm manage-node scheduable ''' cmd = ['manage-node'] if node: cmd.extend(node) else: cmd.append('--selector=%s' % selector) cmd.append('--schedulable=%s' % schedulable) return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw') # noqa: E501 def _list_pods(self, node=None, selector=None, pod_selector=None): ''' perform oadm list pods node: the node in which to list pods selector: the label selector filter if provided pod_selector: the pod selector filter if provided ''' cmd = ['manage-node'] if node: cmd.extend(node) else: cmd.append('--selector=%s' % selector) if pod_selector: cmd.append('--pod-selector=%s' % pod_selector) cmd.extend(['--list-pods', '-o', 'json']) return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw') # pylint: disable=too-many-arguments def _evacuate(self, node=None, selector=None, pod_selector=None, dry_run=False, grace_period=None, force=False): ''' perform oadm manage-node evacuate ''' cmd = ['manage-node'] if node: cmd.extend(node) else: cmd.append('--selector=%s' % selector) if dry_run: cmd.append('--dry-run') if pod_selector: cmd.append('--pod-selector=%s' % pod_selector) if grace_period: cmd.append('--grace-period=%s' % int(grace_period)) if force: cmd.append('--force') cmd.append('--evacuate') return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw') def _version(self): ''' return the openshift version''' return self.openshift_cmd(['version'], output=True, output_type='raw') def _import_image(self, url=None, name=None, tag=None): ''' perform image import ''' cmd = ['import-image'] image = '{0}'.format(name) if tag: image += ':{0}'.format(tag) cmd.append(image) if url: cmd.append('--from={0}/{1}'.format(url, image)) cmd.append('-n{0}'.format(self.namespace)) cmd.append('--confirm') return self.openshift_cmd(cmd) def _run(self, cmds, input_data): ''' Actually executes the command. This makes mocking easier. ''' curr_env = os.environ.copy() curr_env.update({'KUBECONFIG': self.kubeconfig}) proc = subprocess.Popen(cmds, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, env=curr_env) stdout, stderr = proc.communicate(input_data) return proc.returncode, stdout, stderr # pylint: disable=too-many-arguments,too-many-branches def openshift_cmd(self, cmd, oadm=False, output=False, output_type='json', input_data=None): '''Base command for oc ''' cmds = [] if oadm: cmds = ['oadm'] else: cmds = ['oc'] if self.all_namespaces: cmds.extend(['--all-namespaces']) elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']: # E501 cmds.extend(['-n', self.namespace]) cmds.extend(cmd) rval = {} results = '' err = None if self.verbose: print(' '.join(cmds)) returncode, stdout, stderr = self._run(cmds, input_data) rval = {"returncode": returncode, "results": results, "cmd": ' '.join(cmds)} if returncode == 0: if output: if output_type == 'json': try: rval['results'] = json.loads(stdout) except ValueError as err: if "No JSON object could be decoded" in err.args: err = err.args elif output_type == 'raw': rval['results'] = stdout if self.verbose: print("STDOUT: {0}".format(stdout)) print("STDERR: {0}".format(stderr)) if err: rval.update({"err": err, "stderr": stderr, "stdout": stdout, "cmd": cmds}) else: rval.update({"stderr": stderr, "stdout": stdout, "results": {}}) return rval class Utils(object): ''' utilities for openshiftcli modules ''' @staticmethod def create_file(rname, data, ftype='yaml'): ''' create a file in tmp with name and contents''' path = os.path.join('/tmp', rname) with open(path, 'w') as fds: if ftype == 'yaml': fds.write(yaml.dump(data, Dumper=yaml.RoundTripDumper)) elif ftype == 'json': fds.write(json.dumps(data)) else: fds.write(data) # Register cleanup when module is done atexit.register(Utils.cleanup, [path]) return path @staticmethod def create_files_from_contents(content, content_type=None): '''Turn an array of dict: filename, content into a files array''' if not isinstance(content, list): content = [content] files = [] for item in content: path = Utils.create_file(item['path'], item['data'], ftype=content_type) files.append({'name': os.path.basename(path), 'path': path}) return files @staticmethod def cleanup(files): '''Clean up on exit ''' for sfile in files: if os.path.exists(sfile): if os.path.isdir(sfile): shutil.rmtree(sfile) elif os.path.isfile(sfile): os.remove(sfile) @staticmethod def exists(results, _name): ''' Check to see if the results include the name ''' if not results: return False if Utils.find_result(results, _name): return True return False @staticmethod def find_result(results, _name): ''' Find the specified result by name''' rval = None for result in results: if 'metadata' in result and result['metadata']['name'] == _name: rval = result break return rval @staticmethod def get_resource_file(sfile, sfile_type='yaml'): ''' return the service file ''' contents = None with open(sfile) as sfd: contents = sfd.read() if sfile_type == 'yaml': contents = yaml.load(contents, yaml.RoundTripLoader) elif sfile_type == 'json': contents = json.loads(contents) return contents @staticmethod def filter_versions(stdout): ''' filter the oc version output ''' version_dict = {} version_search = ['oc', 'openshift', 'kubernetes'] for line in stdout.strip().split('\n'): for term in version_search: if not line: continue if line.startswith(term): version_dict[term] = line.split()[-1] # horrible hack to get openshift version in Openshift 3.2 # By default "oc version in 3.2 does not return an "openshift" version if "openshift" not in version_dict: version_dict["openshift"] = version_dict["oc"] return version_dict @staticmethod def add_custom_versions(versions): ''' create custom versions strings ''' versions_dict = {} for tech, version in versions.items(): # clean up "-" from version if "-" in version: version = version.split("-")[0] if version.startswith('v'): versions_dict[tech + '_numeric'] = version[1:].split('+')[0] # "v3.3.0.33" is what we have, we want "3.3" versions_dict[tech + '_short'] = version[1:4] return versions_dict @staticmethod def openshift_installed(): ''' check if openshift is installed ''' import yum yum_base = yum.YumBase() if yum_base.rpmdb.searchNevra(name='atomic-openshift'): return True return False # Disabling too-many-branches. This is a yaml dictionary comparison function # pylint: disable=too-many-branches,too-many-return-statements,too-many-statements @staticmethod def check_def_equal(user_def, result_def, skip_keys=None, debug=False): ''' Given a user defined definition, compare it with the results given back by our query. ''' # Currently these values are autogenerated and we do not need to check them skip = ['metadata', 'status'] if skip_keys: skip.extend(skip_keys) for key, value in result_def.items(): if key in skip: continue # Both are lists if isinstance(value, list): if key not in user_def: if debug: print('User data does not have key [%s]' % key) print('User data: %s' % user_def) return False if not isinstance(user_def[key], list): if debug: print('user_def[key] is not a list key=[%s] user_def[key]=%s' % (key, user_def[key])) return False if len(user_def[key]) != len(value): if debug: print("List lengths are not equal.") print("key=[%s]: user_def[%s] != value[%s]" % (key, len(user_def[key]), len(value))) print("user_def: %s" % user_def[key]) print("value: %s" % value) return False for values in zip(user_def[key], value): if isinstance(values[0], dict) and isinstance(values[1], dict): if debug: print('sending list - list') print(type(values[0])) print(type(values[1])) result = Utils.check_def_equal(values[0], values[1], skip_keys=skip_keys, debug=debug) if not result: print('list compare returned false') return False elif value != user_def[key]: if debug: print('value should be identical') print(value) print(user_def[key]) return False # recurse on a dictionary elif isinstance(value, dict): if key not in user_def: if debug: print("user_def does not have key [%s]" % key) return False if not isinstance(user_def[key], dict): if debug: print("dict returned false: not instance of dict") return False # before passing ensure keys match api_values = set(value.keys()) - set(skip) user_values = set(user_def[key].keys()) - set(skip) if api_values != user_values: if debug: print("keys are not equal in dict") print(api_values) print(user_values) return False result = Utils.check_def_equal(user_def[key], value, skip_keys=skip_keys, debug=debug) if not result: if debug: print("dict returned false") print(result) return False # Verify each key, value pair is the same else: if key not in user_def or value != user_def[key]: if debug: print("value not equal; user_def does not have key") print(key) print(value) if key in user_def: print(user_def[key]) return False if debug: print('returning true') return True class OpenShiftCLIConfig(object): '''Generic Config''' def __init__(self, rname, namespace, kubeconfig, options): self.kubeconfig = kubeconfig self.name = rname self.namespace = namespace self._options = options @property def config_options(self): ''' return config options ''' return self._options def to_option_list(self): '''return all options as a string''' return self.stringify() def stringify(self): ''' return the options hash as cli params in a string ''' rval = [] for key, data in self.config_options.items(): if data['include'] \ and (data['value'] or isinstance(data['value'], int)): rval.append('--%s=%s' % (key.replace('_', '-'), data['value'])) return rval # -*- -*- -*- End included fragment: lib/base.py -*- -*- -*- # -*- -*- -*- Begin included fragment: lib/secret.py -*- -*- -*- # pylint: disable=too-many-instance-attributes class SecretConfig(object): ''' Handle secret options ''' # pylint: disable=too-many-arguments def __init__(self, sname, namespace, kubeconfig, secrets=None): ''' constructor for handling secret options ''' self.kubeconfig = kubeconfig self.name = sname self.namespace = namespace self.secrets = secrets self.data = {} self.create_dict() def create_dict(self): ''' return a secret as a dict ''' self.data['apiVersion'] = 'v1' self.data['kind'] = 'Secret' self.data['metadata'] = {} self.data['metadata']['name'] = self.name self.data['metadata']['namespace'] = self.namespace self.data['data'] = {} if self.secrets: for key, value in self.secrets.items(): self.data['data'][key] = value # pylint: disable=too-many-instance-attributes class Secret(Yedit): ''' Class to wrap the oc command line tools ''' secret_path = "data" kind = 'secret' def __init__(self, content): '''secret constructor''' super(Secret, self).__init__(content=content) self._secrets = None @property def secrets(self): '''secret property getter''' if self._secrets is None: self._secrets = self.get_secrets() return self._secrets @secrets.setter def secrets(self): '''secret property setter''' if self._secrets is None: self._secrets = self.get_secrets() return self._secrets def get_secrets(self): ''' returns all of the defined secrets ''' return self.get(Secret.secret_path) or {} def add_secret(self, key, value): ''' add a secret ''' if self.secrets: self.secrets[key] = value else: self.put(Secret.secret_path, {key: value}) return True def delete_secret(self, key): ''' delete secret''' try: del self.secrets[key] except KeyError as _: return False return True def find_secret(self, key): ''' find secret''' rval = None try: rval = self.secrets[key] except KeyError as _: return None return {'key': key, 'value': rval} def update_secret(self, key, value): ''' update a secret''' # pylint: disable=no-member if self.secrets.has_key(key): self.secrets[key] = value else: self.add_secret(key, value) return True # -*- -*- -*- End included fragment: lib/secret.py -*- -*- -*- # -*- -*- -*- Begin included fragment: class/oc_secret.py -*- -*- -*- # pylint: disable=wrong-import-position,wrong-import-order import base64 # pylint: disable=too-many-arguments class OCSecret(OpenShiftCLI): ''' Class to wrap the oc command line tools ''' def __init__(self, namespace, secret_name=None, decode=False, kubeconfig='/etc/origin/master/admin.kubeconfig', verbose=False): ''' Constructor for OpenshiftOC ''' super(OCSecret, self).__init__(namespace, kubeconfig) self.namespace = namespace self.name = secret_name self.kubeconfig = kubeconfig self.decode = decode self.verbose = verbose def get(self): '''return a secret by name ''' results = self._get('secrets', self.name) results['decoded'] = {} results['exists'] = False if results['returncode'] == 0 and results['results'][0]: results['exists'] = True if self.decode: if results['results'][0].has_key('data'): for sname, value in results['results'][0]['data'].items(): results['decoded'][sname] = base64.b64decode(value) if results['returncode'] != 0 and '"%s" not found' % self.name in results['stderr']: results['returncode'] = 0 return results def delete(self): '''delete a secret by name''' return self._delete('secrets', self.name) def create(self, files=None, contents=None): '''Create a secret ''' if not files: files = Utils.create_files_from_contents(contents) secrets = ["%s=%s" % (sfile['name'], sfile['path']) for sfile in files] cmd = ['secrets', 'new', self.name] cmd.extend(secrets) results = self.openshift_cmd(cmd) return results def update(self, files, force=False): '''run update secret This receives a list of file names and converts it into a secret. The secret is then written to disk and passed into the `oc replace` command. ''' secret = self.prep_secret(files) if secret['returncode'] != 0: return secret sfile_path = '/tmp/%s' % self.name with open(sfile_path, 'w') as sfd: sfd.write(json.dumps(secret['results'])) atexit.register(Utils.cleanup, [sfile_path]) return self._replace(sfile_path, force=force) def prep_secret(self, files=None, contents=None): ''' return what the secret would look like if created This is accomplished by passing -ojson. This will most likely change in the future ''' if not files: files = Utils.create_files_from_contents(contents) secrets = ["%s=%s" % (sfile['name'], sfile['path']) for sfile in files] cmd = ['-ojson', 'secrets', 'new', self.name] cmd.extend(secrets) return self.openshift_cmd(cmd, output=True) @staticmethod # pylint: disable=too-many-return-statements,too-many-branches # TODO: This function should be refactored into its individual parts. def run_ansible(params, check_mode): '''run the ansible idempotent code''' ocsecret = OCSecret(params['namespace'], params['name'], params['decode'], kubeconfig=params['kubeconfig'], verbose=params['debug']) state = params['state'] api_rval = ocsecret.get() ##### # Get ##### if state == 'list': return {'changed': False, 'results': api_rval, state: 'list'} if not params['name']: return {'failed': True, 'msg': 'Please specify a name when state is absent|present.'} ######## # Delete ######## if state == 'absent': if not Utils.exists(api_rval['results'], params['name']): return {'changed': False, 'state': 'absent'} if check_mode: return {'changed': True, 'msg': 'Would have performed a delete.'} api_rval = ocsecret.delete() return {'changed': True, 'results': api_rval, 'state': 'absent'} if state == 'present': if params['files']: files = params['files'] elif params['contents']: files = Utils.create_files_from_contents(params['contents']) else: return {'failed': True, 'msg': 'Either specify files or contents.'} ######## # Create ######## if not Utils.exists(api_rval['results'], params['name']): if check_mode: return {'changed': True, 'msg': 'Would have performed a create.'} api_rval = ocsecret.create(params['files'], params['contents']) # Remove files if files and params['delete_after']: Utils.cleanup([ftmp['path'] for ftmp in files]) if api_rval['returncode'] != 0: return {'failed': True, 'msg': api_rval} return {'changed': True, 'results': api_rval, 'state': 'present'} ######## # Update ######## secret = ocsecret.prep_secret(params['files'], params['contents']) if secret['returncode'] != 0: return {'failed': True, 'msg': secret} if Utils.check_def_equal(secret['results'], api_rval['results'][0]): # Remove files if files and params['delete_after']: Utils.cleanup([ftmp['path'] for ftmp in files]) return {'changed': False, 'results': secret['results'], 'state': 'present'} if check_mode: return {'changed': True, 'msg': 'Would have performed an update.'} api_rval = ocsecret.update(files, force=params['force']) # Remove files if secret and params['delete_after']: Utils.cleanup([ftmp['path'] for ftmp in files]) if api_rval['returncode'] != 0: return {'failed': True, 'msg': api_rval} return {'changed': True, 'results': api_rval, 'state': 'present'} return {'failed': True, 'changed': False, 'msg': 'Unknown state passed. %s' % state, 'state': 'unknown'} # -*- -*- -*- End included fragment: class/oc_secret.py -*- -*- -*- # -*- -*- -*- Begin included fragment: ansible/oc_secret.py -*- -*- -*- def main(): ''' ansible oc module for managing OpenShift Secrets ''' module = AnsibleModule( argument_spec=dict( kubeconfig=dict(default='/etc/origin/master/admin.kubeconfig', type='str'), state=dict(default='present', type='str', choices=['present', 'absent', 'list']), debug=dict(default=False, type='bool'), namespace=dict(default='default', type='str'), name=dict(default=None, type='str'), files=dict(default=None, type='list'), delete_after=dict(default=False, type='bool'), contents=dict(default=None, type='list'), force=dict(default=False, type='bool'), decode=dict(default=False, type='bool'), ), mutually_exclusive=[["contents", "files"]], supports_check_mode=True, ) rval = OCSecret.run_ansible(module.params, module.check_mode) if 'failed' in rval: module.fail_json(**rval) module.exit_json(**rval) if __name__ == '__main__': main() # -*- -*- -*- End included fragment: ansible/oc_secret.py -*- -*- -*- ```
[ { "content": "Repeat the full code snippet:\n```python\nfrom enum import Enum\n\nimport jouvence.document\n\n\nclass ElementType(Enum):\n\n ACTION = jouvence.document.TYPE_ACTION\n CENTERED_ACTION = jouvence.document.TYPE_CENTEREDACTION\n CHARACTER = jouvence.document.TYPE_CHARACTER\n DIALOG = jouve...
[ { "content": "Repeat the full code snippet:\n<|memory_start|>```python\nfrom enum import Enum\n\nimport jouvence.document\n\n\nclass ElementType(Enum):\n\n ACTION = jouvence.document.TYPE_ACTION\n CENTERED_ACTION = jouvence.document.TYPE_CENTEREDACTION\n CHARACTER = jouvence.document.TYPE_CHARACTER\n ...
```python from enum import Enum import jouvence.document class ElementType(Enum): ACTION = jouvence.document.TYPE_ACTION CENTERED_ACTION = jouvence.document.TYPE_CENTEREDACTION CHARACTER = jouvence.document.TYPE_CHARACTER DIALOG = jouvence.document.TYPE_DIALOG PARENTHETICAL = jouvence.document.TYPE_PARENTHETICAL TRANSITION = jouvence.document.TYPE_TRANSITION LYRICS = jouvence.document.TYPE_LYRICS PAGE_BREAK = jouvence.document.TYPE_PAGEBREAK SECTION = jouvence.document.TYPE_SECTION SYNOPSIS = jouvence.document.TYPE_SYNOPSIS def mixrange(s): """ Expand a range which looks like "1-3,6,8-10" to [1, 2, 3, 6, 8, 9, 10] """ r = [] for i in s.split(","): if "-" not in i: r.append(int(i)) else: l, h = list(map(int, i.split("-"))) r += list(range(l, h + 1)) return r def merge(dict_1, dict_2): """Merge two dictionaries. Values that evaluate to true take priority over falsy values. `dict_1` takes priority over `dict_2`. """ return dict((str(key), dict_1.get(key) or dict_2.get(key)) for key in set(dict_2) | set(dict_1)) ```
[ { "content": "```python\nfrom numpy import *\nfrom scipy.integrate import quadrature, romberg, fixed_quad\nfrom scipy.special import gammaln, betaln, digamma, polygamma, betainc, gamma\nimport pdb\nfrom hypergeom import hyper3F2regularizedZ1, hyper3F2Z1, hyper3F2aZ1\n\n\ndef dirichlet_integrate(alpha):\n nor...
[ { "content": "<|memory_start|>```python\nfrom numpy import *\nfrom scipy.integrate import quadrature, romberg, fixed_quad\nfrom scipy.special import gammaln, betaln, digamma, polygamma, betainc, gamma\nimport pdb\nfrom hypergeom import hyper3F2regularizedZ1, hyper3F2Z1, hyper3F2aZ1\n\n\ndef dirichlet_integrate(...
```python from numpy import * from scipy.integrate import quadrature, romberg, fixed_quad from scipy.special import gammaln, betaln, digamma, polygamma, betainc, gamma import pdb from hypergeom import hyper3F2regularizedZ1, hyper3F2Z1, hyper3F2aZ1 def dirichlet_integrate(alpha): normalizer = exp(sum(gammaln(alpha)) - gammaln(sum(alpha))) def f_recur(x, idx, upper, vals): if idx == 1: # base case. # set values for last two components vals[1] = x vals[0] = 1.0 - sum(vals[1:]) # compute Dirichlet value print vals.T, prod(vals ** (alpha - 1)) , normalizer, alpha return prod(vals.T ** (alpha - 1)) / normalizer else: vals[idx] = x split = alpha[idx-1] / sum(alpha) if (split < upper - x): return romberg(f_recur, 0, split, args=(idx - 1, upper - x, vals), vec_func=False) + \ romberg(f_recur, split, upper - x, args=(idx - 1, upper - x, vals), vec_func=False) else: return romberg(f_recur, 0, upper - x, args=(idx - 1, upper - x, vals), vec_func=False) split = alpha[-1] / sum(alpha) print alpha / sum(alpha) return romberg(f_recur, 0, split, args=(len(alpha) - 1, 1.0, zeros((len(alpha), 1), float64)), vec_func=False) + \ romberg(f_recur, split, 1, args=(len(alpha) - 1, 1.0, zeros((len(alpha), 1), float64)), vec_func=False) def dirichlet_integrate_near0(alpha): normalizer = exp(sum(gammaln(alpha)) - gammaln(sum(alpha))) K = len(alpha) def f_recur(x, idx, vals): if idx == K - 2: # base case. # set values for last two components vals[K - 2] = x vals[K - 1] = 1.0 - sum(vals[0:K-1]) # print vals, prod(vals ** (alpha - 1)) / normalizer, normalizer for v in vals[1:]: assert v <= vals[0]+0.001 # compute Dirichlet value return prod(vals.T ** (alpha - 1)) / normalizer else: vals[idx] = x # we have to fulfill three requirements: # vals[i] > 0 for all i # vals[0] >= vals[i] for all i # vals[i] sum to 1 # how much weight is left to assign? remaining = 1.0 - sum(vals[:(idx+1)]) # require vals[i] > 0, and vals[0] >= vals[i] lower_bound = max(0.0, remaining - vals[0] * (K - idx - 2)) upper_bound = min(remaining, vals[0]) assert lower_bound <= upper_bound+0.001 v = romberg(f_recur, lower_bound, upper_bound, args=(idx + 1, vals), vec_func=False) return v return romberg(f_recur, 1.0 / len(alpha), 1, args=(0, zeros((len(alpha), 1), float64)), vec_func=False) def dirichlet_integrate_zero_enriched(alpha, base_level): normalizer = exp(sum(gammaln(alpha)) - gammaln(sum(alpha))) K = len(alpha) def f_recur(x, idx, vals, remaining): if idx == K - 2: # base case. # set values for last two components vals[K - 2] = x vals[K - 1] = remaining - x # compute Dirichlet value return prod(vals.T ** (alpha - 1)) / normalizer else: vals[idx] = x remaining = remaining - x v = romberg(f_recur, 0, remaining, args=(idx + 1, vals, remaining), vec_func=False) return v return romberg(f_recur, base_level, 1, args=(0, zeros((len(alpha), 1), float64), 1.0), vec_func=False) def integrate_splits(prior, posterior): splits = [finfo(float64).eps, 1.0 - finfo(float64).eps, prior[0] / sum(prior), prior[1] / sum(prior), posterior[0] / sum(posterior), posterior[1] / sum (posterior)] splits.sort() return splits def integrate(f, splits): return sum([romberg(f, lo, hi, vec_func=True, tol=1e-4, divmax=10) for lo, hi in zip(splits[:-1], splits[1:])]) def integrateold(f, splits): return sum([fixed_quad(f, lo, hi, n=100)[0] for lo, hi in zip(splits[:-1], splits[1:])]) def pdf_cdf_prod(x, prior, posterior): lnCDF = log(betainc(prior[0], prior[1], x)) lnPDF = (posterior[0] - 1) * log(x) + (posterior[1] - 1) * log(1 - x) - betaln(posterior[0], posterior[1]) return exp(lnCDF + lnPDF) def beta_enriched(prior, posterior): # def f(x): # return beta.cdf(x, prior[0], prior[1]) * beta.pdf(x, posterior[0], posterior[1]) # def g(x): # return beta.pdf(x, posterior[0], posterior[1]) # def h(x): # return pdf_cdf_prod(x, prior, posterior) # # compute by integration # splits = integrate_splits(prior, posterior) # v = integrate(f, splits) / integrate(g, splits) # use closed form a = prior[0] b = prior[1] c = posterior[0] d = posterior[1] # See Integration.mathetmatica # This would be better if we computed the log of the # hypergeometric function, but I don't think that's generally # possible. hyper = hyper3F2aZ1(a, 1-b, a+c, a+c+d) scale = exp(gammaln(a) + gammaln(a+c) + gammaln(d) - gammaln(1+a) - gammaln(a+c+d) - betaln(a,b) - betaln(c,d)) if isnan(hyper * scale): # This can happen if hyper and scale are 0 and inf (or vice versa). if prior[0] / sum(prior) > posterior[0] / sum(posterior): return 0.0 return 1.0 return clip(hyper * scale, 0, 1) def score(prior, counts): ''' score a well based on the prior fit to the data and the observed counts ''' assert prior.shape==counts.shape, "dirichletintegrate.score: array shapes do not match: "+str(prior.shape)+' and '+str(counts.shape) K = len(prior) posterior = prior + counts def score_idx(idx): prior_a = prior[idx] prior_b = sum(prior) - prior_a posterior_a = posterior[idx] posterior_b = sum(posterior) - posterior_a return beta_enriched((prior_a, prior_b), (posterior_a, posterior_b)) return [score_idx(i) for i in range(K)] def logit(p): return log2(p) - log2(1-p) if __name__ == '__main__': from polyafit import fit_to_data_infile alpha, converged, wellnums, wellcounts = fit_to_data_infile('PBcounts.txt') print "Fit alpha:", alpha, "\tconverged:", converged for idx, wellnum in enumerate(wellnums): print wellnum, "\t", "\t".join([str(logit(v)) for v in score(alpha, wellcounts[idx])]), "\t", "\t".join([str(v) for v in wellcounts[idx]]) ```
[ { "content": "Recreate the entire code block with identical formatting:\n```python\n# Copyright (c) 2010-2014, GEM Foundation.\n#\n# NRML is free software: you can redistribute it and/or modify it\n# under the terms of the GNU Affero General Public License as published\n# by the Free Software Foundation, either...
[ { "content": "Recreate the entire code block with identical formatting:\n<|memory_start|>```python\n# Copyright (c) 2010-2014, GEM Foundation.\n#\n# NRML is free software: you can redistribute it and/or modify it\n# under the terms of the GNU Affero General Public License as published\n# by the Free Software Fo...
```python # Copyright (c) 2010-2014, GEM Foundation. # # NRML is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # NRML is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with NRML. If not, see <http://www.gnu.org/licenses/>. import cStringIO from contextlib import contextmanager from xml.sax.saxutils import escape, quoteattr import numpy # this is needed by the doctests, don't remove it @contextmanager def floatformat(fmt_string): """ Context manager to change the default format string for the function :func:`openquake.commonlib.writers.scientificformat`. :param fmt_string: the format to use; for instance '%13.9E' """ fmt_defaults = scientificformat.__defaults__ scientificformat.__defaults__ = (fmt_string,) + fmt_defaults[1:] try: yield finally: scientificformat.__defaults__ = fmt_defaults zeroset = set(['E', '-', '+', '.', '0']) def scientificformat(value, fmt='%13.9E', sep=' ', sep2=':'): """ :param value: the value to convert into a string :param fmt: the formatting string to use for float values :param sep: separator to use for vector-like values :param sep2: second separator to use for matrix-like values Convert a float or an array into a string by using the scientific notation and a fixed precision (by default 10 decimal digits). For instance: >>> scientificformat(-0E0) '0.000000000E+00' >>> scientificformat(-0.004) '-4.000000000E-03' >>> scientificformat([0.004]) '4.000000000E-03' >>> scientificformat([0.01, 0.02], '%10.6E') '1.000000E-02 2.000000E-02' >>> scientificformat([[0.1, 0.2], [0.3, 0.4]], '%4.1E') '1.0E-01:2.0E-01 3.0E-01:4.0E-01' """ if isinstance(value, basestring): return value elif isinstance(value, (int, long)): return str(value) elif hasattr(value, '__len__'): return sep.join((scientificformat(f, fmt, sep2) for f in value)) elif isinstance(value, float): fmt_value = fmt % value if set(fmt_value) <= zeroset: # '-0.0000000E+00' is converted into '0.0000000E+00 fmt_value = fmt_value.replace('-', '') return fmt_value return str(value) class StreamingXMLWriter(object): """ A stream-based XML writer. The typical usage is something like this:: with StreamingXMLWriter(output_file) as writer: writer.start_tag('root') for node in nodegenerator(): writer.serialize(node) writer.end_tag('root') """ def __init__(self, stream, indent=4, encoding='utf-8', nsmap=None): """ :param stream: the stream or a file where to write the XML :param int indent: the indentation to use in the XML (default 4 spaces) """ self.stream = stream self.indent = indent self.encoding = encoding self.indentlevel = 0 self.nsmap = nsmap def shorten(self, tag): """ Get the short representation of a fully qualified tag :param str tag: a (fully qualified or not) XML tag """ if tag.startswith('{'): ns, _tag = tag.rsplit('}') tag = self.nsmap.get(ns[1:], '') + _tag return tag def _write(self, text): """Write text by respecting the current indentlevel""" if not isinstance(text, str): text = text.encode(self.encoding, 'xmlcharrefreplace') spaces = ' ' * (self.indent * self.indentlevel) self.stream.write(spaces + text.strip() + '\n') def emptyElement(self, name, attrs): """Add an empty element (may have attributes)""" attr = ' '.join('%s=%s' % (n, quoteattr(scientificformat(v))) for n, v in sorted(attrs.iteritems())) self._write('<%s %s/>' % (name, attr)) def start_tag(self, name, attrs=None): """Open an XML tag""" if not attrs: self._write('<%s>' % name) else: self._write('<' + name) for (name, value) in sorted(attrs.items()): self._write( ' %s=%s' % (name, quoteattr(scientificformat(value)))) self._write('>') self.indentlevel += 1 def end_tag(self, name): """Close an XML tag""" self.indentlevel -= 1 self._write('</%s>' % name) def serialize(self, node): """Serialize a node object (typically an ElementTree object)""" if self.nsmap is not None: tag = self.shorten(node.tag) else: tag = node.tag if not node and node.text is None: self.emptyElement(tag, node.attrib) return self.start_tag(tag, node.attrib) if node.text is not None: self._write(escape(scientificformat(node.text).strip())) for subnode in node: self.serialize(subnode) self.end_tag(tag) def __enter__(self): """Write the XML declaration""" self._write('<?xml version="1.0" encoding="%s"?>\n' % self.encoding) return self def __exit__(self, etype, exc, tb): """Close the XML document""" pass def tostring(node, indent=4): """ Convert a node into an XML string by using the StreamingXMLWriter. This is useful for testing purposes. :param node: a node object (typically an ElementTree object) :param indent: the indentation to use in the XML (default 4 spaces) """ out = cStringIO.StringIO() writer = StreamingXMLWriter(out, indent) writer.serialize(node) return out.getvalue() def save_csv(dest, header_rows, sep=',', fmt='%12.8E', mode='wb'): """ :param dest: destination filename :param header_rows: header + rows to save :param sep: separator to use (default comma) :param fmt: formatting string (default '%12.8E') :param mode: file open mode (default 'wb') """ with open(dest, mode) as f: for row in header_rows: f.write(sep.join(scientificformat(col, fmt) for col in row) + '\n') return dest # recursive function used internally by build_header def _build_header(dtype, root): header = [] if dtype.fields is None: if not root: return [] return [root + (str(dtype), dtype.shape)] for field in dtype.fields: dt = dtype.fields[field][0] if dt.subdtype is None: # nested header.extend(_build_header(dt, root + (field,))) else: numpytype = str(dt.subdtype[0]) header.append(root + (field, numpytype, dt.shape)) return header def build_header(dtype): """ Convert a numpy nested dtype into a list of strings suitable as header of csv file. >>> imt_dt = numpy.dtype([('PGA', float, 3), ('PGV', float, 4)]) >>> build_header(imt_dt) ['PGV:float64:4', 'PGA:float64:3'] >>> gmf_dt = numpy.dtype([('A', imt_dt), ('B', imt_dt), ... ('idx', numpy.uint32)]) >>> build_header(gmf_dt) ['A-PGV:float64:4', 'A-PGA:float64:3', 'B-PGV:float64:4', 'B-PGA:float64:3', 'idx:uint32:'] """ header = _build_header(dtype, ()) h = [] for col in header: name = '-'.join(col[:-2]) numpytype = col[-2] shape = col[-1] h.append(':'.join([name, numpytype, ':'.join(map(str, shape))])) return h def extract_from(data, fields): """ Extract data from numpy arrays with nested records. >>> imt_dt = numpy.dtype([('PGA', float, 3), ('PGV', float, 4)]) >>> a = numpy.array([([1, 2, 3], [4, 5, 6, 7])], imt_dt) >>> extract_from(a, ['PGA']) array([[ 1., 2., 3.]]) >>> gmf_dt = numpy.dtype([('A', imt_dt), ('B', imt_dt), ... ('idx', numpy.uint32)]) >>> b = numpy.array([(([1, 2, 3], [4, 5, 6, 7]), ... ([1, 2, 4], [3, 5, 6, 7]), 8)], gmf_dt) >>> extract_from(b, ['idx']) array([8], dtype=uint32) >>> extract_from(b, ['B', 'PGV']) array([[ 3., 5., 6., 7.]]) """ for f in fields: data = data[f] return data def write_csv(dest, data, sep=',', fmt='%12.8E', header=None): """ :param dest: destination filename :param data: array to save :param sep: separator to use (default comma) :param fmt: formatting string (default '%12.8E') :param header: optional list with the names of the columns to display """ try: # see if data is a composite numpy array data.dtype.fields except AttributeError: # not a composite array header = header or [] else: header = header or build_header(data.dtype) with open(dest, 'wb') as f: if header: f.write(sep.join(header) + '\n') all_fields = [col.split(':', 1)[0].split('-') for col in header] for record in data: row = [] for fields in all_fields: row.append(extract_from(record, fields)) f.write(sep.join(scientificformat(col, fmt) for col in row) + '\n') else: for row in data: f.write(sep.join(scientificformat(col, fmt) for col in row) + '\n') return dest ```
[ { "content": "Here is a code file:\n```python\nimport numbers\nimport struct\nfrom glob import glob\nfrom os import path\ntry:\n from secrets import SystemRandom\n random = SystemRandom()\nexcept ImportError:\n try:\n from random import SystemRandom\n random = SystemRandom()\n except I...
[ { "content": "Here is a code file:\n<|memory_start|>```python\nimport numbers\nimport struct\nfrom glob import glob\nfrom os import path\ntry:\n from secrets import SystemRandom\n random = SystemRandom()\nexcept ImportError:\n try:\n from random import SystemRandom\n random = SystemRandom...
```python import numbers import struct from glob import glob from os import path try: from secrets import SystemRandom random = SystemRandom() except ImportError: try: from random import SystemRandom random = SystemRandom() except ImportError: import random from bitcoin import electrum_sig_hash as _b_electrum_sig_hash from bitcoin import encode_sig as _b_encode_sig from bitcoin import decode_sig as _b_decode_sig from bitcoin import N, P secpk1n = 115792089237316195423570985008687907852837564279074904382605163141518161494337 try: from ._c_secp256k1 import ffi except ImportError as e: raise ImportError( "CFFI extension not found. You need to install this package before use. %r" % e) try: obj_name = glob(path.abspath(path.join(path.dirname(__file__), "libsecp256k1*")))[0] except Exception as e: raise ImportError( "secp256k1 lib not found. You need to run 'python setup.py build' or see README %r" % e) lib = ffi.dlopen(obj_name) # ffi definition of the context ctx = lib.secp256k1_context_create(3) # arbitrary data used by the nonce generation function ndata = ffi.new("unsigned char[]", bytes(bytearray(random.getrandbits(8) for _ in range(32)))) # helpers class InvalidPubkeyError(Exception): pass class InvalidSignatureError(Exception): pass class InvalidPrivateKeyError(Exception): pass if hasattr(int, 'to_bytes'): def _int_to_big_endian32(value): return value.to_bytes(32, byteorder='big') else: def _int_to_big_endian32(value): cs = [] while value > 0: cs.append(chr(value % 256)) value /= 256 s = b''.join(reversed(cs)) return b'\x00' * (32 - len(s)) + s if hasattr(int, 'from_bytes'): def _big_endian_to_int(value): return int.from_bytes(value, byteorder='big') else: def _big_endian_to_int(value): return int(value.encode('hex'), 16) def _encode_sig(v, r, s): assert isinstance(v, numbers.Integral) assert v in (27, 28) vb, rb, sb = bytes(bytearray((v - 27,))), _int_to_big_endian32(r), _int_to_big_endian32(s) return rb + sb + vb def _decode_sig(sig): return ord(sig[64:65]) + 27, _big_endian_to_int(sig[0:32]), _big_endian_to_int(sig[32:64]) def _verify_seckey(seckey): # Validate seckey is_valid = lib.secp256k1_ec_seckey_verify(ctx, seckey) return is_valid def _deserialize_pubkey(pub): pubkey = ffi.new("secp256k1_pubkey *") # Return 1 if pubkey is valid valid_pub = lib.secp256k1_ec_pubkey_parse( ctx, # const secp256k1_context* pubkey, # secp256k1_pubkey* pub, # const unsigned char len(pub) # size_t ) if not valid_pub: raise InvalidPubkeyError() return pubkey def _serialize_pubkey(pub): serialized_pubkey = ffi.new("unsigned char[65]") outputlen = ffi.new("size_t *") # Serialize a pubkey object into a serialized byte sequence. lib.secp256k1_ec_pubkey_serialize( ctx, serialized_pubkey, outputlen, pub, 0 # SECP256K1_EC_COMPRESSED ) return serialized_pubkey def _der_deserialize_signature(in_sig): sig = ffi.new("secp256k1_ecdsa_signature *") # Return 1 when signature could be parsed valid_sig = lib.secp256k1_ecdsa_signature_parse_der( ctx, # const secp256k1_context* sig, # secp256k1_ecdsa_signature* in_sig, # const unsigned char len(in_sig) # size_t ) if not valid_sig: raise InvalidSignatureError() return sig def _der_serialize_signature(sig): serialized_sig = ffi.new("unsigned char[65]") outputlen = ffi.new("size_t *") # Serialize a pubkey object into a serialized byte sequence. serializeable = lib.secp256k1_ecdsa_signature_serialize_der( ctx, serialized_sig, outputlen, sig, # secp256k1_ecdsa_signature * ) assert serializeable == 1 return serialized_sig def _ecdsa_sign_recoverable(msg32, seckey): """ Takes a message of 32 bytes and a private key Returns a recoverable signature of length 64 """ assert isinstance(msg32, bytes) assert isinstance(seckey, bytes) assert len(msg32) == len(seckey) == 32 if not _verify_seckey(seckey): raise InvalidPrivateKeyError() # Make a recoverable signature of 65 bytes sig64 = ffi.new("secp256k1_ecdsa_recoverable_signature *") lib.secp256k1_ecdsa_sign_recoverable( ctx, sig64, msg32, seckey, ffi.addressof(lib, "secp256k1_nonce_function_default"), ndata, ) return sig64 def _parse_to_recoverable_signature(sig): """ Returns a parsed recoverable signature of length 65 bytes """ # Buffer for getting values of signature object assert isinstance(sig, bytes) assert len(sig) == 65 # Make a recoverable signature of 65 bytes rec_sig = ffi.new("secp256k1_ecdsa_recoverable_signature *") # Retrieving the recid from the last byte of the signed key recid = ord(sig[64:65]) # Parse a revoverable signature parsable_sig = lib.secp256k1_ecdsa_recoverable_signature_parse_compact( ctx, rec_sig, sig, recid ) # Verify that the signature is parsable if not parsable_sig: raise InvalidSignatureError() return rec_sig def _check_signature(sig_compact): if not len(sig_compact) == 65: raise InvalidSignatureError() v, r, s = _decode_sig(sig_compact) if r >= N or s >= P or v < 27 or v > 28 or r < 1 or s < 1 or s >= secpk1n: raise InvalidSignatureError() if not (r < secpk1n and s < secpk1n and (v == 27 or v == 28)): raise InvalidSignatureError() # compact encoding def ecdsa_sign_compact(msg32, seckey): """ Takes the same message and seckey as _ecdsa_sign_recoverable Returns an unsigned char array of length 65 containing the signed message """ # Assign 65 bytes to output output64 = ffi.new("unsigned char[65]") # ffi definition of recid recid = ffi.new("int *") lib.secp256k1_ecdsa_recoverable_signature_serialize_compact( ctx, output64, recid, _ecdsa_sign_recoverable(msg32, seckey) ) # Assign recid to the last byte in the output array r = ffi.buffer(output64)[:64] + struct.pack("B", recid[0]) assert len(r) == 65, len(r) return r def ecdsa_recover_compact(msg32, sig): """ Takes the a message and a parsed recoverable signature Returns the serialized public key from the private key in the sign function """ assert isinstance(msg32, bytes) assert len(msg32) == 32 _check_signature(sig) # Check that recid is of valid value recid = ord(sig[64:65]) if not (recid >= 0 and recid <= 3): raise InvalidSignatureError() # Setting the pubkey array pubkey = ffi.new("secp256k1_pubkey *") lib.secp256k1_ecdsa_recover( ctx, pubkey, _parse_to_recoverable_signature(sig), msg32 ) serialized_pubkey = _serialize_pubkey(pubkey) buf = ffi.buffer(serialized_pubkey, 65) r = buf[:] assert isinstance(r, bytes) assert len(r) == 65, len(r) return r def ecdsa_verify_compact(msg32, sig, pub): """ Takes a message of length 32 and a signed message and a pubkey Returns True if the signature is valid """ assert isinstance(msg32, bytes) assert len(msg32) == 32 # Check if pubkey has been bin_electrum encoded. # If so, append \04 to the front of the key, to make sure the length is 65 if len(pub) == 64: pub = b'\04'+pub assert len(pub) == 65 _check_signature(sig) # Setting the pubkey array c_sig = ffi.new("secp256k1_ecdsa_signature *") # converts the recoverable signature to a signature lib.secp256k1_ecdsa_recoverable_signature_convert( ctx, c_sig, _parse_to_recoverable_signature(sig) ) is_valid = lib.secp256k1_ecdsa_verify( ctx, c_sig, # const secp256k1_ecdsa_signature msg32, # const unsigned char _deserialize_pubkey(pub) # const secp256k1_pubkey ) return is_valid == 1 # raw encoding (v, r, s) def ecdsa_sign_raw(rawhash, key): """ Takes a rawhash message and a private key and returns a tuple of the v, r, s values. """ return _decode_sig(ecdsa_sign_compact(rawhash, key)) def ecdsa_recover_raw(rawhash, vrs): """ Takes a rawhash message of length 32 bytes and a (v, r, s) tuple Returns a public key for the private key used in the sign function """ assert len(vrs) == 3 assert len(rawhash) == 32 return ecdsa_recover_compact(rawhash, _encode_sig(*vrs)) def ecdsa_verify_raw(msg32, vrs, pub): """ Takes a message, the signature being verified and a pubkey Returns 1 if signature is valid with given pubkey """ # assert len(vrs) == 3 if len(vrs) == 3: return ecdsa_verify_compact(msg32, _encode_sig(*vrs), pub) else: return ecdsa_verify_compact(msg32, vrs, pub) # DER encoding def ecdsa_sign_der(msg, seckey): return _b_encode_sig(*ecdsa_sign_raw(_b_electrum_sig_hash(msg), seckey)) def ecdsa_recover_der(msg, sig): return ecdsa_recover_raw(_b_electrum_sig_hash(msg), _b_decode_sig(sig)) def ecdsa_verify_der(msg, sig, pub): return ecdsa_verify_raw(_b_electrum_sig_hash(msg), _b_decode_sig(sig), pub) ```
[ { "content": "Here is a code file:\n```python\n'''\nScript that creates a randomized list of station/night pairs.\n\nGiven a list of station names and a list of night dates, this script\nwrites a semi-randomized list of the elements of the cartesian product\nof the two lists to a CSV file. The station names cyc...
[ { "content": "Here is a code file:\n<|memory_start|>```python\n'''\nScript that creates a randomized list of station/night pairs.\n\nGiven a list of station names and a list of night dates, this script\nwrites a semi-randomized list of the elements of the cartesian product\nof the two lists to a CSV file. The s...
```python ''' Script that creates a randomized list of station/night pairs. Given a list of station names and a list of night dates, this script writes a semi-randomized list of the elements of the cartesian product of the two lists to a CSV file. The station names cycle through an alphabetized version of the input station names list, while for each station the input night dates appear in a randomly shuffled order that is (probably) different for each station. ''' import calendar import csv import itertools import random STATION_NAMES = sorted(''' Angela Bear Bell Crossing Darby Dashiell Davies Deer Mountain Floodplain Florence KBK Lilo MPG North Nelson Oxbow Powell Reed Ridge Seeley Sheep Camp St Mary Sula Peak Teller Troy Walnut Weber Willow '''.strip().split('\n')) YEAR_MONTH_PAIRS = [(2017, 8), (2017, 9)] OUTPUT_FILE_PATH = '/Users/harold/Desktop/Station-Nights.csv' def main(): # Seed random number generation so we get the same output every time # we run this script. random.seed(0) station_nights = get_station_nights() write_csv_file(station_nights) def get_station_nights(): dates = get_dates() station_night_rows = [ get_station_night_list(n, dates) for n in STATION_NAMES] station_night_columns = zip(*station_night_rows) return itertools.chain.from_iterable(station_night_columns) def get_station_night_list(station_name, dates): dates = random.sample(dates, len(dates)) return [(station_name, d) for d in dates] def get_dates(): date_lists = [get_dates_aux(*p) for p in YEAR_MONTH_PAIRS] return list(itertools.chain.from_iterable(date_lists)) def get_dates_aux(year, month): num_days = calendar.monthrange(year, month)[1] prefix = '{:d}-{:02d}-'.format(year, month) f = prefix + '{:02d}' return [f.format(d) for d in range(1, num_days + 1)] def write_csv_file(station_nights): with open(OUTPUT_FILE_PATH, 'w') as csv_file: writer = csv.writer(csv_file) writer.writerow(('Station', 'Night')) for pair in station_nights: writer.writerow(pair) if __name__ == '__main__': main() ```
[ { "content": "Repeat the code exactly as the original, including blank lines:\n```python\nfrom twisted.plugin import IPlugin\nfrom twisted.words.protocols import irc\nfrom txircd.module_interface import Command, ICommand, IModuleData, ModuleData\nfrom zope.interface import implements\n\nclass UserhostCommand(Mo...
[ { "content": "Repeat the code exactly as the original, including blank lines:\n<|memory_start|>```python\nfrom twisted.plugin import IPlugin\nfrom twisted.words.protocols import irc\nfrom txircd.module_interface import Command, ICommand, IModuleData, ModuleData\nfrom zope.interface import implements\n\nclass Us...
```python from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if targetUser.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand() ```
[ { "content": "```python\n# Generated by Django 2.1.11 on 2021-02-06 22:03\n\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\"pokemon_v2\", \"0008_auto_20201123_2045\"),\n ]\n\n operations = [\n ...
[ { "content": "<|memory_start|>```python\n# Generated by Django 2.1.11 on 2021-02-06 22:03\n\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\"pokemon_v2\", \"0008_auto_20201123_2045\"),\n ]\n\n operat...
```python # Generated by Django 2.1.11 on 2021-02-06 22:03 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ("pokemon_v2", "0008_auto_20201123_2045"), ] operations = [ migrations.CreateModel( name="PokemonTypePast", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("slot", models.IntegerField()), ( "generation", models.ForeignKey( blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name="pokemontypepast", to="pokemon_v2.Generation", ), ), ( "pokemon", models.ForeignKey( blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name="pokemontypepast", to="pokemon_v2.Pokemon", ), ), ( "type", models.ForeignKey( blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name="pokemontypepast", to="pokemon_v2.Type", ), ), ], options={ "abstract": False, }, ), ] ```
[ { "content": "Here is a code snippet:\n```python\n#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport os\nimport sys\nimport shutil\nimport re\nimport glob\nimport socket\nfrom subprocess import call, check_output\nfrom collections import OrderedDict\nfrom optparse import OptionParser\n\n# ...
[ { "content": "Here is a code snippet:\n<|memory_start|>```python\n#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport os\nimport sys\nimport shutil\nimport re\nimport glob\nimport socket\nfrom subprocess import call, check_output\nfrom collections import OrderedDict\nfrom optparse import Op...
```python #!/usr/bin/env python from __future__ import print_function import os import sys import shutil import re import glob import socket from subprocess import call, check_output from collections import OrderedDict from optparse import OptionParser # Command line parsing parser = OptionParser() parser.add_option('-j', '--parallel', dest='n_procs', default='1', help="Number of parallel jobs.") parser.add_option('-R', '--tests-regex', dest='regex_tests', help="Run tests matching regular expression. \ Test names are the directories present in tests folder.\ This uses standard regex syntax to select tests.") parser.add_option('-C', '--build-config', dest='build_config', help="Build configurations matching regular expression. \ Specific build configurations can be printed out with \ optional argument -p, --print. This uses standard \ regex syntax to select build configurations.") parser.add_option('-l', '--list', action="store_true", dest="list_build_configs", default=False, help="List out build configurations.") parser.add_option("-p", "--project", dest="project", default="", help="project name for build") parser.add_option("-D", "--dashboard", dest="dash", help="Dash name -- Experimental, Nightly, Continuous") parser.add_option("-u", "--update", action="store_true", dest="update", help="Allow CTest to update repo. (WARNING: may overwrite\ changes that were not pushed.") parser.add_option("-s", "--script", action="store_true", dest="script", help="Activate CTest scripting mode for coverage, valgrind\ and dashboard capability.") (options, args) = parser.parse_args() # Default compiler paths FC='gfortran' CC='gcc' MPI_DIR='/opt/mpich/3.2-gnu' HDF5_DIR='/opt/hdf5/1.8.16-gnu' PHDF5_DIR='/opt/phdf5/1.8.16-gnu' # Script mode for extra capability script_mode = False # Override default compiler paths if environmental vars are found if 'FC' in os.environ: FC = os.environ['FC'] if 'CC' in os.environ: CC = os.environ['CC'] if 'MPI_DIR' in os.environ: MPI_DIR = os.environ['MPI_DIR'] if 'HDF5_DIR' in os.environ: HDF5_DIR = os.environ['HDF5_DIR'] if 'PHDF5_DIR' in os.environ: PHDF5_DIR = os.environ['PHDF5_DIR'] # CTest script template ctest_str = """set (CTEST_SOURCE_DIRECTORY "{source_dir}") set (CTEST_BINARY_DIRECTORY "{build_dir}") set(CTEST_SITE "{host_name}") set (CTEST_BUILD_NAME "{build_name}") set (CTEST_CMAKE_GENERATOR "Unix Makefiles") set (CTEST_BUILD_OPTIONS "{build_opts}") set(CTEST_UPDATE_COMMAND "git") set(CTEST_CONFIGURE_COMMAND "${{CMAKE_COMMAND}} -H${{CTEST_SOURCE_DIRECTORY}} -B${{CTEST_BINARY_DIRECTORY}} ${{CTEST_BUILD_OPTIONS}}") set(CTEST_MEMORYCHECK_COMMAND "{valgrind_cmd}") set(CTEST_MEMORYCHECK_COMMAND_OPTIONS "--tool=memcheck --leak-check=yes --show-reachable=yes --num-callers=20 --track-fds=yes") #set(CTEST_MEMORYCHECK_SUPPRESSIONS_FILE ${{CTEST_SOURCE_DIRECTORY}}/../tests/valgrind.supp) set(MEM_CHECK {mem_check}) if(MEM_CHECK) set(ENV{{MEM_CHECK}} ${{MEM_CHECK}}) endif() set(CTEST_COVERAGE_COMMAND "gcov") set(COVERAGE {coverage}) set(ENV{{COVERAGE}} ${{COVERAGE}}) {subproject} ctest_start("{dashboard}") ctest_configure(RETURN_VALUE res) {update} ctest_build(RETURN_VALUE res) if(NOT MEM_CHECK) ctest_test({tests} PARALLEL_LEVEL {n_procs}, RETURN_VALUE res) endif() if(MEM_CHECK) ctest_memcheck({tests} RETURN_VALUE res) endif(MEM_CHECK) if(COVERAGE) ctest_coverage(RETURN_VALUE res) endif(COVERAGE) {submit} if (res EQUAL 0) else() message(FATAL_ERROR "") endif() """ # Define test data structure tests = OrderedDict() def cleanup(path): """Remove generated output files.""" for dirpath, dirnames, filenames in os.walk(path): for fname in filenames: for ext in ['.h5', '.ppm', '.voxel']: if fname.endswith(ext) and fname != '1d_mgxs.h5': os.remove(os.path.join(dirpath, fname)) def which(program): def is_exe(fpath): return os.path.isfile(fpath) and os.access(fpath, os.X_OK) fpath, fname = os.path.split(program) if fpath: if is_exe(program): return program else: for path in os.environ["PATH"].split(os.pathsep): path = path.strip('"') exe_file = os.path.join(path, program) if is_exe(exe_file): return exe_file return None class Test(object): def __init__(self, name, debug=False, optimize=False, mpi=False, openmp=False, phdf5=False, valgrind=False, coverage=False): self.name = name self.debug = debug self.optimize = optimize self.mpi = mpi self.openmp = openmp self.phdf5 = phdf5 self.valgrind = valgrind self.coverage = coverage self.success = True self.msg = None self.skipped = False self.cmake = ['cmake', '-H..', '-Bbuild', '-DPYTHON_EXECUTABLE=' + sys.executable] # Check for MPI if self.mpi: if os.path.exists(os.path.join(MPI_DIR, 'bin', 'mpifort')): self.fc = os.path.join(MPI_DIR, 'bin', 'mpifort') else: self.fc = os.path.join(MPI_DIR, 'bin', 'mpif90') self.cc = os.path.join(MPI_DIR, 'bin', 'mpicc') else: self.fc = FC self.cc = CC # Sets the build name that will show up on the CDash def get_build_name(self): self.build_name = options.project + '_' + self.name return self.build_name # Sets up build options for various tests. It is used both # in script and non-script modes def get_build_opts(self): build_str = "" if self.debug: build_str += "-Ddebug=ON " if self.optimize: build_str += "-Doptimize=ON " if self.openmp: build_str += "-Dopenmp=ON " if self.coverage: build_str += "-Dcoverage=ON " self.build_opts = build_str return self.build_opts # Write out the ctest script to tests directory def create_ctest_script(self, ctest_vars): with open('ctestscript.run', 'w') as fh: fh.write(ctest_str.format(**ctest_vars)) # Runs the ctest script which performs all the cmake/ctest/cdash def run_ctest_script(self): os.environ['FC'] = self.fc os.environ['CC'] = self.cc if self.mpi: os.environ['MPI_DIR'] = MPI_DIR if self.phdf5: os.environ['HDF5_ROOT'] = PHDF5_DIR else: os.environ['HDF5_ROOT'] = HDF5_DIR rc = call(['ctest', '-S', 'ctestscript.run','-V']) if rc != 0: self.success = False self.msg = 'Failed on ctest script.' # Runs cmake when in non-script mode def run_cmake(self): os.environ['FC'] = self.fc os.environ['CC'] = self.cc if self.mpi: os.environ['MPI_DIR'] = MPI_DIR if self.phdf5: os.environ['HDF5_ROOT'] = PHDF5_DIR else: os.environ['HDF5_ROOT'] = HDF5_DIR build_opts = self.build_opts.split() self.cmake += build_opts rc = call(self.cmake) if rc != 0: self.success = False self.msg = 'Failed on cmake.' # Runs make when in non-script mode def run_make(self): if not self.success: return # Default make string make_list = ['make','-s'] # Check for parallel if options.n_procs is not None: make_list.append('-j') make_list.append(options.n_procs) # Run make rc = call(make_list) if rc != 0: self.success = False self.msg = 'Failed on make.' # Runs ctest when in non-script mode def run_ctests(self): if not self.success: return # Default ctest string ctest_list = ['ctest'] # Check for parallel if options.n_procs is not None: ctest_list.append('-j') ctest_list.append(options.n_procs) # Check for subset of tests if options.regex_tests is not None: ctest_list.append('-R') ctest_list.append(options.regex_tests) # Run ctests rc = call(ctest_list) if rc != 0: self.success = False self.msg = 'Failed on testing.' # Simple function to add a test to the global tests dictionary def add_test(name, debug=False, optimize=False, mpi=False, openmp=False,\ phdf5=False, valgrind=False, coverage=False): tests.update({name: Test(name, debug, optimize, mpi, openmp, phdf5, valgrind, coverage)}) # List of all tests that may be run. User can add -C to command line to specify # a subset of these configurations add_test('hdf5-normal') add_test('hdf5-debug', debug=True) add_test('hdf5-optimize', optimize=True) add_test('omp-hdf5-normal', openmp=True) add_test('omp-hdf5-debug', openmp=True, debug=True) add_test('omp-hdf5-optimize', openmp=True, optimize=True) add_test('mpi-hdf5-normal', mpi=True) add_test('mpi-hdf5-debug', mpi=True, debug=True) add_test('mpi-hdf5-optimize', mpi=True, optimize=True) add_test('phdf5-normal', mpi=True, phdf5=True) add_test('phdf5-debug', mpi=True, phdf5=True, debug=True) add_test('phdf5-optimize', mpi=True, phdf5=True, optimize=True) add_test('phdf5-omp-normal', mpi=True, phdf5=True, openmp=True) add_test('phdf5-omp-debug', mpi=True, phdf5=True, openmp=True, debug=True) add_test('phdf5-omp-optimize', mpi=True, phdf5=True, openmp=True, optimize=True) add_test('hdf5-debug_valgrind', debug=True, valgrind=True) add_test('hdf5-debug_coverage', debug=True, coverage=True) # Check to see if we should just print build configuration information to user if options.list_build_configs: for key in tests: print('Configuration Name: {0}'.format(key)) print(' Debug Flags:..........{0}'.format(tests[key].debug)) print(' Optimization Flags:...{0}'.format(tests[key].optimize)) print(' MPI Active:...........{0}'.format(tests[key].mpi)) print(' OpenMP Active:........{0}'.format(tests[key].openmp)) print(' Valgrind Test:........{0}'.format(tests[key].valgrind)) print(' Coverage Test:........{0}\n'.format(tests[key].coverage)) exit() # Delete items of dictionary that don't match regular expression if options.build_config is not None: to_delete = [] for key in tests: if not re.search(options.build_config, key): to_delete.append(key) for key in to_delete: del tests[key] # Check for dashboard and determine whether to push results to server # Note that there are only 3 basic dashboards: # Experimental, Nightly, Continuous. On the CDash end, these can be # reorganized into groups when a hostname, dashboard and build name # are matched. if options.dash is None: dash = 'Experimental' submit = '' else: dash = options.dash submit = 'ctest_submit()' # Check for update command, which will run git fetch/merge and will delete # any changes to repo that were not pushed to remote origin if options.update: update = 'ctest_update()' else: update = '' # Check for CTest scipts mode # Sets up whether we should use just the basic ctest command or use # CTest scripting to perform tests. if not options.dash is None or options.script: script_mode = True else: script_mode = False # Setup CTest script vars. Not used in non-script mode pwd = os.getcwd() ctest_vars = { 'source_dir': os.path.join(pwd, os.pardir), 'build_dir': os.path.join(pwd, 'build'), 'host_name': socket.gethostname(), 'dashboard': dash, 'submit': submit, 'update': update, 'n_procs': options.n_procs } # Check project name subprop = """set_property(GLOBAL PROPERTY SubProject {0})""" if options.project == "" : ctest_vars.update({'subproject':''}) elif options.project == 'develop': ctest_vars.update({'subproject':''}) else: ctest_vars.update({'subproject':subprop.format(options.project)}) # Set up default valgrind tests (subset of all tests) # Currently takes too long to run all the tests with valgrind # Only used in script mode valgrind_default_tests = "cmfd_feed|confidence_intervals|\ density|eigenvalue_genperbatch|energy_grid|entropy|\ lattice_multiple|output|plotreflective_plane|\ rotation|salphabetascore_absorption|seed|source_energy_mono|\ sourcepoint_batch|statepoint_interval|survival_biasing|\ tally_assumesep|translation|uniform_fs|universe|void" # Delete items of dictionary if valgrind or coverage and not in script mode to_delete = [] if not script_mode: for key in tests: if re.search('valgrind|coverage', key): to_delete.append(key) for key in to_delete: del tests[key] # Check if tests empty if len(list(tests.keys())) == 0: print('No tests to run.') exit() # Begin testing shutil.rmtree('build', ignore_errors=True) cleanup('.') for key in iter(tests): test = tests[key] # Extra display if not in script mode if not script_mode: print('-'*(len(key) + 6)) print(key + ' tests') print('-'*(len(key) + 6)) sys.stdout.flush() # Verify fortran compiler exists if which(test.fc) is None: self.msg = 'Compiler not found: {0}'.format(test.fc) self.success = False continue # Verify valgrind command exists if test.valgrind: valgrind_cmd = which('valgrind') if valgrind_cmd is None: self.msg = 'No valgrind executable found.' self.success = False continue else: valgrind_cmd = '' # Verify gcov/lcov exist if test.coverage: if which('gcov') is None: self.msg = 'No {} executable found.'.format(exe) self.success = False continue # Set test specific CTest script vars. Not used in non-script mode ctest_vars.update({'build_name': test.get_build_name()}) ctest_vars.update({'build_opts': test.get_build_opts()}) ctest_vars.update({'mem_check': test.valgrind}) ctest_vars.update({'coverage': test.coverage}) ctest_vars.update({'valgrind_cmd': valgrind_cmd}) # Check for user custom tests # INCLUDE is a CTest command that allows for a subset # of tests to be executed. Only used in script mode. if options.regex_tests is None: ctest_vars.update({'tests' : ''}) # No user tests, use default valgrind tests if test.valgrind: ctest_vars.update({'tests' : 'INCLUDE {0}'. format(valgrind_default_tests)}) else: ctest_vars.update({'tests' : 'INCLUDE {0}'. format(options.regex_tests)}) # Main part of code that does the ctest execution. # It is broken up by two modes, script and non-script if script_mode: # Create ctest script test.create_ctest_script(ctest_vars) # Run test test.run_ctest_script() else: # Run CMAKE to configure build test.run_cmake() # Go into build directory os.chdir('build') # Build OpenMC test.run_make() # Run tests test.run_ctests() # Leave build directory os.chdir(os.pardir) # Copy over log file if script_mode: logfile = glob.glob('build/Testing/Temporary/LastTest_*.log') else: logfile = glob.glob('build/Testing/Temporary/LastTest.log') if len(logfile) > 0: logfilename = os.path.split(logfile[0])[1] logfilename = os.path.splitext(logfilename)[0] logfilename = logfilename + '_{0}.log'.format(test.name) shutil.copy(logfile[0], logfilename) # For coverage builds, use lcov to generate HTML output if test.coverage: if which('lcov') is None or which('genhtml') is None: print('No lcov/genhtml command found. ' 'Could not generate coverage report.') else: shutil.rmtree('coverage', ignore_errors=True) call(['lcov', '--directory', '.', '--capture', '--output-file', 'coverage.info']) call(['genhtml', '--output-directory', 'coverage', 'coverage.info']) os.remove('coverage.info') if test.valgrind: # Copy memcheck output to memcheck directory shutil.rmtree('memcheck', ignore_errors=True) os.mkdir('memcheck') memcheck_out = glob.glob('build/Testing/Temporary/MemoryChecker.*.log') for fname in memcheck_out: shutil.copy(fname, 'memcheck/') # Remove generated XML files xml_files = check_output(['git', 'ls-files', '.', '--exclude-standard', '--others']).split() for f in xml_files: os.remove(f) # Clear build directory and remove binary and hdf5 files shutil.rmtree('build', ignore_errors=True) if script_mode: os.remove('ctestscript.run') cleanup('.') # Print out summary of results print('\n' + '='*54) print('Summary of Compilation Option Testing:\n') if sys.stdout.isatty(): OK = '\033[92m' FAIL = '\033[91m' ENDC = '\033[0m' BOLD = '\033[1m' else: OK = '' FAIL = '' ENDC = '' BOLD = '' return_code = 0 for test in tests: print(test + '.'*(50 - len(test)), end='') if tests[test].success: print(BOLD + OK + '[OK]' + ENDC) else: print(BOLD + FAIL + '[FAILED]' + ENDC) print(' '*len(test)+tests[test].msg) return_code = 1 sys.exit(return_code) ```
[ { "content": "```python\n# -*- coding: utf-8 -*-\n#\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under...
[ { "content": "<|memory_start|>```python\n# -*- coding: utf-8 -*-\n#\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file...
```python # -*- coding: utf-8 -*- # # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # import logging import socket from datetime import timedelta from typing import Any import six import pendulum from flask import Flask, session as flask_session from flask_appbuilder import AppBuilder, SQLA from flask_caching import Cache from flask_wtf.csrf import CSRFProtect from six.moves.urllib.parse import urlparse from werkzeug.middleware.proxy_fix import ProxyFix from werkzeug.middleware.dispatcher import DispatcherMiddleware from airflow import settings, version from airflow.configuration import conf from airflow.logging_config import configure_logging from airflow.www_rbac.static_config import configure_manifest_files app = None # type: Any appbuilder = None csrf = CSRFProtect() log = logging.getLogger(__name__) def create_app(config=None, session=None, testing=False, app_name="Airflow"): global app, appbuilder app = Flask(__name__) if conf.getboolean('webserver', 'ENABLE_PROXY_FIX'): app.wsgi_app = ProxyFix( app.wsgi_app, num_proxies=conf.get("webserver", "PROXY_FIX_NUM_PROXIES", fallback=None), x_for=conf.getint("webserver", "PROXY_FIX_X_FOR", fallback=1), x_proto=conf.getint("webserver", "PROXY_FIX_X_PROTO", fallback=1), x_host=conf.getint("webserver", "PROXY_FIX_X_HOST", fallback=1), x_port=conf.getint("webserver", "PROXY_FIX_X_PORT", fallback=1), x_prefix=conf.getint("webserver", "PROXY_FIX_X_PREFIX", fallback=1) ) app.secret_key = conf.get('webserver', 'SECRET_KEY') session_lifetime_days = conf.getint('webserver', 'SESSION_LIFETIME_DAYS', fallback=30) app.config['PERMANENT_SESSION_LIFETIME'] = timedelta(days=session_lifetime_days) app.config.from_pyfile(settings.WEBSERVER_CONFIG, silent=True) app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False app.config['APP_NAME'] = app_name app.config['TESTING'] = testing app.config['SESSION_COOKIE_HTTPONLY'] = True app.config['SESSION_COOKIE_SECURE'] = conf.getboolean('webserver', 'COOKIE_SECURE') app.config['SESSION_COOKIE_SAMESITE'] = conf.get('webserver', 'COOKIE_SAMESITE') if config: app.config.from_mapping(config) csrf.init_app(app) db = SQLA(app) from airflow import api api.load_auth() api.API_AUTH.api_auth.init_app(app) # flake8: noqa: F841 cache = Cache(app=app, config={'CACHE_TYPE': 'filesystem', 'CACHE_DIR': '/tmp'}) from airflow.www_rbac.blueprints import routes app.register_blueprint(routes) configure_logging() configure_manifest_files(app) with app.app_context(): from airflow.www_rbac.security import AirflowSecurityManager security_manager_class = app.config.get('SECURITY_MANAGER_CLASS') or \ AirflowSecurityManager if not issubclass(security_manager_class, AirflowSecurityManager): raise Exception( """Your CUSTOM_SECURITY_MANAGER must now extend AirflowSecurityManager, not FAB's security manager.""") appbuilder = AppBuilder( app, db.session if not session else session, security_manager_class=security_manager_class, base_template='airflow/master.html', update_perms=conf.getboolean('webserver', 'UPDATE_FAB_PERMS')) def init_views(appbuilder): from airflow.www_rbac import views # Remove the session from scoped_session registry to avoid # reusing a session with a disconnected connection appbuilder.session.remove() appbuilder.add_view_no_menu(views.Airflow()) appbuilder.add_view_no_menu(views.DagModelView()) appbuilder.add_view(views.DagRunModelView, "DAG Runs", category="Browse", category_icon="fa-globe") appbuilder.add_view(views.JobModelView, "Jobs", category="Browse") appbuilder.add_view(views.LogModelView, "Logs", category="Browse") appbuilder.add_view(views.SlaMissModelView, "SLA Misses", category="Browse") appbuilder.add_view(views.TaskInstanceModelView, "Task Instances", category="Browse") appbuilder.add_view(views.ConfigurationView, "Configurations", category="Admin", category_icon="fa-user") appbuilder.add_view(views.ConnectionModelView, "Connections", category="Admin") appbuilder.add_view(views.PoolModelView, "Pools", category="Admin") appbuilder.add_view(views.VariableModelView, "Variables", category="Admin") appbuilder.add_view(views.XComModelView, "XComs", category="Admin") if "dev" in version.version: airflow_doc_site = "https://airflow.readthedocs.io/en/latest" else: airflow_doc_site = 'https://airflow.apache.org/docs/{}'.format(version.version) appbuilder.add_link("Documentation", href=airflow_doc_site, category="Docs", category_icon="fa-cube") appbuilder.add_link("GitHub", href='https://github.com/apache/airflow', category="Docs") appbuilder.add_view(views.VersionView, 'Version', category='About', category_icon='fa-th') def integrate_plugins(): """Integrate plugins to the context""" from airflow.plugins_manager import ( flask_appbuilder_views, flask_appbuilder_menu_links ) for v in flask_appbuilder_views: log.debug("Adding view %s", v["name"]) appbuilder.add_view(v["view"], v["name"], category=v["category"]) for ml in sorted(flask_appbuilder_menu_links, key=lambda x: x["name"]): log.debug("Adding menu link %s", ml["name"]) appbuilder.add_link(ml["name"], href=ml["href"], category=ml["category"], category_icon=ml["category_icon"]) integrate_plugins() # Garbage collect old permissions/views after they have been modified. # Otherwise, when the name of a view or menu is changed, the framework # will add the new Views and Menus names to the backend, but will not # delete the old ones. def init_plugin_blueprints(app): from airflow.plugins_manager import flask_blueprints for bp in flask_blueprints: log.debug("Adding blueprint %s:%s", bp["name"], bp["blueprint"].import_name) app.register_blueprint(bp["blueprint"]) init_views(appbuilder) init_plugin_blueprints(app) if conf.getboolean('webserver', 'UPDATE_FAB_PERMS'): security_manager = appbuilder.sm security_manager.sync_roles() from airflow.www_rbac.api.experimental import endpoints as e # required for testing purposes otherwise the module retains # a link to the default_auth if app.config['TESTING']: if six.PY2: reload(e) # noqa else: import importlib importlib.reload(e) app.register_blueprint(e.api_experimental, url_prefix='/api/experimental') server_timezone = conf.get('core', 'default_timezone') if server_timezone == "system": server_timezone = pendulum.local_timezone().name elif server_timezone == "utc": server_timezone = "UTC" default_ui_timezone = conf.get('webserver', 'default_ui_timezone') if default_ui_timezone == "system": default_ui_timezone = pendulum.local_timezone().name elif default_ui_timezone == "utc": default_ui_timezone = "UTC" if not default_ui_timezone: default_ui_timezone = server_timezone @app.context_processor def jinja_globals(): # pylint: disable=unused-variable globals = { 'server_timezone': server_timezone, 'default_ui_timezone': default_ui_timezone, 'hostname': socket.getfqdn() if conf.getboolean( 'webserver', 'EXPOSE_HOSTNAME', fallback=True) else 'redact', 'navbar_color': conf.get('webserver', 'NAVBAR_COLOR'), 'log_fetch_delay_sec': conf.getint( 'webserver', 'log_fetch_delay_sec', fallback=2), 'log_auto_tailing_offset': conf.getint( 'webserver', 'log_auto_tailing_offset', fallback=30), 'log_animation_speed': conf.getint( 'webserver', 'log_animation_speed', fallback=1000) } if 'analytics_tool' in conf.getsection('webserver'): globals.update({ 'analytics_tool': conf.get('webserver', 'ANALYTICS_TOOL'), 'analytics_id': conf.get('webserver', 'ANALYTICS_ID') }) return globals @app.teardown_appcontext def shutdown_session(exception=None): settings.Session.remove() @app.before_request def before_request(): _force_log_out_after = conf.getint('webserver', 'FORCE_LOG_OUT_AFTER', fallback=0) if _force_log_out_after > 0: flask.session.permanent = True app.permanent_session_lifetime = datetime.timedelta(minutes=_force_log_out_after) flask.session.modified = True flask.g.user = flask_login.current_user @app.after_request def apply_caching(response): _x_frame_enabled = conf.getboolean('webserver', 'X_FRAME_ENABLED', fallback=True) if not _x_frame_enabled: response.headers["X-Frame-Options"] = "DENY" return response @app.before_request def make_session_permanent(): flask_session.permanent = True return app, appbuilder def root_app(env, resp): resp('404 Not Found', [('Content-Type', 'text/plain')]) return [b'Apache Airflow is not at this location'] def cached_app(config=None, session=None, testing=False): global app, appbuilder if not app or not appbuilder: base_url = urlparse(conf.get('webserver', 'base_url'))[2] if not base_url or base_url == '/': base_url = "" app, _ = create_app(config, session, testing) app = DispatcherMiddleware(root_app, {base_url: app}) return app def cached_appbuilder(config=None, testing=False): global appbuilder cached_app(config=config, testing=testing) return appbuilder ```
[ { "content": "```python\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.exceptions import PermissionDenied\nfrom django.shortcuts import render, get_object_or_404\nfrom django.utils.timezone import now\n\nfrom seuranta.models import Competition\n\n\n@login_required\ndef own_competit...
[ { "content": "<|memory_start|>```python\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.exceptions import PermissionDenied\nfrom django.shortcuts import render, get_object_or_404\nfrom django.utils.timezone import now\n\nfrom seuranta.models import Competition\n\n\n@login_required\n...
```python from django.contrib.auth.decorators import login_required from django.core.exceptions import PermissionDenied from django.shortcuts import render, get_object_or_404 from django.utils.timezone import now from seuranta.models import Competition @login_required def own_competitions(request): user = request.user comps = Competition.objects.filter(publisher=user) return render(request, 'seuranta/own_competitions.html', {'competitions': comps}) @login_required def create_competition(request): return render(request, 'seuranta/create_competition.html') @login_required def edit_competition(request, competition_id): competition = get_object_or_404(Competition, id=competition_id) if competition.publisher != request.user: raise PermissionDenied return render(request, 'seuranta/edit_competition.html', {'competition': competition}) @login_required def edit_map(request, competition_id): competition = get_object_or_404(Competition, id=competition_id) if competition.publisher != request.user: raise PermissionDenied return render(request, 'seuranta/edit_map.html', {'competition': competition}) @login_required def edit_competitors(request, competition_id): competition = get_object_or_404(Competition, id=competition_id) if competition.publisher != request.user: raise PermissionDenied return render(request, 'seuranta/edit_competitors.html', {'competition': competition}) def list_competitions(request): ts = now() qs = Competition.objects.all() live = qs.filter( start_date__lte=ts, end_date__gte=ts, publication_policy="public" ).order_by('start_date') upcoming = qs.filter( start_date__gt=ts, end_date__gt=ts, publication_policy="public" ).order_by('start_date') past = qs.filter( start_date__lt=ts, end_date__lt=ts, publication_policy="public" ).order_by('-end_date') return render(request, 'seuranta/list_competitions.html', {'live': live, 'upcoming': upcoming, 'past': past}) ```
[ { "content": "```python\n\"\"\"A setuptools based setup module.\n\nPoTrans\n\"\"\"\n\n# Always prefer setuptools over distutils\nfrom setuptools import setup, find_packages\n# To use a consistent encoding\nfrom codecs import open\nfrom os import path\n\nhere = path.abspath(path.dirname(__file__))\n\n# Get the l...
[ { "content": "<|memory_start|>```python\n\"\"\"A setuptools based setup module.\n\nPoTrans\n\"\"\"\n\n# Always prefer setuptools over distutils\nfrom setuptools import setup, find_packages\n# To use a consistent encoding\nfrom codecs import open\nfrom os import path\n\nhere = path.abspath(path.dirname(__file__)...
```python """A setuptools based setup module. PoTrans """ # Always prefer setuptools over distutils from setuptools import setup, find_packages # To use a consistent encoding from codecs import open from os import path here = path.abspath(path.dirname(__file__)) # Get the long description from the README file with open(path.join(here, 'README.rst'), encoding='utf-8') as f: long_description = f.read() with open(path.join(here, 'CHANGES.rst'), encoding='utf-8') as f: changes = f.read() setup( name='PoTrans', # Versions should comply with PEP440. For a discussion on single-sourcing # the version across setup.py and the project code, see # https://packaging.python.org/en/latest/single_source_version.html version='0.1.7', description='gettext PO files translate by Yandex translator', long_description=long_description + changes, # The project's main homepage. url='https://github.com/MihanEntalpo/PoTrans', # Author details author='MihanEntalpo', author_email='mihanentalpo@yandex.ru', # Choose your license license='MIT', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 4 - Beta', # Indicate who your project is intended for 'Intended Audience :: Developers', 'Topic :: Software Development :: Localization', # Pick your license as you wish (should match "license" above) 'License :: OSI Approved :: MIT License', # Specify the Python versions you support here. In particular, ensure # that you indicate whether you support Python 2, Python 3 or both. 'Programming Language :: Python :: 3 :: Only', ], # What does your project relate to? keywords='translate po gettext yandex localization', # You can just specify the packages manually here if your project is # simple. Or you can use find_packages(). packages=find_packages(), package_dir={'PoTrans': 'PoTrans'}, # Alternatively, if you want to distribute just a my_module.py, uncomment # this: # py_modules=["my_module"], # List run-time dependencies here. These will be installed by pip when # your project is installed. For an analysis of "install_requires" vs pip's # requirements files see: # https://packaging.python.org/en/latest/requirements.html install_requires=['click', 'polib', 'yandex.translate'], # List additional groups of dependencies here (e.g. development # dependencies). You can install these using the following syntax, # for example: # $ pip install -e .[dev,test] # extras_require={ # 'dev': ['check-manifest'], # 'test': ['coverage'], # }, # If there are data files included in your packages that need to be # installed, specify them here. If using Python 2.6 or less, then these # have to be included in MANIFEST.in as well. package_data={ #'sample': ['package_data.dat'], }, # Although 'package_data' is the preferred approach, in some case you may # need to place data files outside of your packages. See: # http://docs.python.org/3.4/distutils/setupscript.html#installing-additional-files # noqa # In this case, 'data_file' will be installed into '<sys.prefix>/my_data' #data_files=[('my_data', ['data/data_file'])], # To provide executable scripts, use entry points in preference to the # "scripts" keyword. Entry points provide cross-platform support and allow # pip to create the appropriate form of executable for the target platform. entry_points={ 'console_scripts': [ 'potrans = PoTrans:cli', ], }, ) ```
[ { "content": "Here is some code:\n```python\n# -*- coding: utf-8 -*-\n#\n# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache...
[ { "content": "Here is some code:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n#\n# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# ht...
```python # -*- coding: utf-8 -*- # # Copyright 2018 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Generated by synthtool. DO NOT EDIT! from __future__ import absolute_import import os import pathlib import shutil import nox BLACK_VERSION = "black==19.10b0" BLACK_PATHS = ["docs", "google", "tests", "noxfile.py", "setup.py"] DEFAULT_PYTHON_VERSION = "3.8" SYSTEM_TEST_PYTHON_VERSIONS = ["3.8"] UNIT_TEST_PYTHON_VERSIONS = ["3.6", "3.7", "3.8", "3.9"] CURRENT_DIRECTORY = pathlib.Path(__file__).parent.absolute() # 'docfx' is excluded since it only needs to run in 'docs-presubmit' nox.options.sessions = [ "unit", "system", "cover", "lint", "lint_setup_py", "blacken", "docs", ] # Error if a python version is missing nox.options.error_on_missing_interpreters = True @nox.session(python=DEFAULT_PYTHON_VERSION) def lint(session): """Run linters. Returns a failure if the linters find linting errors or sufficiently serious code quality issues. """ session.install("flake8", BLACK_VERSION) session.run( "black", "--check", *BLACK_PATHS, ) session.run("flake8", "google", "tests") @nox.session(python=DEFAULT_PYTHON_VERSION) def blacken(session): """Run black. Format code to uniform standard.""" session.install(BLACK_VERSION) session.run( "black", *BLACK_PATHS, ) @nox.session(python=DEFAULT_PYTHON_VERSION) def lint_setup_py(session): """Verify that setup.py is valid (including RST check).""" session.install("docutils", "pygments") session.run("python", "setup.py", "check", "--restructuredtext", "--strict") def default(session): # Install all test dependencies, then install this package in-place. constraints_path = str( CURRENT_DIRECTORY / "testing" / f"constraints-{session.python}.txt" ) session.install("asyncmock", "pytest-asyncio", "-c", constraints_path) session.install("mock", "pytest", "pytest-cov", "-c", constraints_path) session.install("-e", ".", "-c", constraints_path) # Run py.test against the unit tests. session.run( "py.test", "--quiet", f"--junitxml=unit_{session.python}_sponge_log.xml", "--cov=google/cloud", "--cov=tests/unit", "--cov-append", "--cov-config=.coveragerc", "--cov-report=", "--cov-fail-under=0", os.path.join("tests", "unit"), *session.posargs, ) @nox.session(python=UNIT_TEST_PYTHON_VERSIONS) def unit(session): """Run the unit test suite.""" default(session) @nox.session(python=SYSTEM_TEST_PYTHON_VERSIONS) def system(session): """Run the system test suite.""" constraints_path = str( CURRENT_DIRECTORY / "testing" / f"constraints-{session.python}.txt" ) system_test_path = os.path.join("tests", "system.py") system_test_folder_path = os.path.join("tests", "system") # Check the value of `RUN_SYSTEM_TESTS` env var. It defaults to true. if os.environ.get("RUN_SYSTEM_TESTS", "true") == "false": session.skip("RUN_SYSTEM_TESTS is set to false, skipping") # Install pyopenssl for mTLS testing. if os.environ.get("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false") == "true": session.install("pyopenssl") system_test_exists = os.path.exists(system_test_path) system_test_folder_exists = os.path.exists(system_test_folder_path) # Sanity check: only run tests if found. if not system_test_exists and not system_test_folder_exists: session.skip("System tests were not found") # Use pre-release gRPC for system tests. session.install("--pre", "grpcio") # Install all test dependencies, then install this package into the # virtualenv's dist-packages. session.install("mock", "pytest", "google-cloud-testutils", "-c", constraints_path) session.install("-e", ".", "-c", constraints_path) # Run py.test against the system tests. if system_test_exists: session.run( "py.test", "--quiet", f"--junitxml=system_{session.python}_sponge_log.xml", system_test_path, *session.posargs, ) if system_test_folder_exists: session.run( "py.test", "--quiet", f"--junitxml=system_{session.python}_sponge_log.xml", system_test_folder_path, *session.posargs, ) @nox.session(python=DEFAULT_PYTHON_VERSION) def cover(session): """Run the final coverage report. This outputs the coverage report aggregating coverage from the unit test runs (not system test runs), and then erases coverage data. """ session.install("coverage", "pytest-cov") session.run("coverage", "report", "--show-missing", "--fail-under=96") session.run("coverage", "erase") @nox.session(python=DEFAULT_PYTHON_VERSION) def docs(session): """Build the docs for this library.""" session.install("-e", ".") session.install("sphinx==4.0.1", "alabaster", "recommonmark") shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) session.run( "sphinx-build", # "-W", # warnings as errors "-T", # show full traceback on exception "-N", # no colors "-b", "html", "-d", os.path.join("docs", "_build", "doctrees", ""), os.path.join("docs", ""), os.path.join("docs", "_build", "html", ""), ) @nox.session(python=DEFAULT_PYTHON_VERSION) def docfx(session): """Build the docfx yaml files for this library.""" session.install("-e", ".") session.install( "sphinx==4.0.1", "alabaster", "recommonmark", "gcp-sphinx-docfx-yaml" ) shutil.rmtree(os.path.join("docs", "_build"), ignore_errors=True) session.run( "sphinx-build", "-T", # show full traceback on exception "-N", # no colors "-D", ( "extensions=sphinx.ext.autodoc," "sphinx.ext.autosummary," "docfx_yaml.extension," "sphinx.ext.intersphinx," "sphinx.ext.coverage," "sphinx.ext.napoleon," "sphinx.ext.todo," "sphinx.ext.viewcode," "recommonmark" ), "-b", "html", "-d", os.path.join("docs", "_build", "doctrees", ""), os.path.join("docs", ""), os.path.join("docs", "_build", "html", ""), ) ```
[ { "content": "```python\nfrom bitarray import bitarray\n\n\ndef MissingNumbers(array):\n bArray = bitarray(100)\n missingNums = []\n tempPair = []\n strList = []\n itr = iter(range(100))\n strRange = \"\"\n\n\n for num in array:\n bArray[num] = 1\n\n for i in itr:\n print(i...
[ { "content": "<|memory_start|>```python\nfrom bitarray import bitarray\n\n\ndef MissingNumbers(array):\n bArray = bitarray(100)\n missingNums = []\n tempPair = []\n strList = []\n itr = iter(range(100))\n strRange = \"\"\n\n\n for num in array:\n bArray[num] = 1\n\n for i in itr:\...
```python from bitarray import bitarray def MissingNumbers(array): bArray = bitarray(100) missingNums = [] tempPair = [] strList = [] itr = iter(range(100)) strRange = "" for num in array: bArray[num] = 1 for i in itr: print(i) if bArray[i] == 0: tempPair.append(str(i)) while bArray[i] == 0: print(i) try: i = next(itr) except StopIteration: break tempPair.append(str(i - 1)) missingNums.append(tempPair) tempPair = [] for pair in missingNums: if pair[0] == pair[1]: strList.append(pair[0]) strList.append(", ") else: strList.append(pair[0]) strList.append('-') strList.append(pair[1]) strList.append(', ') return strRange.join(strList) def main(): array = [] for i in range(90): if i < 40 or i > 60: array.append(i) array = [0, 2, 4, 6, 9, 13, 15, 18, 21, 22, 24, 25] print(MissingNumbers(array)) main() ```
[ { "content": "```python\nimport pandas as pd\nfrom faker import Faker\nfrom uuid import uuid4\nimport logging\nimport numpy as np\nfrom datetime import datetime\n\n\nclass DataSet:\n def __init__(self, length, **kwargs):\n self.data = self.create(length, **kwargs)\n\n def get_data(self):\n r...
[ { "content": "<|memory_start|>```python\nimport pandas as pd\nfrom faker import Faker\nfrom uuid import uuid4\nimport logging\nimport numpy as np\nfrom datetime import datetime\n\n\nclass DataSet:\n def __init__(self, length, **kwargs):\n self.data = self.create(length, **kwargs)\n\n def get_data(s...
```python import pandas as pd from faker import Faker from uuid import uuid4 import logging import numpy as np from datetime import datetime class DataSet: def __init__(self, length, **kwargs): self.data = self.create(length, **kwargs) def get_data(self): return self.data def num_data(self, ty, length): a = ty['min'] b = ty['max'] return pd.Series(np.random.uniform(a, b, length)) def num_int(self, ty, length): a = ty['min'] b = ty['max'] return pd.Series(np.random.random_integers(a, b, length)) def norm_data(self, ty, length): if len(ty) == 1: return pd.Series(np.random.standard_normal(size=length)) mean = ty['mean'] sd = ty['sd'] return pd.Series(np.random.normal(mean, sd, length)) def exp_data(self, ty, length): B = float(1) / float(ty['lam']) return pd.Series(np.random.exponential(B, length)) def binom_data(self, ty, length): n = ty['n'] p = ty['p'] return pd.Series(np.random.binomial(n, p, length)) def poisson_data(self, ty, length): lam = ty['lam'] return pd.Series(np.random.poisson(lam, length)) def text_data(self, ty, length): res = [] f = Faker() for _ in range(0, length - 1): res.append(f.text()) return pd.Series(res) def name_data(self, ty, length): res = [] f = Faker() for _ in range(0, length - 1): res.append(f.name()) return pd.Series(res) def cats_data(self, ty, length): res = [] f = Faker() for _ in range(0, length - 1): res.append(f.name()) return pd.Series(res) def date_data(self, ty, length): # TODO add error handling and validation for date strings passed res = [] f = Faker() begin = datetime.strptime(ty['begin'], '%Y-%m-%d') end = datetime.strptime(ty['end'], '%Y-%m-%d') for _ in range(0, length - 1): res.append(f.date_time_between_dates(datetime_start=begin, datetime_end=end)) return pd.Series(res) def coords_data(self, ty, length): lat_min = ty['lat_min'] lat_max = ty['lat_max'] lon_min = ty['lon_min'] lon_max = ty['lon_max'] if lat_min not in range(-90, 90) or lat_min > lat_max: logging.error('lat ranges unacceptable; not in [-90, 90] or lat_min > lat_max') if lon_min not in range(-180, 180) or lon_min > lon_max: logging.error('lon ranges unacceptable; not in [-180, 180] or lon_min > lon_max') return pd.Series(zip(np.random.uniform(lat_min, lat_max, length), np.random.uniform(lat_min, lat_max, length))) def address_data(self, ty, length): res = [] f = Faker() for _ in range(0, length - 1): res.append(f.address()) return pd.Series(res) def zip_data(self, ty, length): res = [] f = Faker() for _ in range(0, length - 1): res.append(f.name()) return pd.Series(res) @staticmethod def uuid_data(ty, length): """ Generate a column of random uuids. :param length: The number of uuids. :type length: int. :return: The column of uuids. :rtype: pd.Series """ return pd.Series(list(map(lambda _: uuid4(), range(length)))) @staticmethod def faker_data(ty, length): """ Generate a column based on any faker data type. :param ty: A configuration for the faker data. Must contain faker provider and related args as dict. :param length: The number of rows wanted. :param ty: dict. :param length: The number of rows wanted. :type length: int. :return: The column of Faker data. :rtype: pd.Series """ try: provider = ty["provider"] del ty["provider"] return pd.Series(list(map(lambda _: getattr(Faker(), provider)(**ty), range(length)))) except KeyError: raise KeyError("You have to define the Faker provider.") except AttributeError: raise AttributeError("Faker().{}() is not a valid Faker provider.".format(provider)) def create(self, length, cols=None, types=None, coltypes=None): series_res = {} ops = {'num': self.num_data, 'int': self.num_int, 'norm': self.norm_data, 'exp': self.exp_data, 'bin': self.binom_data, 'pois': self.poisson_data, 'txt': self.text_data, 'name': self.name_data, 'addr': self.address_data, 'zip': self.zip_data, 'date': self.date_data, 'uuid': self.uuid_data, 'faker': self.faker_data} if cols and types and coltypes: logging.error('coltypes should not be defined when cols and types are defined') if (cols and not types) or (types and not cols): logging.error('cols and types must both be defined together, as lists') if (cols and types): validate_types(types) if len(cols) != len(types): logging.error('cols and types must be lists of equal length') for i in len(cols): series_res[col[i]] = ops[types[i]['type']](types[i], length) else: if not coltypes: logging.error('please define either cols and types or coltypes') # Assure iteritems compatibility throught 2.7 and 3+ try: coltypes_items = coltypes.iteritems() except AttributeError: coltypes_items = coltypes.items() for col, typ in coltypes_items: data_builder = ops[typ['type']] del typ['type'] series_res[col] = data_builder(typ, length) return pd.DataFrame(series_res) ```
[ { "content": "Reproduce the code exactly as provided (keep formatting):\n```python\nfrom _pydevd_bundle.pydevd_constants import dict_contains\nimport sys\nfrom _pydevd_bundle import pydevd_xml\nfrom os.path import basename\nimport traceback\ntry:\n from urllib import quote, quote_plus, unquote, unquote_plus\...
[ { "content": "Reproduce the code exactly as provided (keep formatting):\n<|memory_start|>```python\nfrom _pydevd_bundle.pydevd_constants import dict_contains\nimport sys\nfrom _pydevd_bundle import pydevd_xml\nfrom os.path import basename\nimport traceback\ntry:\n from urllib import quote, quote_plus, unquot...
```python from _pydevd_bundle.pydevd_constants import dict_contains import sys from _pydevd_bundle import pydevd_xml from os.path import basename import traceback try: from urllib import quote, quote_plus, unquote, unquote_plus except: from urllib.parse import quote, quote_plus, unquote, unquote_plus #@Reimport @UnresolvedImport #=================================================================================================== # print_var_node #=================================================================================================== def print_var_node(xml_node, stream): name = xml_node.getAttribute('name') value = xml_node.getAttribute('value') val_type = xml_node.getAttribute('type') found_as = xml_node.getAttribute('found_as') stream.write('Name: ') stream.write(unquote_plus(name)) stream.write(', Value: ') stream.write(unquote_plus(value)) stream.write(', Type: ') stream.write(unquote_plus(val_type)) if found_as: stream.write(', Found as: %s' % (unquote_plus(found_as),)) stream.write('\n') #=================================================================================================== # print_referrers #=================================================================================================== def print_referrers(obj, stream=None): if stream is None: stream = sys.stdout result = get_referrer_info(obj) from xml.dom.minidom import parseString dom = parseString(result) xml = dom.getElementsByTagName('xml')[0] for node in xml.childNodes: if node.nodeType == node.TEXT_NODE: continue if node.localName == 'for': stream.write('Searching references for: ') for child in node.childNodes: if child.nodeType == node.TEXT_NODE: continue print_var_node(child, stream) elif node.localName == 'var': stream.write('Referrer found: ') print_var_node(node, stream) else: sys.stderr.write('Unhandled node: %s\n' % (node,)) return result #=================================================================================================== # get_referrer_info #=================================================================================================== def get_referrer_info(searched_obj): DEBUG = 0 if DEBUG: sys.stderr.write('Getting referrers info.\n') try: try: if searched_obj is None: ret = ['<xml>\n'] ret.append('<for>\n') ret.append(pydevd_xml.var_to_xml( searched_obj, 'Skipping getting referrers for None', additional_in_xml=' id="%s"' % (id(searched_obj),))) ret.append('</for>\n') ret.append('</xml>') ret = ''.join(ret) return ret obj_id = id(searched_obj) try: if DEBUG: sys.stderr.write('Getting referrers...\n') import gc referrers = gc.get_referrers(searched_obj) except: traceback.print_exc() ret = ['<xml>\n'] ret.append('<for>\n') ret.append(pydevd_xml.var_to_xml( searched_obj, 'Exception raised while trying to get_referrers.', additional_in_xml=' id="%s"' % (id(searched_obj),))) ret.append('</for>\n') ret.append('</xml>') ret = ''.join(ret) return ret if DEBUG: sys.stderr.write('Found %s referrers.\n' % (len(referrers),)) curr_frame = sys._getframe() frame_type = type(curr_frame) #Ignore this frame and any caller frame of this frame ignore_frames = {} #Should be a set, but it's not available on all python versions. while curr_frame is not None: if basename(curr_frame.f_code.co_filename).startswith('pydev'): ignore_frames[curr_frame] = 1 curr_frame = curr_frame.f_back ret = ['<xml>\n'] ret.append('<for>\n') if DEBUG: sys.stderr.write('Searching Referrers of obj with id="%s"\n' % (obj_id,)) ret.append(pydevd_xml.var_to_xml( searched_obj, 'Referrers of obj with id="%s"' % (obj_id,))) ret.append('</for>\n') all_objects = None for r in referrers: try: if dict_contains(ignore_frames, r): continue #Skip the references we may add ourselves except: pass #Ok: unhashable type checked... if r is referrers: continue r_type = type(r) r_id = str(id(r)) representation = str(r_type) found_as = '' if r_type == frame_type: if DEBUG: sys.stderr.write('Found frame referrer: %r\n' % (r,)) for key, val in r.f_locals.items(): if val is searched_obj: found_as = key break elif r_type == dict: if DEBUG: sys.stderr.write('Found dict referrer: %r\n' % (r,)) # Try to check if it's a value in the dict (and under which key it was found) for key, val in r.items(): if val is searched_obj: found_as = key if DEBUG: sys.stderr.write(' Found as %r in dict\n' % (found_as,)) break #Ok, there's one annoying thing: many times we find it in a dict from an instance, #but with this we don't directly have the class, only the dict, so, to workaround that #we iterate over all reachable objects ad check if one of those has the given dict. if all_objects is None: all_objects = gc.get_objects() for x in all_objects: try: if getattr(x, '__dict__', None) is r: r = x r_type = type(x) r_id = str(id(r)) representation = str(r_type) break except: pass #Just ignore any error here (i.e.: ReferenceError, etc.) elif r_type in (tuple, list): if DEBUG: sys.stderr.write('Found tuple referrer: %r\n' % (r,)) #Don't use enumerate() because not all Python versions have it. i = 0 for x in r: if x is searched_obj: found_as = '%s[%s]' % (r_type.__name__, i) if DEBUG: sys.stderr.write(' Found as %s in tuple: \n' % (found_as,)) break i += 1 if found_as: if not isinstance(found_as, str): found_as = str(found_as) found_as = ' found_as="%s"' % (pydevd_xml.make_valid_xml_value(found_as),) ret.append(pydevd_xml.var_to_xml( r, representation, additional_in_xml=' id="%s"%s' % (r_id, found_as))) finally: if DEBUG: sys.stderr.write('Done searching for references.\n') #If we have any exceptions, don't keep dangling references from this frame to any of our objects. all_objects = None referrers = None searched_obj = None r = None x = None key = None val = None curr_frame = None ignore_frames = None except: traceback.print_exc() ret = ['<xml>\n'] ret.append('<for>\n') ret.append(pydevd_xml.var_to_xml( searched_obj, 'Error getting referrers for:', additional_in_xml=' id="%s"' % (id(searched_obj),))) ret.append('</for>\n') ret.append('</xml>') ret = ''.join(ret) return ret ret.append('</xml>') ret = ''.join(ret) return ret ```
[ { "content": "Repeat the code exactly as the original, including blank lines:\n```python\n#Constraining Order - a simple constraint satisfaction library\n#\n#Copyright (c) 2015 Johannes Reinhardt <jreinhardt@ist-dein-freund.de>\n#\n#Permission is hereby granted, free of charge, to any person obtaining a copy\n#...
[ { "content": "Repeat the code exactly as the original, including blank lines:\n<|memory_start|>```python\n#Constraining Order - a simple constraint satisfaction library\n#\n#Copyright (c) 2015 Johannes Reinhardt <jreinhardt@ist-dein-freund.de>\n#\n#Permission is hereby granted, free of charge, to any person obt...
```python #Constraining Order - a simple constraint satisfaction library # #Copyright (c) 2015 Johannes Reinhardt <jreinhardt@ist-dein-freund.de> # #Permission is hereby granted, free of charge, to any person obtaining a copy #of this software and associated documentation files (the "Software"), to deal #in the Software without restriction, including without limitation the rights #to use, copy, modify, merge, publish, distribute, sublicense, and/or sell #copies of the Software, and to permit persons to whom the Software is #furnished to do so, subject to the following conditions: # #The above copyright notice and this permission notice shall be included in all #copies or substantial portions of the Software. # #THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR #IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, #FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE #AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER #LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, #OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE #SOFTWARE. """ This module contains functions for solving and reducing CSPs """ from __future__ import unicode_literals from itertools import product from constrainingorder import Space from constrainingorder.constraints import FixedValue from constrainingorder.sets import DiscreteSet, IntervalSet def ac3(space): """ AC-3 algorithm. This reduces the domains of the variables by propagating constraints to ensure arc consistency. :param Space space: The space to reduce """ #determine arcs arcs = {} for name in space.variables: arcs[name] = set([]) for const in space.constraints: for vname1,vname2 in product(const.vnames,const.vnames): if vname1 != vname2: #this is pessimistic, we assume that each constraint #pairwisely couples all variables it affects arcs[vname1].add(vname2) #enforce node consistency for vname in space.variables: for const in space.constraints: _unary(space,const,vname) #assemble work list worklist = set([]) for v1 in space.variables: for v2 in space.variables: for const in space.constraints: if _binary(space,const,v1,v2): for name in arcs[v1]: worklist.add((v1,name)) #work through work list while worklist: v1,v2 = worklist.pop() for const in space.constraints: if _binary(space,const,v1,v2): for vname in arcs[v1]: worklist.add((v1,vname)) def _unary(space,const,name): """ Reduce the domain of variable name to be node-consistent with this constraint, i.e. remove those values for the variable that are not consistent with the constraint. returns True if the domain of name was modified """ if not name in const.vnames: return False if space.variables[name].discrete: values = const.domains[name] else: values = const.domains[name] space.domains[name] = space.domains[name].intersection(values) return True def _binary(space,const,name1,name2): """ reduce the domain of variable name1 to be two-consistent (arc-consistent) with this constraint, i.e. remove those values for the variable name1, for which no values for name2 exist such that this pair is consistent with the constraint returns True if the domain of name1 was modified """ if not (name1 in const.vnames and name2 in const.vnames): return False remove = set([]) for v1 in space.domains[name1].iter_members(): for v2 in space.domains[name2].iter_members(): if const.consistent({name1 : v1, name2 : v2}): break else: remove.add(v1) if len(remove) > 0: if space.variables[name1].discrete: remove = DiscreteSet(remove) else: remove = IntervalSet.from_values(remove) space.domains[name1] = space.domains[name1].difference(remove) return True else: return False def solve(space,method='backtrack',ordering=None): """ Generator for all solutions. :param str method: the solution method to employ :param ordering: an optional parameter ordering :type ordering: sequence of parameter names Methods: :"backtrack": simple chronological backtracking :"ac-lookahead": full lookahead """ if ordering is None: ordering = list(space.variables.keys()) if not space.is_discrete(): raise ValueError("Can not backtrack on non-discrete space") if method=='backtrack': for label in _backtrack(space,{},ordering): yield label elif method=='ac-lookahead': for label in _lookahead(space,{},ordering): yield label else: raise ValueError("Unknown solution method: %s" % method) def _backtrack(space,label,ordering): level = len(label) if level == len(space.variables): if space.satisfied(label): yield label elif space.consistent(label): vname = ordering[level] newlabel = label.copy() for val in space.domains[vname].iter_members(): newlabel[vname] = val for sol in _backtrack(space,newlabel,ordering): yield sol def _lookahead(space,label,ordering): level = len(label) if len(label) == len(space.variables): if space.satisfied(label): yield label elif space.consistent(label): vname = ordering[level] var = space.variables[vname] newlabel = label.copy() for val in space.domains[vname].iter_members(): nspace = Space(list(space.variables.values()), space.constraints + [FixedValue(var,val)]) newlabel[vname] = val ac3(nspace) for sol in _lookahead(nspace,newlabel,ordering): yield sol ```
[ { "content": "Here is the code block:\n```python\nfrom __future__ import division, absolute_import, print_function\n\nimport copy\nimport pickle\nimport sys\nimport platform\nimport gc\nimport copy\nimport warnings\nimport tempfile\nfrom os import path\nfrom io import BytesIO\n\nimport numpy as np\nfrom numpy.t...
[ { "content": "Here is the code block:\n<|memory_start|>```python\nfrom __future__ import division, absolute_import, print_function\n\nimport copy\nimport pickle\nimport sys\nimport platform\nimport gc\nimport copy\nimport warnings\nimport tempfile\nfrom os import path\nfrom io import BytesIO\n\nimport numpy as ...
```python from __future__ import division, absolute_import, print_function import copy import pickle import sys import platform import gc import copy import warnings import tempfile from os import path from io import BytesIO import numpy as np from numpy.testing import ( run_module_suite, TestCase, assert_, assert_equal, assert_almost_equal, assert_array_equal, assert_array_almost_equal, assert_raises, assert_warns, dec ) from numpy.testing.utils import _assert_valid_refcount from numpy.compat import asbytes, asunicode, asbytes_nested, long, sixu rlevel = 1 class TestRegression(TestCase): def test_invalid_round(self,level=rlevel): """Ticket #3""" v = 4.7599999999999998 assert_array_equal(np.array([v]), np.array(v)) def test_mem_empty(self,level=rlevel): """Ticket #7""" np.empty((1,), dtype=[('x', np.int64)]) def test_pickle_transposed(self,level=rlevel): """Ticket #16""" a = np.transpose(np.array([[2, 9], [7, 0], [3, 8]])) f = BytesIO() pickle.dump(a, f) f.seek(0) b = pickle.load(f) f.close() assert_array_equal(a, b) def test_typeNA(self,level=rlevel): """Ticket #31""" assert_equal(np.typeNA[np.int64], 'Int64') assert_equal(np.typeNA[np.uint64], 'UInt64') def test_dtype_names(self,level=rlevel): """Ticket #35""" dt = np.dtype([(('name', 'label'), np.int32, 3)]) def test_reduce(self,level=rlevel): """Ticket #40""" assert_almost_equal(np.add.reduce([1., .5], dtype=None), 1.5) def test_zeros_order(self,level=rlevel): """Ticket #43""" np.zeros([3], int, 'C') np.zeros([3], order='C') np.zeros([3], int, order='C') def test_asarray_with_order(self,level=rlevel): """Check that nothing is done when order='F' and array C/F-contiguous""" a = np.ones(2) assert_(a is np.asarray(a, order='F')) def test_ravel_with_order(self,level=rlevel): """Check that ravel works when order='F' and array C/F-contiguous""" a = np.ones(2) assert_(not a.ravel('F').flags.owndata) def test_sort_bigendian(self,level=rlevel): """Ticket #47""" a = np.linspace(0, 10, 11) c = a.astype(np.dtype('<f8')) c.sort() assert_array_almost_equal(c, a) def test_negative_nd_indexing(self,level=rlevel): """Ticket #49""" c = np.arange(125).reshape((5, 5, 5)) origidx = np.array([-1, 0, 1]) idx = np.array(origidx) c[idx] assert_array_equal(idx, origidx) def test_char_dump(self,level=rlevel): """Ticket #50""" f = BytesIO() ca = np.char.array(np.arange(1000, 1010), itemsize=4) ca.dump(f) f.seek(0) ca = np.load(f) f.close() def test_noncontiguous_fill(self,level=rlevel): """Ticket #58.""" a = np.zeros((5, 3)) b = a[:, :2,] def rs(): b.shape = (10,) self.assertRaises(AttributeError, rs) def test_bool(self,level=rlevel): """Ticket #60""" x = np.bool_(1) def test_indexing1(self,level=rlevel): """Ticket #64""" descr = [('x', [('y', [('z', 'c16', (2,)),]),]),] buffer = ((([6j, 4j],),),) h = np.array(buffer, dtype=descr) h['x']['y']['z'] def test_indexing2(self,level=rlevel): """Ticket #65""" descr = [('x', 'i4', (2,))] buffer = ([3, 2],) h = np.array(buffer, dtype=descr) h['x'] def test_round(self,level=rlevel): """Ticket #67""" x = np.array([1+2j]) assert_almost_equal(x**(-1), [1/(1+2j)]) def test_scalar_compare(self,level=rlevel): """Ticket #72""" a = np.array(['test', 'auto']) assert_array_equal(a == 'auto', np.array([False, True])) self.assertTrue(a[1] == 'auto') self.assertTrue(a[0] != 'auto') b = np.linspace(0, 10, 11) self.assertTrue(b != 'auto') self.assertTrue(b[0] != 'auto') def test_unicode_swapping(self,level=rlevel): """Ticket #79""" ulen = 1 ucs_value = sixu('\U0010FFFF') ua = np.array([[[ucs_value*ulen]*2]*3]*4, dtype='U%s' % ulen) ua2 = ua.newbyteorder() def test_object_array_fill(self,level=rlevel): """Ticket #86""" x = np.zeros(1, 'O') x.fill([]) def test_mem_dtype_align(self,level=rlevel): """Ticket #93""" self.assertRaises(TypeError, np.dtype, {'names':['a'],'formats':['foo']}, align=1) @dec.knownfailureif((sys.version_info[0] >= 3) or (sys.platform == "win32" and platform.architecture()[0] == "64bit"), "numpy.intp('0xff', 16) not supported on Py3, " "as it does not inherit from Python int") def test_intp(self,level=rlevel): """Ticket #99""" i_width = np.int_(0).nbytes*2 - 1 np.intp('0x' + 'f'*i_width, 16) self.assertRaises(OverflowError, np.intp, '0x' + 'f'*(i_width+1), 16) self.assertRaises(ValueError, np.intp, '0x1', 32) assert_equal(255, np.intp('0xFF', 16)) assert_equal(1024, np.intp(1024)) def test_endian_bool_indexing(self,level=rlevel): """Ticket #105""" a = np.arange(10., dtype='>f8') b = np.arange(10., dtype='<f8') xa = np.where((a>2) & (a<6)) xb = np.where((b>2) & (b<6)) ya = ((a>2) & (a<6)) yb = ((b>2) & (b<6)) assert_array_almost_equal(xa, ya.nonzero()) assert_array_almost_equal(xb, yb.nonzero()) assert_(np.all(a[ya] > 0.5)) assert_(np.all(b[yb] > 0.5)) def test_endian_where(self,level=rlevel): """GitHuB issue #369""" net = np.zeros(3, dtype='>f4') net[1] = 0.00458849 net[2] = 0.605202 max_net = net.max() test = np.where(net <= 0., max_net, net) correct = np.array([ 0.60520202, 0.00458849, 0.60520202]) assert_array_almost_equal(test, correct) def test_endian_recarray(self,level=rlevel): """Ticket #2185""" dt = np.dtype([ ('head', '>u4'), ('data', '>u4', 2), ]) buf = np.recarray(1, dtype=dt) buf[0]['head'] = 1 buf[0]['data'][:] = [1, 1] h = buf[0]['head'] d = buf[0]['data'][0] buf[0]['head'] = h buf[0]['data'][0] = d assert_(buf[0]['head'] == 1) def test_mem_dot(self,level=rlevel): """Ticket #106""" x = np.random.randn(0, 1) y = np.random.randn(10, 1) # Dummy array to detect bad memory access: _z = np.ones(10) _dummy = np.empty((0, 10)) z = np.lib.stride_tricks.as_strided(_z, _dummy.shape, _dummy.strides) np.dot(x, np.transpose(y), out=z) assert_equal(_z, np.ones(10)) # Do the same for the built-in dot: np.core.multiarray.dot(x, np.transpose(y), out=z) assert_equal(_z, np.ones(10)) def test_arange_endian(self,level=rlevel): """Ticket #111""" ref = np.arange(10) x = np.arange(10, dtype='<f8') assert_array_equal(ref, x) x = np.arange(10, dtype='>f8') assert_array_equal(ref, x) # Longfloat support is not consistent enough across # platforms for this test to be meaningful. # def test_longfloat_repr(self,level=rlevel): # """Ticket #112""" # if np.longfloat(0).itemsize > 8: # a = np.exp(np.array([1000],dtype=np.longfloat)) # assert_(str(a)[1:9] == str(a[0])[:8]) def test_argmax(self,level=rlevel): """Ticket #119""" a = np.random.normal(0, 1, (4, 5, 6, 7, 8)) for i in range(a.ndim): aargmax = a.argmax(i) def test_mem_divmod(self,level=rlevel): """Ticket #126""" for i in range(10): divmod(np.array([i])[0], 10) def test_hstack_invalid_dims(self,level=rlevel): """Ticket #128""" x = np.arange(9).reshape((3, 3)) y = np.array([0, 0, 0]) self.assertRaises(ValueError, np.hstack, (x, y)) def test_squeeze_type(self,level=rlevel): """Ticket #133""" a = np.array([3]) b = np.array(3) assert_(type(a.squeeze()) is np.ndarray) assert_(type(b.squeeze()) is np.ndarray) def test_add_identity(self,level=rlevel): """Ticket #143""" assert_equal(0, np.add.identity) def test_numpy_float_python_long_addition(self): # Check that numpy float and python longs can be added correctly. a = np.float_(23.) + 2**135 assert_equal(a, 23. + 2**135) def test_binary_repr_0(self,level=rlevel): """Ticket #151""" assert_equal('0', np.binary_repr(0)) def test_rec_iterate(self,level=rlevel): """Ticket #160""" descr = np.dtype([('i', int), ('f', float), ('s', '|S3')]) x = np.rec.array([(1, 1.1, '1.0'), (2, 2.2, '2.0')], dtype=descr) x[0].tolist() [i for i in x[0]] def test_unicode_string_comparison(self,level=rlevel): """Ticket #190""" a = np.array('hello', np.unicode_) b = np.array('world') a == b def test_tobytes_FORTRANORDER_discontiguous(self,level=rlevel): """Fix in r2836""" # Create discontiguous Fortran-ordered array x = np.array(np.random.rand(3, 3), order='F')[:, :2] assert_array_almost_equal(x.ravel(), np.fromstring(x.tobytes())) def test_flat_assignment(self,level=rlevel): """Correct behaviour of ticket #194""" x = np.empty((3, 1)) x.flat = np.arange(3) assert_array_almost_equal(x, [[0], [1], [2]]) x.flat = np.arange(3, dtype=float) assert_array_almost_equal(x, [[0], [1], [2]]) def test_broadcast_flat_assignment(self,level=rlevel): """Ticket #194""" x = np.empty((3, 1)) def bfa(): x[:] = np.arange(3) def bfb(): x[:] = np.arange(3, dtype=float) self.assertRaises(ValueError, bfa) self.assertRaises(ValueError, bfb) def test_nonarray_assignment(self): # See also Issue gh-2870, test for nonarray assignment # and equivalent unsafe casted array assignment a = np.arange(10) b = np.ones(10, dtype=bool) r = np.arange(10) def assign(a, b, c): a[b] = c assert_raises(ValueError, assign, a, b, np.nan) a[b] = np.array(np.nan) # but not this. assert_raises(ValueError, assign, a, r, np.nan) a[r] = np.array(np.nan) def test_unpickle_dtype_with_object(self,level=rlevel): """Implemented in r2840""" dt = np.dtype([('x', int), ('y', np.object_), ('z', 'O')]) f = BytesIO() pickle.dump(dt, f) f.seek(0) dt_ = pickle.load(f) f.close() assert_equal(dt, dt_) def test_mem_array_creation_invalid_specification(self,level=rlevel): """Ticket #196""" dt = np.dtype([('x', int), ('y', np.object_)]) # Wrong way self.assertRaises(ValueError, np.array, [1, 'object'], dt) # Correct way np.array([(1, 'object')], dt) def test_recarray_single_element(self,level=rlevel): """Ticket #202""" a = np.array([1, 2, 3], dtype=np.int32) b = a.copy() r = np.rec.array(a, shape=1, formats=['3i4'], names=['d']) assert_array_equal(a, b) assert_equal(a, r[0][0]) def test_zero_sized_array_indexing(self,level=rlevel): """Ticket #205""" tmp = np.array([]) def index_tmp(): tmp[np.array(10)] self.assertRaises(IndexError, index_tmp) def test_chararray_rstrip(self,level=rlevel): """Ticket #222""" x = np.chararray((1,), 5) x[0] = asbytes('a ') x = x.rstrip() assert_equal(x[0], asbytes('a')) def test_object_array_shape(self,level=rlevel): """Ticket #239""" assert_equal(np.array([[1, 2], 3, 4], dtype=object).shape, (3,)) assert_equal(np.array([[1, 2], [3, 4]], dtype=object).shape, (2, 2)) assert_equal(np.array([(1, 2), (3, 4)], dtype=object).shape, (2, 2)) assert_equal(np.array([], dtype=object).shape, (0,)) assert_equal(np.array([[], [], []], dtype=object).shape, (3, 0)) assert_equal(np.array([[3, 4], [5, 6], None], dtype=object).shape, (3,)) def test_mem_around(self,level=rlevel): """Ticket #243""" x = np.zeros((1,)) y = [0] decimal = 6 np.around(abs(x-y), decimal) <= 10.0**(-decimal) def test_character_array_strip(self,level=rlevel): """Ticket #246""" x = np.char.array(("x", "x ", "x ")) for c in x: assert_equal(c, "x") def test_lexsort(self,level=rlevel): """Lexsort memory error""" v = np.array([1, 2, 3, 4, 5, 6, 7, 8, 9, 10]) assert_equal(np.lexsort(v), 0) def test_lexsort_invalid_sequence(self): # Issue gh-4123 class BuggySequence(object): def __len__(self): return 4 def __getitem__(self, key): raise KeyError assert_raises(KeyError, np.lexsort, BuggySequence()) def test_pickle_dtype(self,level=rlevel): """Ticket #251""" pickle.dumps(np.float) def test_swap_real(self, level=rlevel): """Ticket #265""" assert_equal(np.arange(4, dtype='>c8').imag.max(), 0.0) assert_equal(np.arange(4, dtype='<c8').imag.max(), 0.0) assert_equal(np.arange(4, dtype='>c8').real.max(), 3.0) assert_equal(np.arange(4, dtype='<c8').real.max(), 3.0) def test_object_array_from_list(self, level=rlevel): """Ticket #270""" a = np.array([1, 'A', None]) def test_multiple_assign(self, level=rlevel): """Ticket #273""" a = np.zeros((3, 1), int) a[[1, 2]] = 1 def test_empty_array_type(self, level=rlevel): assert_equal(np.array([]).dtype, np.zeros(0).dtype) def test_void_copyswap(self, level=rlevel): dt = np.dtype([('one', '<i4'), ('two', '<i4')]) x = np.array((1, 2), dtype=dt) x = x.byteswap() assert_(x['one'] > 1 and x['two'] > 2) def test_method_args(self, level=rlevel): # Make sure methods and functions have same default axis # keyword and arguments funcs1= ['argmax', 'argmin', 'sum', ('product', 'prod'), ('sometrue', 'any'), ('alltrue', 'all'), 'cumsum', ('cumproduct', 'cumprod'), 'ptp', 'cumprod', 'prod', 'std', 'var', 'mean', 'round', 'min', 'max', 'argsort', 'sort'] funcs2 = ['compress', 'take', 'repeat'] for func in funcs1: arr = np.random.rand(8, 7) arr2 = arr.copy() if isinstance(func, tuple): func_meth = func[1] func = func[0] else: func_meth = func res1 = getattr(arr, func_meth)() res2 = getattr(np, func)(arr2) if res1 is None: res1 = arr if res1.dtype.kind in 'uib': assert_((res1 == res2).all(), func) else: assert_(abs(res1-res2).max() < 1e-8, func) for func in funcs2: arr1 = np.random.rand(8, 7) arr2 = np.random.rand(8, 7) res1 = None if func == 'compress': arr1 = arr1.ravel() res1 = getattr(arr2, func)(arr1) else: arr2 = (15*arr2).astype(int).ravel() if res1 is None: res1 = getattr(arr1, func)(arr2) res2 = getattr(np, func)(arr1, arr2) assert_(abs(res1-res2).max() < 1e-8, func) def test_mem_lexsort_strings(self, level=rlevel): """Ticket #298""" lst = ['abc', 'cde', 'fgh'] np.lexsort((lst,)) def test_fancy_index(self, level=rlevel): """Ticket #302""" x = np.array([1, 2])[np.array([0])] assert_equal(x.shape, (1,)) def test_recarray_copy(self, level=rlevel): """Ticket #312""" dt = [('x', np.int16), ('y', np.float64)] ra = np.array([(1, 2.3)], dtype=dt) rb = np.rec.array(ra, dtype=dt) rb['x'] = 2. assert_(ra['x'] != rb['x']) def test_rec_fromarray(self, level=rlevel): """Ticket #322""" x1 = np.array([[1, 2], [3, 4], [5, 6]]) x2 = np.array(['a', 'dd', 'xyz']) x3 = np.array([1.1, 2, 3]) np.rec.fromarrays([x1, x2, x3], formats="(2,)i4,a3,f8") def test_object_array_assign(self, level=rlevel): x = np.empty((2, 2), object) x.flat[2] = (1, 2, 3) assert_equal(x.flat[2], (1, 2, 3)) def test_ndmin_float64(self, level=rlevel): """Ticket #324""" x = np.array([1, 2, 3], dtype=np.float64) assert_equal(np.array(x, dtype=np.float32, ndmin=2).ndim, 2) assert_equal(np.array(x, dtype=np.float64, ndmin=2).ndim, 2) def test_ndmin_order(self, level=rlevel): """Issue #465 and related checks""" assert_(np.array([1, 2], order='C', ndmin=3).flags.c_contiguous) assert_(np.array([1, 2], order='F', ndmin=3).flags.f_contiguous) assert_(np.array(np.ones((2, 2), order='F'), ndmin=3).flags.f_contiguous) assert_(np.array(np.ones((2, 2), order='C'), ndmin=3).flags.c_contiguous) def test_mem_axis_minimization(self, level=rlevel): """Ticket #327""" data = np.arange(5) data = np.add.outer(data, data) def test_mem_float_imag(self, level=rlevel): """Ticket #330""" np.float64(1.0).imag def test_dtype_tuple(self, level=rlevel): """Ticket #334""" assert_(np.dtype('i4') == np.dtype(('i4', ()))) def test_dtype_posttuple(self, level=rlevel): """Ticket #335""" np.dtype([('col1', '()i4')]) def test_numeric_carray_compare(self, level=rlevel): """Ticket #341""" assert_equal(np.array(['X'], 'c'), asbytes('X')) def test_string_array_size(self, level=rlevel): """Ticket #342""" self.assertRaises(ValueError, np.array, [['X'], ['X', 'X', 'X']], '|S1') def test_dtype_repr(self, level=rlevel): """Ticket #344""" dt1=np.dtype(('uint32', 2)) dt2=np.dtype(('uint32', (2,))) assert_equal(dt1.__repr__(), dt2.__repr__()) def test_reshape_order(self, level=rlevel): """Make sure reshape order works.""" a = np.arange(6).reshape(2, 3, order='F') assert_equal(a, [[0, 2, 4], [1, 3, 5]]) a = np.array([[1, 2], [3, 4], [5, 6], [7, 8]]) b = a[:, 1] assert_equal(b.reshape(2, 2, order='F'), [[2, 6], [4, 8]]) def test_reshape_zero_strides(self, level=rlevel): """Issue #380, test reshaping of zero strided arrays""" a = np.ones(1) a = np.lib.stride_tricks.as_strided(a, shape=(5,), strides=(0,)) assert_(a.reshape(5, 1).strides[0] == 0) def test_reshape_zero_size(self, level=rlevel): """Github Issue #2700, setting shape failed for 0-sized arrays""" a = np.ones((0, 2)) a.shape = (-1, 2) # Cannot test if NPY_RELAXED_STRIDES_CHECKING changes the strides. # With NPY_RELAXED_STRIDES_CHECKING the test becomes superfluous. @dec.skipif(np.ones(1).strides[0] == np.iinfo(np.intp).max) def test_reshape_trailing_ones_strides(self): # Github issue gh-2949, bad strides for trailing ones of new shape a = np.zeros(12, dtype=np.int32)[::2] # not contiguous strides_c = (16, 8, 8, 8) strides_f = (8, 24, 48, 48) assert_equal(a.reshape(3, 2, 1, 1).strides, strides_c) assert_equal(a.reshape(3, 2, 1, 1, order='F').strides, strides_f) assert_equal(np.array(0, dtype=np.int32).reshape(1, 1).strides, (4, 4)) def test_repeat_discont(self, level=rlevel): """Ticket #352""" a = np.arange(12).reshape(4, 3)[:, 2] assert_equal(a.repeat(3), [2, 2, 2, 5, 5, 5, 8, 8, 8, 11, 11, 11]) def test_array_index(self, level=rlevel): """Make sure optimization is not called in this case.""" a = np.array([1, 2, 3]) a2 = np.array([[1, 2, 3]]) assert_equal(a[np.where(a==3)], a2[np.where(a2==3)]) def test_object_argmax(self, level=rlevel): a = np.array([1, 2, 3], dtype=object) assert_(a.argmax() == 2) def test_recarray_fields(self, level=rlevel): """Ticket #372""" dt0 = np.dtype([('f0', 'i4'), ('f1', 'i4')]) dt1 = np.dtype([('f0', 'i8'), ('f1', 'i8')]) for a in [np.array([(1, 2), (3, 4)], "i4,i4"), np.rec.array([(1, 2), (3, 4)], "i4,i4"), np.rec.array([(1, 2), (3, 4)]), np.rec.fromarrays([(1, 2), (3, 4)], "i4,i4"), np.rec.fromarrays([(1, 2), (3, 4)])]: assert_(a.dtype in [dt0, dt1]) def test_random_shuffle(self, level=rlevel): """Ticket #374""" a = np.arange(5).reshape((5, 1)) b = a.copy() np.random.shuffle(b) assert_equal(np.sort(b, axis=0), a) def test_refcount_vdot(self, level=rlevel): """Changeset #3443""" _assert_valid_refcount(np.vdot) def test_startswith(self, level=rlevel): ca = np.char.array(['Hi', 'There']) assert_equal(ca.startswith('H'), [True, False]) def test_noncommutative_reduce_accumulate(self, level=rlevel): """Ticket #413""" tosubtract = np.arange(5) todivide = np.array([2.0, 0.5, 0.25]) assert_equal(np.subtract.reduce(tosubtract), -10) assert_equal(np.divide.reduce(todivide), 16.0) assert_array_equal(np.subtract.accumulate(tosubtract), np.array([0, -1, -3, -6, -10])) assert_array_equal(np.divide.accumulate(todivide), np.array([2., 4., 16.])) def test_convolve_empty(self, level=rlevel): """Convolve should raise an error for empty input array.""" self.assertRaises(ValueError, np.convolve, [], [1]) self.assertRaises(ValueError, np.convolve, [1], []) def test_multidim_byteswap(self, level=rlevel): """Ticket #449""" r=np.array([(1, (0, 1, 2))], dtype="i2,3i2") assert_array_equal(r.byteswap(), np.array([(256, (0, 256, 512))], r.dtype)) def test_string_NULL(self, level=rlevel): """Changeset 3557""" assert_equal(np.array("a\x00\x0b\x0c\x00").item(), 'a\x00\x0b\x0c') def test_junk_in_string_fields_of_recarray(self, level=rlevel): """Ticket #483""" r = np.array([[asbytes('abc')]], dtype=[('var1', '|S20')]) assert_(asbytes(r['var1'][0][0]) == asbytes('abc')) def test_take_output(self, level=rlevel): """Ensure that 'take' honours output parameter.""" x = np.arange(12).reshape((3, 4)) a = np.take(x, [0, 2], axis=1) b = np.zeros_like(a) np.take(x, [0, 2], axis=1, out=b) assert_array_equal(a, b) def test_take_object_fail(self): # Issue gh-3001 d = 123. a = np.array([d, 1], dtype=object) ref_d = sys.getrefcount(d) try: a.take([0, 100]) except IndexError: pass assert_(ref_d == sys.getrefcount(d)) def test_array_str_64bit(self, level=rlevel): """Ticket #501""" s = np.array([1, np.nan], dtype=np.float64) with np.errstate(all='raise'): sstr = np.array_str(s) def test_frompyfunc_endian(self, level=rlevel): """Ticket #503""" from math import radians uradians = np.frompyfunc(radians, 1, 1) big_endian = np.array([83.4, 83.5], dtype='>f8') little_endian = np.array([83.4, 83.5], dtype='<f8') assert_almost_equal(uradians(big_endian).astype(float), uradians(little_endian).astype(float)) def test_mem_string_arr(self, level=rlevel): """Ticket #514""" s = "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa" t = [] np.hstack((t, s )) def test_arr_transpose(self, level=rlevel): """Ticket #516""" x = np.random.rand(*(2,)*16) y = x.transpose(list(range(16))) def test_string_mergesort(self, level=rlevel): """Ticket #540""" x = np.array(['a']*32) assert_array_equal(x.argsort(kind='m'), np.arange(32)) def test_argmax_byteorder(self, level=rlevel): """Ticket #546""" a = np.arange(3, dtype='>f') assert_(a[a.argmax()] == a.max()) def test_rand_seed(self, level=rlevel): """Ticket #555""" for l in np.arange(4): np.random.seed(l) def test_mem_deallocation_leak(self, level=rlevel): """Ticket #562""" a = np.zeros(5, dtype=float) b = np.array(a, dtype=float) del a, b def test_mem_on_invalid_dtype(self): "Ticket #583" self.assertRaises(ValueError, np.fromiter, [['12', ''], ['13', '']], str) def test_dot_negative_stride(self, level=rlevel): """Ticket #588""" x = np.array([[1, 5, 25, 125., 625]]) y = np.array([[20.], [160.], [640.], [1280.], [1024.]]) z = y[::-1].copy() y2 = y[::-1] assert_equal(np.dot(x, z), np.dot(x, y2)) def test_object_casting(self, level=rlevel): # This used to trigger the object-type version of # the bitwise_or operation, because float64 -> object # casting succeeds def rs(): x = np.ones([484, 286]) y = np.zeros([484, 286]) x |= y self.assertRaises(TypeError, rs) def test_unicode_scalar(self, level=rlevel): """Ticket #600""" x = np.array(["DROND", "DROND1"], dtype="U6") el = x[1] new = pickle.loads(pickle.dumps(el)) assert_equal(new, el) def test_arange_non_native_dtype(self, level=rlevel): """Ticket #616""" for T in ('>f4', '<f4'): dt = np.dtype(T) assert_equal(np.arange(0, dtype=dt).dtype, dt) assert_equal(np.arange(0.5, dtype=dt).dtype, dt) assert_equal(np.arange(5, dtype=dt).dtype, dt) def test_bool_indexing_invalid_nr_elements(self, level=rlevel): s = np.ones(10, dtype=float) x = np.array((15,), dtype=float) def ia(x, s, v): x[(s>0)]=v self.assertRaises(ValueError, ia, x, s, np.zeros(9, dtype=float)) self.assertRaises(ValueError, ia, x, s, np.zeros(11, dtype=float)) # Old special case (different code path): self.assertRaises(ValueError, ia, x.flat, s, np.zeros(9, dtype=float)) def test_mem_scalar_indexing(self, level=rlevel): """Ticket #603""" x = np.array([0], dtype=float) index = np.array(0, dtype=np.int32) x[index] def test_binary_repr_0_width(self, level=rlevel): assert_equal(np.binary_repr(0, width=3), '000') def test_fromstring(self, level=rlevel): assert_equal(np.fromstring("12:09:09", dtype=int, sep=":"), [12, 9, 9]) def test_searchsorted_variable_length(self, level=rlevel): x = np.array(['a', 'aa', 'b']) y = np.array(['d', 'e']) assert_equal(x.searchsorted(y), [3, 3]) def test_string_argsort_with_zeros(self, level=rlevel): """Check argsort for strings containing zeros.""" x = np.fromstring("\x00\x02\x00\x01", dtype="|S2") assert_array_equal(x.argsort(kind='m'), np.array([1, 0])) assert_array_equal(x.argsort(kind='q'), np.array([1, 0])) def test_string_sort_with_zeros(self, level=rlevel): """Check sort for strings containing zeros.""" x = np.fromstring("\x00\x02\x00\x01", dtype="|S2") y = np.fromstring("\x00\x01\x00\x02", dtype="|S2") assert_array_equal(np.sort(x, kind="q"), y) def test_copy_detection_zero_dim(self, level=rlevel): """Ticket #658""" np.indices((0, 3, 4)).T.reshape(-1, 3) def test_flat_byteorder(self, level=rlevel): """Ticket #657""" x = np.arange(10) assert_array_equal(x.astype('>i4'), x.astype('<i4').flat[:]) assert_array_equal(x.astype('>i4').flat[:], x.astype('<i4')) def test_uint64_from_negative(self, level=rlevel) : assert_equal(np.uint64(-2), np.uint64(18446744073709551614)) def test_sign_bit(self, level=rlevel): x = np.array([0, -0.0, 0]) assert_equal(str(np.abs(x)), '[ 0. 0. 0.]') def test_flat_index_byteswap(self, level=rlevel): for dt in (np.dtype('<i4'), np.dtype('>i4')): x = np.array([-1, 0, 1], dtype=dt) assert_equal(x.flat[0].dtype, x[0].dtype) def test_copy_detection_corner_case(self, level=rlevel): """Ticket #658""" np.indices((0, 3, 4)).T.reshape(-1, 3) # Cannot test if NPY_RELAXED_STRIDES_CHECKING changes the strides. # With NPY_RELAXED_STRIDES_CHECKING the test becomes superfluous, # 0-sized reshape itself is tested elsewhere. @dec.skipif(np.ones(1).strides[0] == np.iinfo(np.intp).max) def test_copy_detection_corner_case2(self, level=rlevel): """Ticket #771: strides are not set correctly when reshaping 0-sized arrays""" b = np.indices((0, 3, 4)).T.reshape(-1, 3) assert_equal(b.strides, (3 * b.itemsize, b.itemsize)) def test_object_array_refcounting(self, level=rlevel): """Ticket #633""" if not hasattr(sys, 'getrefcount'): return # NB. this is probably CPython-specific cnt = sys.getrefcount a = object() b = object() c = object() cnt0_a = cnt(a) cnt0_b = cnt(b) cnt0_c = cnt(c) # -- 0d -> 1d broadcasted slice assignment arr = np.zeros(5, dtype=np.object_) arr[:] = a assert_equal(cnt(a), cnt0_a + 5) arr[:] = b assert_equal(cnt(a), cnt0_a) assert_equal(cnt(b), cnt0_b + 5) arr[:2] = c assert_equal(cnt(b), cnt0_b + 3) assert_equal(cnt(c), cnt0_c + 2) del arr # -- 1d -> 2d broadcasted slice assignment arr = np.zeros((5, 2), dtype=np.object_) arr0 = np.zeros(2, dtype=np.object_) arr0[0] = a assert_(cnt(a) == cnt0_a + 1) arr0[1] = b assert_(cnt(b) == cnt0_b + 1) arr[:,:] = arr0 assert_(cnt(a) == cnt0_a + 6) assert_(cnt(b) == cnt0_b + 6) arr[:, 0] = None assert_(cnt(a) == cnt0_a + 1) del arr, arr0 # -- 2d copying + flattening arr = np.zeros((5, 2), dtype=np.object_) arr[:, 0] = a arr[:, 1] = b assert_(cnt(a) == cnt0_a + 5) assert_(cnt(b) == cnt0_b + 5) arr2 = arr.copy() assert_(cnt(a) == cnt0_a + 10) assert_(cnt(b) == cnt0_b + 10) arr2 = arr[:, 0].copy() assert_(cnt(a) == cnt0_a + 10) assert_(cnt(b) == cnt0_b + 5) arr2 = arr.flatten() assert_(cnt(a) == cnt0_a + 10) assert_(cnt(b) == cnt0_b + 10) del arr, arr2 # -- concatenate, repeat, take, choose arr1 = np.zeros((5, 1), dtype=np.object_) arr2 = np.zeros((5, 1), dtype=np.object_) arr1[...] = a arr2[...] = b assert_(cnt(a) == cnt0_a + 5) assert_(cnt(b) == cnt0_b + 5) arr3 = np.concatenate((arr1, arr2)) assert_(cnt(a) == cnt0_a + 5 + 5) assert_(cnt(b) == cnt0_b + 5 + 5) arr3 = arr1.repeat(3, axis=0) assert_(cnt(a) == cnt0_a + 5 + 3*5) arr3 = arr1.take([1, 2, 3], axis=0) assert_(cnt(a) == cnt0_a + 5 + 3) x = np.array([[0], [1], [0], [1], [1]], int) arr3 = x.choose(arr1, arr2) assert_(cnt(a) == cnt0_a + 5 + 2) assert_(cnt(b) == cnt0_b + 5 + 3) def test_mem_custom_float_to_array(self, level=rlevel): """Ticket 702""" class MyFloat(object): def __float__(self): return 1.0 tmp = np.atleast_1d([MyFloat()]) tmp2 = tmp.astype(float) def test_object_array_refcount_self_assign(self, level=rlevel): """Ticket #711""" class VictimObject(object): deleted = False def __del__(self): self.deleted = True d = VictimObject() arr = np.zeros(5, dtype=np.object_) arr[:] = d del d arr[:] = arr # refcount of 'd' might hit zero here assert_(not arr[0].deleted) arr[:] = arr # trying to induce a segfault by doing it again... assert_(not arr[0].deleted) def test_mem_fromiter_invalid_dtype_string(self, level=rlevel): x = [1, 2, 3] self.assertRaises(ValueError, np.fromiter, [xi for xi in x], dtype='S') def test_reduce_big_object_array(self, level=rlevel): """Ticket #713""" oldsize = np.setbufsize(10*16) a = np.array([None]*161, object) assert_(not np.any(a)) np.setbufsize(oldsize) def test_mem_0d_array_index(self, level=rlevel): """Ticket #714""" np.zeros(10)[np.array(0)] def test_floats_from_string(self, level=rlevel): """Ticket #640, floats from string""" fsingle = np.single('1.234') fdouble = np.double('1.234') flongdouble = np.longdouble('1.234') assert_almost_equal(fsingle, 1.234) assert_almost_equal(fdouble, 1.234) assert_almost_equal(flongdouble, 1.234) def test_nonnative_endian_fill(self, level=rlevel): """ Non-native endian arrays were incorrectly filled with scalars before r5034. """ if sys.byteorder == 'little': dtype = np.dtype('>i4') else: dtype = np.dtype('<i4') x = np.empty([1], dtype=dtype) x.fill(1) assert_equal(x, np.array([1], dtype=dtype)) def test_dot_alignment_sse2(self, level=rlevel): """Test for ticket #551, changeset r5140""" x = np.zeros((30, 40)) y = pickle.loads(pickle.dumps(x)) # y is now typically not aligned on a 8-byte boundary z = np.ones((1, y.shape[0])) # This shouldn't cause a segmentation fault: np.dot(z, y) def test_astype_copy(self, level=rlevel): """Ticket #788, changeset r5155""" # The test data file was generated by scipy.io.savemat. # The dtype is float64, but the isbuiltin attribute is 0. data_dir = path.join(path.dirname(__file__), 'data') filename = path.join(data_dir, "astype_copy.pkl") if sys.version_info[0] >= 3: f = open(filename, 'rb') xp = pickle.load(f, encoding='latin1') f.close() else: f = open(filename) xp = pickle.load(f) f.close() xpd = xp.astype(np.float64) assert_((xp.__array_interface__['data'][0] != xpd.__array_interface__['data'][0])) def test_compress_small_type(self, level=rlevel): """Ticket #789, changeset 5217. """ # compress with out argument segfaulted if cannot cast safely import numpy as np a = np.array([[1, 2], [3, 4]]) b = np.zeros((2, 1), dtype = np.single) try: a.compress([True, False], axis = 1, out = b) raise AssertionError("compress with an out which cannot be " \ "safely casted should not return "\ "successfully") except TypeError: pass def test_attributes(self, level=rlevel): """Ticket #791 """ class TestArray(np.ndarray): def __new__(cls, data, info): result = np.array(data) result = result.view(cls) result.info = info return result def __array_finalize__(self, obj): self.info = getattr(obj, 'info', '') dat = TestArray([[1, 2, 3, 4], [5, 6, 7, 8]], 'jubba') assert_(dat.info == 'jubba') dat.resize((4, 2)) assert_(dat.info == 'jubba') dat.sort() assert_(dat.info == 'jubba') dat.fill(2) assert_(dat.info == 'jubba') dat.put([2, 3, 4], [6, 3, 4]) assert_(dat.info == 'jubba') dat.setfield(4, np.int32, 0) assert_(dat.info == 'jubba') dat.setflags() assert_(dat.info == 'jubba') assert_(dat.all(1).info == 'jubba') assert_(dat.any(1).info == 'jubba') assert_(dat.argmax(1).info == 'jubba') assert_(dat.argmin(1).info == 'jubba') assert_(dat.argsort(1).info == 'jubba') assert_(dat.astype(TestArray).info == 'jubba') assert_(dat.byteswap().info == 'jubba') assert_(dat.clip(2, 7).info == 'jubba') assert_(dat.compress([0, 1, 1]).info == 'jubba') assert_(dat.conj().info == 'jubba') assert_(dat.conjugate().info == 'jubba') assert_(dat.copy().info == 'jubba') dat2 = TestArray([2, 3, 1, 0], 'jubba') choices = [[0, 1, 2, 3], [10, 11, 12, 13], [20, 21, 22, 23], [30, 31, 32, 33]] assert_(dat2.choose(choices).info == 'jubba') assert_(dat.cumprod(1).info == 'jubba') assert_(dat.cumsum(1).info == 'jubba') assert_(dat.diagonal().info == 'jubba') assert_(dat.flatten().info == 'jubba') assert_(dat.getfield(np.int32, 0).info == 'jubba') assert_(dat.imag.info == 'jubba') assert_(dat.max(1).info == 'jubba') assert_(dat.mean(1).info == 'jubba') assert_(dat.min(1).info == 'jubba') assert_(dat.newbyteorder().info == 'jubba') assert_(dat.nonzero()[0].info == 'jubba') assert_(dat.nonzero()[1].info == 'jubba') assert_(dat.prod(1).info == 'jubba') assert_(dat.ptp(1).info == 'jubba') assert_(dat.ravel().info == 'jubba') assert_(dat.real.info == 'jubba') assert_(dat.repeat(2).info == 'jubba') assert_(dat.reshape((2, 4)).info == 'jubba') assert_(dat.round().info == 'jubba') assert_(dat.squeeze().info == 'jubba') assert_(dat.std(1).info == 'jubba') assert_(dat.sum(1).info == 'jubba') assert_(dat.swapaxes(0, 1).info == 'jubba') assert_(dat.take([2, 3, 5]).info == 'jubba') assert_(dat.transpose().info == 'jubba') assert_(dat.T.info == 'jubba') assert_(dat.var(1).info == 'jubba') assert_(dat.view(TestArray).info == 'jubba') def test_recarray_tolist(self, level=rlevel): """Ticket #793, changeset r5215 """ # Comparisons fail for NaN, so we can't use random memory # for the test. buf = np.zeros(40, dtype=np.int8) a = np.recarray(2, formats="i4,f8,f8", names="id,x,y", buf=buf) b = a.tolist() assert_( a[0].tolist() == b[0]) assert_( a[1].tolist() == b[1]) def test_nonscalar_item_method(self): # Make sure that .item() fails graciously when it should a = np.arange(5) assert_raises(ValueError, a.item) def test_char_array_creation(self, level=rlevel): a = np.array('123', dtype='c') b = np.array(asbytes_nested(['1', '2', '3'])) assert_equal(a, b) def test_unaligned_unicode_access(self, level=rlevel) : """Ticket #825""" for i in range(1, 9) : msg = 'unicode offset: %d chars'%i t = np.dtype([('a', 'S%d'%i), ('b', 'U2')]) x = np.array([(asbytes('a'), sixu('b'))], dtype=t) if sys.version_info[0] >= 3: assert_equal(str(x), "[(b'a', 'b')]", err_msg=msg) else: assert_equal(str(x), "[('a', u'b')]", err_msg=msg) def test_sign_for_complex_nan(self, level=rlevel): """Ticket 794.""" with np.errstate(invalid='ignore'): C = np.array([-np.inf, -2+1j, 0, 2-1j, np.inf, np.nan]) have = np.sign(C) want = np.array([-1+0j, -1+0j, 0+0j, 1+0j, 1+0j, np.nan]) assert_equal(have, want) def test_for_equal_names(self, level=rlevel): """Ticket #674""" dt = np.dtype([('foo', float), ('bar', float)]) a = np.zeros(10, dt) b = list(a.dtype.names) b[0] = "notfoo" a.dtype.names = b assert_(a.dtype.names[0] == "notfoo") assert_(a.dtype.names[1] == "bar") def test_for_object_scalar_creation(self, level=rlevel): """Ticket #816""" a = np.object_() b = np.object_(3) b2 = np.object_(3.0) c = np.object_([4, 5]) d = np.object_([None, {}, []]) assert_(a is None) assert_(type(b) is int) assert_(type(b2) is float) assert_(type(c) is np.ndarray) assert_(c.dtype == object) assert_(d.dtype == object) def test_array_resize_method_system_error(self): """Ticket #840 - order should be an invalid keyword.""" x = np.array([[0, 1], [2, 3]]) self.assertRaises(TypeError, x.resize, (2, 2), order='C') def test_for_zero_length_in_choose(self, level=rlevel): "Ticket #882" a = np.array(1) self.assertRaises(ValueError, lambda x: x.choose([]), a) def test_array_ndmin_overflow(self): "Ticket #947." self.assertRaises(ValueError, lambda: np.array([1], ndmin=33)) def test_errobj_reference_leak(self, level=rlevel): """Ticket #955""" with np.errstate(all="ignore"): z = int(0) p = np.int32(-1) gc.collect() n_before = len(gc.get_objects()) z**p # this shouldn't leak a reference to errobj gc.collect() n_after = len(gc.get_objects()) assert_(n_before >= n_after, (n_before, n_after)) def test_void_scalar_with_titles(self, level=rlevel): """No ticket""" data = [('john', 4), ('mary', 5)] dtype1 = [(('source:yy', 'name'), 'O'), (('source:xx', 'id'), int)] arr = np.array(data, dtype=dtype1) assert_(arr[0][0] == 'john') assert_(arr[0][1] == 4) def test_void_scalar_constructor(self): #Issue #1550 #Create test string data, construct void scalar from data and assert #that void scalar contains original data. test_string = np.array("test") test_string_void_scalar = np.core.multiarray.scalar( np.dtype(("V", test_string.dtype.itemsize)), test_string.tobytes()) assert_(test_string_void_scalar.view(test_string.dtype) == test_string) #Create record scalar, construct from data and assert that #reconstructed scalar is correct. test_record = np.ones((), "i,i") test_record_void_scalar = np.core.multiarray.scalar( test_record.dtype, test_record.tobytes()) assert_(test_record_void_scalar == test_record) #Test pickle and unpickle of void and record scalars assert_(pickle.loads(pickle.dumps(test_string)) == test_string) assert_(pickle.loads(pickle.dumps(test_record)) == test_record) def test_blasdot_uninitialized_memory(self): """Ticket #950""" for m in [0, 1, 2]: for n in [0, 1, 2]: for k in range(3): # Try to ensure that x->data contains non-zero floats x = np.array([123456789e199], dtype=np.float64) x.resize((m, 0)) y = np.array([123456789e199], dtype=np.float64) y.resize((0, n)) # `dot` should just return zero (m,n) matrix z = np.dot(x, y) assert_(np.all(z == 0)) assert_(z.shape == (m, n)) def test_zeros(self): """Regression test for #1061.""" # Set a size which cannot fit into a 64 bits signed integer sz = 2 ** 64 good = 'Maximum allowed dimension exceeded' try: np.empty(sz) except ValueError as e: if not str(e) == good: self.fail("Got msg '%s', expected '%s'" % (e, good)) except Exception as e: self.fail("Got exception of type %s instead of ValueError" % type(e)) def test_huge_arange(self): """Regression test for #1062.""" # Set a size which cannot fit into a 64 bits signed integer sz = 2 ** 64 good = 'Maximum allowed size exceeded' try: a = np.arange(sz) self.assertTrue(np.size == sz) except ValueError as e: if not str(e) == good: self.fail("Got msg '%s', expected '%s'" % (e, good)) except Exception as e: self.fail("Got exception of type %s instead of ValueError" % type(e)) def test_fromiter_bytes(self): """Ticket #1058""" a = np.fromiter(list(range(10)), dtype='b') b = np.fromiter(list(range(10)), dtype='B') assert_(np.alltrue(a == np.array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]))) assert_(np.alltrue(b == np.array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]))) def test_array_from_sequence_scalar_array(self): """Ticket #1078: segfaults when creating an array with a sequence of 0d arrays.""" a = np.array((np.ones(2), np.array(2))) assert_equal(a.shape, (2,)) assert_equal(a.dtype, np.dtype(object)) assert_equal(a[0], np.ones(2)) assert_equal(a[1], np.array(2)) a = np.array(((1,), np.array(1))) assert_equal(a.shape, (2,)) assert_equal(a.dtype, np.dtype(object)) assert_equal(a[0], (1,)) assert_equal(a[1], np.array(1)) def test_array_from_sequence_scalar_array2(self): """Ticket #1081: weird array with strange input...""" t = np.array([np.array([]), np.array(0, object)]) assert_equal(t.shape, (2,)) assert_equal(t.dtype, np.dtype(object)) def test_array_too_big(self): """Ticket #1080.""" assert_raises(ValueError, np.zeros, [975]*7, np.int8) assert_raises(ValueError, np.zeros, [26244]*5, np.int8) def test_dtype_keyerrors_(self): """Ticket #1106.""" dt = np.dtype([('f1', np.uint)]) assert_raises(KeyError, dt.__getitem__, "f2") assert_raises(IndexError, dt.__getitem__, 1) assert_raises(ValueError, dt.__getitem__, 0.0) def test_lexsort_buffer_length(self): """Ticket #1217, don't segfault.""" a = np.ones(100, dtype=np.int8) b = np.ones(100, dtype=np.int32) i = np.lexsort((a[::-1], b)) assert_equal(i, np.arange(100, dtype=np.int)) def test_object_array_to_fixed_string(self): """Ticket #1235.""" a = np.array(['abcdefgh', 'ijklmnop'], dtype=np.object_) b = np.array(a, dtype=(np.str_, 8)) assert_equal(a, b) c = np.array(a, dtype=(np.str_, 5)) assert_equal(c, np.array(['abcde', 'ijklm'])) d = np.array(a, dtype=(np.str_, 12)) assert_equal(a, d) e = np.empty((2, ), dtype=(np.str_, 8)) e[:] = a[:] assert_equal(a, e) def test_unicode_to_string_cast(self): """Ticket #1240.""" a = np.array( [ [sixu('abc'), sixu('\u03a3')], [sixu('asdf'), sixu('erw')] ], dtype='U') def fail(): b = np.array(a, 'S4') self.assertRaises(UnicodeEncodeError, fail) def test_mixed_string_unicode_array_creation(self): a = np.array(['1234', sixu('123')]) assert_(a.itemsize == 16) a = np.array([sixu('123'), '1234']) assert_(a.itemsize == 16) a = np.array(['1234', sixu('123'), '12345']) assert_(a.itemsize == 20) a = np.array([sixu('123'), '1234', sixu('12345')]) assert_(a.itemsize == 20) a = np.array([sixu('123'), '1234', sixu('1234')]) assert_(a.itemsize == 16) def test_misaligned_objects_segfault(self): """Ticket #1198 and #1267""" a1 = np.zeros((10,), dtype='O,c') a2 = np.array(['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j'], 'S10') a1['f0'] = a2 r = repr(a1) np.argmax(a1['f0']) a1['f0'][1] = "FOO" a1['f0'] = "FOO" a3 = np.array(a1['f0'], dtype='S') np.nonzero(a1['f0']) a1.sort() a4 = copy.deepcopy(a1) def test_misaligned_scalars_segfault(self): """Ticket #1267""" s1 = np.array(('a', 'Foo'), dtype='c,O') s2 = np.array(('b', 'Bar'), dtype='c,O') s1['f1'] = s2['f1'] s1['f1'] = 'Baz' def test_misaligned_dot_product_objects(self): """Ticket #1267""" # This didn't require a fix, but it's worth testing anyway, because # it may fail if .dot stops enforcing the arrays to be BEHAVED a = np.array([[(1, 'a'), (0, 'a')], [(0, 'a'), (1, 'a')]], dtype='O,c') b = np.array([[(4, 'a'), (1, 'a')], [(2, 'a'), (2, 'a')]], dtype='O,c') np.dot(a['f0'], b['f0']) def test_byteswap_complex_scalar(self): """Ticket #1259 and gh-441""" for dtype in [np.dtype('<'+t) for t in np.typecodes['Complex']]: z = np.array([2.2-1.1j], dtype) x = z[0] # always native-endian y = x.byteswap() if x.dtype.byteorder == z.dtype.byteorder: # little-endian machine assert_equal(x, np.fromstring(y.tobytes(), dtype=dtype.newbyteorder())) else: # big-endian machine assert_equal(x, np.fromstring(y.tobytes(), dtype=dtype)) # double check real and imaginary parts: assert_equal(x.real, y.real.byteswap()) assert_equal(x.imag, y.imag.byteswap()) def test_structured_arrays_with_objects1(self): """Ticket #1299""" stra = 'aaaa' strb = 'bbbb' x = np.array([[(0, stra), (1, strb)]], 'i8,O') x[x.nonzero()] = x.ravel()[:1] assert_(x[0, 1] == x[0, 0]) def test_structured_arrays_with_objects2(self): """Ticket #1299 second test""" stra = 'aaaa' strb = 'bbbb' numb = sys.getrefcount(strb) numa = sys.getrefcount(stra) x = np.array([[(0, stra), (1, strb)]], 'i8,O') x[x.nonzero()] = x.ravel()[:1] assert_(sys.getrefcount(strb) == numb) assert_(sys.getrefcount(stra) == numa + 2) def test_duplicate_title_and_name(self): """Ticket #1254""" def func(): x = np.dtype([(('a', 'a'), 'i'), ('b', 'i')]) self.assertRaises(ValueError, func) def test_signed_integer_division_overflow(self): """Ticket #1317.""" def test_type(t): min = np.array([np.iinfo(t).min]) min //= -1 with np.errstate(divide="ignore"): for t in (np.int8, np.int16, np.int32, np.int64, np.int, np.long): test_type(t) def test_buffer_hashlib(self): try: from hashlib import md5 except ImportError: from md5 import new as md5 x = np.array([1, 2, 3], dtype=np.dtype('<i4')) assert_equal(md5(x).hexdigest(), '2a1dd1e1e59d0a384c26951e316cd7e6') def test_0d_string_scalar(self): # Bug #1436; the following should succeed np.asarray('x', '>c') def test_log1p_compiler_shenanigans(self): # Check if log1p is behaving on 32 bit intel systems. assert_(np.isfinite(np.log1p(np.exp2(-53)))) def test_fromiter_comparison(self, level=rlevel): a = np.fromiter(list(range(10)), dtype='b') b = np.fromiter(list(range(10)), dtype='B') assert_(np.alltrue(a == np.array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]))) assert_(np.alltrue(b == np.array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]))) def test_fromstring_crash(self): # Ticket #1345: the following should not cause a crash np.fromstring(asbytes('aa, aa, 1.0'), sep=',') def test_ticket_1539(self): dtypes = [x for x in np.typeDict.values() if (issubclass(x, np.number) and not issubclass(x, np.timedelta64))] a = np.array([], dtypes[0]) failures = [] # ignore complex warnings with warnings.catch_warnings(): warnings.simplefilter('ignore', np.ComplexWarning) for x in dtypes: b = a.astype(x) for y in dtypes: c = a.astype(y) try: np.dot(b, c) except TypeError as e: failures.append((x, y)) if failures: raise AssertionError("Failures: %r" % failures) def test_ticket_1538(self): x = np.finfo(np.float32) for name in 'eps epsneg max min resolution tiny'.split(): assert_equal(type(getattr(x, name)), np.float32, err_msg=name) def test_ticket_1434(self): # Check that the out= argument in var and std has an effect data = np.array(((1, 2, 3), (4, 5, 6), (7, 8, 9))) out = np.zeros((3,)) ret = data.var(axis=1, out=out) assert_(ret is out) assert_array_equal(ret, data.var(axis=1)) ret = data.std(axis=1, out=out) assert_(ret is out) assert_array_equal(ret, data.std(axis=1)) def test_complex_nan_maximum(self): cnan = complex(0, np.nan) assert_equal(np.maximum(1, cnan), cnan) def test_subclass_int_tuple_assignment(self): # ticket #1563 class Subclass(np.ndarray): def __new__(cls, i): return np.ones((i,)).view(cls) x = Subclass(5) x[(0,)] = 2 # shouldn't raise an exception assert_equal(x[0], 2) def test_ufunc_no_unnecessary_views(self): # ticket #1548 class Subclass(np.ndarray): pass x = np.array([1, 2, 3]).view(Subclass) y = np.add(x, x, x) assert_equal(id(x), id(y)) def test_take_refcount(self): # ticket #939 a = np.arange(16, dtype=np.float) a.shape = (4, 4) lut = np.ones((5 + 3, 4), np.float) rgba = np.empty(shape=a.shape + (4,), dtype=lut.dtype) c1 = sys.getrefcount(rgba) try: lut.take(a, axis=0, mode='clip', out=rgba) except TypeError: pass c2 = sys.getrefcount(rgba) assert_equal(c1, c2) def test_fromfile_tofile_seeks(self): # On Python 3, tofile/fromfile used to get (#1610) the Python # file handle out of sync f0 = tempfile.NamedTemporaryFile() f = f0.file f.write(np.arange(255, dtype='u1').tobytes()) f.seek(20) ret = np.fromfile(f, count=4, dtype='u1') assert_equal(ret, np.array([20, 21, 22, 23], dtype='u1')) assert_equal(f.tell(), 24) f.seek(40) np.array([1, 2, 3], dtype='u1').tofile(f) assert_equal(f.tell(), 43) f.seek(40) data = f.read(3) assert_equal(data, asbytes("\x01\x02\x03")) f.seek(80) f.read(4) data = np.fromfile(f, dtype='u1', count=4) assert_equal(data, np.array([84, 85, 86, 87], dtype='u1')) f.close() def test_complex_scalar_warning(self): for tp in [np.csingle, np.cdouble, np.clongdouble]: x = tp(1+2j) assert_warns(np.ComplexWarning, float, x) with warnings.catch_warnings(): warnings.simplefilter('ignore') assert_equal(float(x), float(x.real)) def test_complex_scalar_complex_cast(self): for tp in [np.csingle, np.cdouble, np.clongdouble]: x = tp(1+2j) assert_equal(complex(x), 1+2j) def test_complex_boolean_cast(self): """Ticket #2218""" for tp in [np.csingle, np.cdouble, np.clongdouble]: x = np.array([0, 0+0.5j, 0.5+0j], dtype=tp) assert_equal(x.astype(bool), np.array([0, 1, 1], dtype=bool)) assert_(np.any(x)) assert_(np.all(x[1:])) def test_uint_int_conversion(self): x = 2**64 - 1 assert_equal(int(np.uint64(x)), x) def test_duplicate_field_names_assign(self): ra = np.fromiter(((i*3, i*2) for i in range(10)), dtype='i8,f8') ra.dtype.names = ('f1', 'f2') rep = repr(ra) # should not cause a segmentation fault assert_raises(ValueError, setattr, ra.dtype, 'names', ('f1', 'f1')) def test_eq_string_and_object_array(self): # From e-mail thread "__eq__ with str and object" (Keith Goodman) a1 = np.array(['a', 'b'], dtype=object) a2 = np.array(['a', 'c']) assert_array_equal(a1 == a2, [True, False]) assert_array_equal(a2 == a1, [True, False]) def test_nonzero_byteswap(self): a = np.array([0x80000000, 0x00000080, 0], dtype=np.uint32) a.dtype = np.float32 assert_equal(a.nonzero()[0], [1]) a = a.byteswap().newbyteorder() assert_equal(a.nonzero()[0], [1]) # [0] if nonzero() ignores swap def test_find_common_type_boolean(self): # Ticket #1695 assert_(np.find_common_type([], ['?', '?']) == '?') def test_empty_mul(self): a = np.array([1.]) a[1:1] *= 2 assert_equal(a, [1.]) def test_array_side_effect(self): assert_equal(np.dtype('S10').itemsize, 10) A = np.array([['abc', 2], ['long ', '0123456789']], dtype=np.string_) # This was throwing an exception because in ctors.c, # discover_itemsize was calling PyObject_Length without checking # the return code. This failed to get the length of the number 2, # and the exception hung around until something checked # PyErr_Occurred() and returned an error. assert_equal(np.dtype('S10').itemsize, 10) def test_any_float(self): # all and any for floats a = np.array([0.1, 0.9]) assert_(np.any(a)) assert_(np.all(a)) def test_large_float_sum(self): a = np.arange(10000, dtype='f') assert_equal(a.sum(dtype='d'), a.astype('d').sum()) def test_ufunc_casting_out(self): a = np.array(1.0, dtype=np.float32) b = np.array(1.0, dtype=np.float64) c = np.array(1.0, dtype=np.float32) np.add(a, b, out=c) assert_equal(c, 2.0) def test_array_scalar_contiguous(self): # Array scalars are both C and Fortran contiguous assert_(np.array(1.0).flags.c_contiguous) assert_(np.array(1.0).flags.f_contiguous) assert_(np.array(np.float32(1.0)).flags.c_contiguous) assert_(np.array(np.float32(1.0)).flags.f_contiguous) def test_squeeze_contiguous(self): """Similar to GitHub issue #387""" a = np.zeros((1, 2)).squeeze() b = np.zeros((2, 2, 2), order='F')[:,:, ::2].squeeze() assert_(a.flags.c_contiguous) assert_(a.flags.f_contiguous) assert_(b.flags.f_contiguous) def test_reduce_contiguous(self): """GitHub issue #387""" a = np.add.reduce(np.zeros((2, 1, 2)), (0, 1)) b = np.add.reduce(np.zeros((2, 1, 2)), 1) assert_(a.flags.c_contiguous) assert_(a.flags.f_contiguous) assert_(b.flags.c_contiguous) def test_object_array_self_reference(self): # Object arrays with references to themselves can cause problems a = np.array(0, dtype=object) a[()] = a assert_raises(TypeError, int, a) assert_raises(TypeError, long, a) assert_raises(TypeError, float, a) assert_raises(TypeError, oct, a) assert_raises(TypeError, hex, a) # Test the same for a circular reference. b = np.array(a, dtype=object) a[()] = b assert_raises(TypeError, int, a) # Numpy has no tp_traverse currently, so circular references # cannot be detected. So resolve it: a[()] = 0 # This was causing a to become like the above a = np.array(0, dtype=object) a[...] += 1 assert_equal(a, 1) def test_object_array_self_copy(self): # An object array being copied into itself DECREF'ed before INCREF'ing # causing segmentation faults (gh-3787) a = np.array(object(), dtype=object) np.copyto(a, a) assert_equal(sys.getrefcount(a[()]), 2) a[()].__class__ # will segfault if object was deleted def test_zerosize_accumulate(self): "Ticket #1733" x = np.array([[42, 0]], dtype=np.uint32) assert_equal(np.add.accumulate(x[:-1, 0]), []) def test_objectarray_setfield(self): # Setfield directly manipulates the raw array data, # so is invalid for object arrays. x = np.array([1, 2, 3], dtype=object) assert_raises(RuntimeError, x.setfield, 4, np.int32, 0) def test_setting_rank0_string(self): "Ticket #1736" s1 = asbytes("hello1") s2 = asbytes("hello2") a = np.zeros((), dtype="S10") a[()] = s1 assert_equal(a, np.array(s1)) a[()] = np.array(s2) assert_equal(a, np.array(s2)) a = np.zeros((), dtype='f4') a[()] = 3 assert_equal(a, np.array(3)) a[()] = np.array(4) assert_equal(a, np.array(4)) def test_string_astype(self): "Ticket #1748" s1 = asbytes('black') s2 = asbytes('white') s3 = asbytes('other') a = np.array([[s1], [s2], [s3]]) assert_equal(a.dtype, np.dtype('S5')) b = a.astype(np.dtype('S0')) assert_equal(b.dtype, np.dtype('S5')) def test_ticket_1756(self): """Ticket #1756 """ s = asbytes('0123456789abcdef') a = np.array([s]*5) for i in range(1, 17): a1 = np.array(a, "|S%d"%i) a2 = np.array([s[:i]]*5) assert_equal(a1, a2) def test_fields_strides(self): "Ticket #1760" r=np.fromstring('abcdefghijklmnop'*4*3, dtype='i4,(2,3)u2') assert_equal(r[0:3:2]['f1'], r['f1'][0:3:2]) assert_equal(r[0:3:2]['f1'][0], r[0:3:2][0]['f1']) assert_equal(r[0:3:2]['f1'][0][()], r[0:3:2][0]['f1'][()]) assert_equal(r[0:3:2]['f1'][0].strides, r[0:3:2][0]['f1'].strides) def test_alignment_update(self): """Check that alignment flag is updated on stride setting""" a = np.arange(10) assert_(a.flags.aligned) a.strides = 3 assert_(not a.flags.aligned) def test_ticket_1770(self): "Should not segfault on python 3k" import numpy as np try: a = np.zeros((1,), dtype=[('f1', 'f')]) a['f1'] = 1 a['f2'] = 1 except ValueError: pass except: raise AssertionError def test_ticket_1608(self): "x.flat shouldn't modify data" x = np.array([[1, 2], [3, 4]]).T y = np.array(x.flat) assert_equal(x, [[1, 3], [2, 4]]) def test_pickle_string_overwrite(self): import re data = np.array([1], dtype='b') blob = pickle.dumps(data, protocol=1) data = pickle.loads(blob) # Check that loads does not clobber interned strings s = re.sub("a(.)", "\x01\\1", "a_") assert_equal(s[0], "\x01") data[0] = 0xbb s = re.sub("a(.)", "\x01\\1", "a_") assert_equal(s[0], "\x01") def test_pickle_bytes_overwrite(self): if sys.version_info[0] >= 3: data = np.array([1], dtype='b') data = pickle.loads(pickle.dumps(data)) data[0] = 0xdd bytestring = "\x01 ".encode('ascii') assert_equal(bytestring[0:1], '\x01'.encode('ascii')) def test_structured_type_to_object(self): a_rec = np.array([(0, 1), (3, 2)], dtype='i4,i8') a_obj = np.empty((2,), dtype=object) a_obj[0] = (0, 1) a_obj[1] = (3, 2) # astype records -> object assert_equal(a_rec.astype(object), a_obj) # '=' records -> object b = np.empty_like(a_obj) b[...] = a_rec assert_equal(b, a_obj) # '=' object -> records b = np.empty_like(a_rec) b[...] = a_obj assert_equal(b, a_rec) def test_assign_obj_listoflists(self): # Ticket # 1870 # The inner list should get assigned to the object elements a = np.zeros(4, dtype=object) b = a.copy() a[0] = [1] a[1] = [2] a[2] = [3] a[3] = [4] b[...] = [[1], [2], [3], [4]] assert_equal(a, b) # The first dimension should get broadcast a = np.zeros((2, 2), dtype=object) a[...] = [[1, 2]] assert_equal(a, [[1, 2], [1, 2]]) def test_memoryleak(self): # Ticket #1917 - ensure that array data doesn't leak for i in range(1000): # 100MB times 1000 would give 100GB of memory usage if it leaks a = np.empty((100000000,), dtype='i1') del a def test_ufunc_reduce_memoryleak(self): a = np.arange(6) acnt = sys.getrefcount(a) res = np.add.reduce(a) assert_equal(sys.getrefcount(a), acnt) def test_search_sorted_invalid_arguments(self): # Ticket #2021, should not segfault. x = np.arange(0, 4, dtype='datetime64[D]') assert_raises(TypeError, x.searchsorted, 1) def test_string_truncation(self): # Ticket #1990 - Data can be truncated in creation of an array from a # mixed sequence of numeric values and strings for val in [True, 1234, 123.4, complex(1, 234)]: for tostr in [asunicode, asbytes]: b = np.array([val, tostr('xx')]) assert_equal(tostr(b[0]), tostr(val)) b = np.array([tostr('xx'), val]) assert_equal(tostr(b[1]), tostr(val)) # test also with longer strings b = np.array([val, tostr('xxxxxxxxxx')]) assert_equal(tostr(b[0]), tostr(val)) b = np.array([tostr('xxxxxxxxxx'), val]) assert_equal(tostr(b[1]), tostr(val)) def test_string_truncation_ucs2(self): # Ticket #2081. Python compiled with two byte unicode # can lead to truncation if itemsize is not properly # adjusted for Numpy's four byte unicode. if sys.version_info[0] >= 3: a = np.array(['abcd']) else: a = np.array([sixu('abcd')]) assert_equal(a.dtype.itemsize, 16) def test_unique_stable(self): # Ticket #2063 must always choose stable sort for argsort to # get consistent results v = np.array(([0]*5 + [1]*6 + [2]*6)*4) res = np.unique(v, return_index=True) tgt = (np.array([0, 1, 2]), np.array([ 0, 5, 11])) assert_equal(res, tgt) def test_unicode_alloc_dealloc_match(self): # Ticket #1578, the mismatch only showed up when running # python-debug for python versions >= 2.7, and then as # a core dump and error message. a = np.array(['abc'], dtype=np.unicode)[0] del a def test_refcount_error_in_clip(self): # Ticket #1588 a = np.zeros((2,), dtype='>i2').clip(min=0) x = a + a # This used to segfault: y = str(x) # Check the final string: assert_(y == "[0 0]") def test_searchsorted_wrong_dtype(self): # Ticket #2189, it used to segfault, so we check that it raises the # proper exception. a = np.array([('a', 1)], dtype='S1, int') assert_raises(TypeError, np.searchsorted, a, 1.2) # Ticket #2066, similar problem: dtype = np.format_parser(['i4', 'i4'], [], []) a = np.recarray((2, ), dtype) assert_raises(TypeError, np.searchsorted, a, 1) def test_complex64_alignment(self): # Issue gh-2668 (trac 2076), segfault on sparc due to misalignment dtt = np.complex64 arr = np.arange(10, dtype=dtt) # 2D array arr2 = np.reshape(arr, (2, 5)) # Fortran write followed by (C or F) read caused bus error data_str = arr2.tobytes('F') data_back = np.ndarray(arr2.shape, arr2.dtype, buffer=data_str, order='F') assert_array_equal(arr2, data_back) def test_structured_count_nonzero(self): arr = np.array([0, 1]).astype('i4, (2)i4')[:1] count = np.count_nonzero(arr) assert_equal(count, 0) def test_copymodule_preserves_f_contiguity(self): a = np.empty((2, 2), order='F') b = copy.copy(a) c = copy.deepcopy(a) assert_(b.flags.fortran) assert_(b.flags.f_contiguous) assert_(c.flags.fortran) assert_(c.flags.f_contiguous) def test_fortran_order_buffer(self): import numpy as np a = np.array([['Hello', 'Foob']], dtype='U5', order='F') arr = np.ndarray(shape=[1, 2, 5], dtype='U1', buffer=a) arr2 = np.array([[[sixu('H'), sixu('e'), sixu('l'), sixu('l'), sixu('o')], [sixu('F'), sixu('o'), sixu('o'), sixu('b'), sixu('')]]]) assert_array_equal(arr, arr2) def test_assign_from_sequence_error(self): # Ticket #4024. arr = np.array([1, 2, 3]) assert_raises(ValueError, arr.__setitem__, slice(None), [9, 9]) arr.__setitem__(slice(None), [9]) assert_equal(arr, [9, 9, 9]) def test_format_on_flex_array_element(self): # Ticket #4369. dt = np.dtype([('date', '<M8[D]'), ('val', '<f8')]) arr = np.array([('2000-01-01', 1)], dt) formatted = '{0}'.format(arr[0]) assert_equal(formatted, str(arr[0])) def test_deepcopy_on_0d_array(self): # Ticket #3311. arr = np.array(3) arr_cp = copy.deepcopy(arr) assert_equal(arr, arr_cp) assert_equal(arr.shape, arr_cp.shape) assert_equal(int(arr), int(arr_cp)) self.assertTrue(arr is not arr_cp) self.assertTrue(isinstance(arr_cp, type(arr))) if __name__ == "__main__": run_module_suite() ```
[ { "content": "```python\n##\n# Copyright 2012-2013 Ghent University\n#\n# This file is part of EasyBuild,\n# originally created by the HPC team of Ghent University (http://ugent.be/hpc/en),\n# with support of Ghent University (http://ugent.be/hpc),\n# the Flemish Supercomputer Centre (VSC) (https://vscentrum.be...
[ { "content": "<|memory_start|>```python\n##\n# Copyright 2012-2013 Ghent University\n#\n# This file is part of EasyBuild,\n# originally created by the HPC team of Ghent University (http://ugent.be/hpc/en),\n# with support of Ghent University (http://ugent.be/hpc),\n# the Flemish Supercomputer Centre (VSC) (http...
```python ## # Copyright 2012-2013 Ghent University # # This file is part of EasyBuild, # originally created by the HPC team of Ghent University (http://ugent.be/hpc/en), # with support of Ghent University (http://ugent.be/hpc), # the Flemish Supercomputer Centre (VSC) (https://vscentrum.be/nl/en), # the Hercules foundation (http://www.herculesstichting.be/in_English) # and the Department of Economy, Science and Innovation (EWI) (http://www.ewi-vlaanderen.be/en). # # http://github.com/hpcugent/easybuild # # EasyBuild is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation v2. # # EasyBuild is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with EasyBuild. If not, see <http://www.gnu.org/licenses/>. ## """ This script can be used to install easybuild-easyconfigs, e.g. using: easy_install --user . or python setup.py --prefix=$HOME/easybuild @author: Kenneth Hoste (Ghent University) """ import glob import os import re import shutil import sys from distutils import log # note: release candidates should be versioned as a pre-release, e.g. "1.1rc1" # 1.1-rc1 would indicate a post-release, i.e., and update of 1.1, so beware! VERSION = "1.13.0.0" API_VERSION = VERSION.split('.')[0] EB_VERSION = '.'.join(VERSION.split('.')[0:2]) suff = '' rc_regexp = re.compile("^.*(rc[0-9]*)$") res = rc_regexp.search(str(VERSION)) if res: suff = res.group(1) dev_regexp = re.compile("^.*[0-9]dev$") if dev_regexp.match(VERSION): suff = 'dev' API_VERSION += suff EB_VERSION += suff # log levels: 0 = WARN (default), 1 = INFO, 2 = DEBUG log.set_verbosity(1) # try setuptools, fall back to distutils if needed try: from setuptools import setup log.info("Installing with setuptools.setup...") install_package = 'setuptools' except ImportError, err: log.info("Failed to import setuptools.setup (%s), so falling back to distutils.setup" % err) from distutils.core import setup install_package = 'distutils' # utility function to read README file def read(fname): """Read contents of given file.""" return open(os.path.join(os.path.dirname(__file__), fname)).read() # utility function to get list of data files (i.e. easyconfigs) def get_data_files(): """ Return list of data files, i.e. easyconfigs, patches, etc., and retain directory structure. """ data_files = [] for dirname,dirs,files in os.walk(os.path.join('easybuild', 'easyconfigs')): if files: data_files.append((dirname, [os.path.join(dirname, f) for f in files])) return data_files log.info("Installing version %s (required versions: API >= %s, easyblocks >= %s)" % (VERSION, API_VERSION, EB_VERSION)) setup( name = "easybuild-easyconfigs", version = VERSION, author = "EasyBuild community", author_email = "easybuild@lists.ugent.be", description = """EasyBuild is a software installation framework in Python that allows you to \ install software in a structured and robust way. This package contains a collection of easyconfigs, i.e. simple text files written in Python syntax \ that specify the build parameters for software packages (version, compiler toolchain, dependency \ versions, etc.)""", license = "GPLv2", keywords = "software build building installation installing compilation HPC scientific", url = "http://hpcugent.github.com/easybuild", data_files = get_data_files(), long_description = read("README.rst"), classifiers = [ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: System Administrators", "License :: OSI Approved :: GNU General Public License v2 (GPLv2)", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.4", "Topic :: Software Development :: Build Tools", ], platforms = "Linux", # install_requires list is not enforced, because of 'old-and-unmanageable' setup? # do we even want the dependency, since it's artificial? install_requires = [ "easybuild-framework >= %s" % API_VERSION, "easybuild-easyblocks >= %s" % EB_VERSION ], zip_safe = False ) ```
[ { "content": "Here is a code file:\n```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2013 S. Daniel Francis <francis@sugarlabs.org>\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# t...
[ { "content": "Here is a code file:\n<|memory_start|>```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2013 S. Daniel Francis <francis@sugarlabs.org>\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as p...
```python #!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2013 S. Daniel Francis <francis@sugarlabs.org> # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, # MA 02110-1301, USA. from distutils.core import setup from distutils.command.build import build from distutils.core import Command import os srcdir = os.path.dirname(os.path.abspath(__file__)) docdir = os.path.join(srcdir, 'doc') docgettextdir = os.path.join(docdir, 'gettext') mandir = os.path.join(docdir, 'man') data_files = [] class build_manpage(Command): description = 'Generate man pages.' user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): for i in os.listdir(docgettextdir) + ['man1.1']: name, ext = i.split('.') if ext != 'pot' and name: build_dir = os.path.join(mandir, name) if not os.path.exists(build_dir): os.makedirs(build_dir) langopt = ('-Dlanguage=%s' % name) if name != 'man1' else '' print 'Generating %s/dictate.1.gz' % build_dir os.system('sphinx-build -b man %s %s %s' % (langopt, docdir, build_dir)) if os.path.exists('%s/dictate.1.gz' % build_dir): os.remove('%s/dictate.1.gz' % build_dir) os.system('gzip %s/*.1' % build_dir) self.install_man('doc/man') def install_man(self, directory): for i in os.listdir(directory): path = os.path.join(directory, i) if os.path.isdir(path) and i != '.doctrees': install_path = os.path.join('share', 'man', i, 'man1') if i == 'man1': install_path = os.path.join('share', 'man', 'man1') files = [] for filename in os.listdir(path): if filename.split('.')[-1] == 'gz': files.append(os.path.join(path, filename)) data_files.append((install_path, files)) class build_trans(Command): description = 'Compile .po files into .mo files' def initialize_options(self): pass def finalize_options(self): pass def run(self): self._srcdir = os.path.join(os.path.abspath(os.curdir)) translations = [(os.path.join(self._srcdir, 'po'), os.path.join(self._srcdir, 'locale'), 'dictate'), (os.path.join(self._srcdir, 'doc', 'gettext'), os.path.join(self._srcdir, 'doc', 'locale'), 'index')] for po_dir, locale_dir, module in translations: os.system('%s/i18nhelpers/buildmo.py %s %s %s' % (srcdir, po_dir, locale_dir, module)) self.append_mo(translations[0][1]) def append_mo(self, directory): for lang in os.listdir(directory): lang_dir = os.path.join('share', 'locale', lang, 'LC_MESSAGES') lang_file = os.path.join(self._srcdir, 'locale', lang, 'LC_MESSAGES', 'dictate.mo') data_files.append((lang_dir, [lang_file])) build.sub_commands.append(('build_trans', None)) build.sub_commands.append(('build_manpage', None)) setup(name='dictate', version='0.3', description='Command-line dictation utility.', author='Daniel Francis', author_email='francis@sugarlabs.org', license='GPLv3', url='https://github.com/sdanielf/dictate/', packages=['dictation'], scripts=['dictate'], cmdclass={'build_manpage': build_manpage, 'build_trans': build_trans}, data_files=data_files, long_description="""Dictation is an eSpeak-based dictation utility. It reads a text slowly, allowing users to write it. Also can pause the dictation, spell difficult words and identify punctuation marks.""", classifiers=['Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Education', 'License :: OSI Approved :: GNU General Public License v3 \ or later (GPLv3+)', 'Operating System :: POSIX', 'Programming Language :: Python :: 2.7', 'Topic :: Education', 'Topic :: Multimedia :: Sound/Audio :: Speech', 'Topic :: Utilities']) ```
[ { "content": "Here is the script:\n```python\n#!/bin/env python3.1\n# Bradley N. Miller, David L. Ranum\n# Introduction to Data Structures and Algorithms in Python\n# Copyright 2005, 2010\n# \n\nclass BinarySearchTree:\n '''\n Author: Brad Miller\n Date: 1/15/2005\n Description: Imlement a binary...
[ { "content": "Here is the script:\n<|memory_start|>```python\n#!/bin/env python3.1\n# Bradley N. Miller, David L. Ranum\n# Introduction to Data Structures and Algorithms in Python\n# Copyright 2005, 2010\n# \n\nclass BinarySearchTree:\n '''\n Author: Brad Miller\n Date: 1/15/2005\n Description: I...
```python #!/bin/env python3.1 # Bradley N. Miller, David L. Ranum # Introduction to Data Structures and Algorithms in Python # Copyright 2005, 2010 # class BinarySearchTree: ''' Author: Brad Miller Date: 1/15/2005 Description: Imlement a binary search tree with the following interface functions: __contains__(y) <==> y in x __getitem__(y) <==> x[y] __init__() __len__() <==> len(x) __setitem__(k,v) <==> x[k] = v clear() get(k) items() keys() values() put(k,v) in del <==> ''' def __init__(self): self.root = None self.size = 0 def put(self,key,val): if self.root: self._put(key,val,self.root) else: self.root = TreeNode(key,val) self.size = self.size + 1 def _put(self,key,val,currentNode): if key < currentNode.key: if currentNode.hasLeftChild(): self._put(key,val,currentNode.leftChild) else: currentNode.leftChild = TreeNode(key,val,parent=currentNode) else: if currentNode.hasRightChild(): self._put(key,val,currentNode.rightChild) else: currentNode.rightChild = TreeNode(key,val,parent=currentNode) def __setitem__(self,k,v): self.put(k,v) def get(self,key): if self.root: res = self._get(key,self.root) if res: return res.payload else: return None else: return None def _get(self,key,currentNode): if not currentNode: return None elif currentNode.key == key: return currentNode elif key < currentNode.key: return self._get(key,currentNode.leftChild) else: return self._get(key,currentNode.rightChild) def __getitem__(self,key): res = self.get(key) if res: return res else: raise KeyError('Error, key not in tree') def __contains__(self,key): if self._get(key,self.root): return True else: return False def length(self): return self.size def __len__(self): return self.size def __iter__(self): return self.root.__iter__() def delete(self,key): if self.size > 1: nodeToRemove = self._get(key,self.root) if nodeToRemove: self.remove(nodeToRemove) self.size = self.size-1 else: raise KeyError('Error, key not in tree') elif self.size == 1 and self.root.key == key: self.root = None self.size = self.size - 1 else: raise KeyError('Error, key not in tree') def __delitem__(self,key): self.delete(key) def remove(self,currentNode): if currentNode.isLeaf(): #leaf if currentNode == currentNode.parent.leftChild: currentNode.parent.leftChild = None else: currentNode.parent.rightChild = None elif currentNode.hasBothChildren(): #interior succ = currentNode.findSuccessor() succ.spliceOut() currentNode.key = succ.key currentNode.payload = succ.payload else: # this node has one child if currentNode.hasLeftChild(): if currentNode.isLeftChild(): currentNode.leftChild.parent = currentNode.parent currentNode.parent.leftChild = currentNode.leftChild elif currentNode.isRightChild(): currentNode.leftChild.parent = currentNode.parent currentNode.parent.rightChild = currentNode.leftChild else: currentNode.replaceNodeData(currentNode.leftChild.key, currentNode.leftChild.payload, currentNode.leftChild.leftChild, currentNode.leftChild.rightChild) else: if currentNode.isLeftChild(): currentNode.rightChild.parent = currentNode.parent currentNode.parent.leftChild = currentNode.rightChild elif currentNode.isRightChild(): currentNode.rightChild.parent = currentNode.parent currentNode.parent.rightChild = currentNode.rightChild else: currentNode.replaceNodeData(currentNode.rightChild.key, currentNode.rightChild.payload, currentNode.rightChild.leftChild, currentNode.rightChild.rightChild) def inorder(self): self._inorder(self.root) def _inorder(self,tree): if tree != None: self._inorder(tree.leftChild) print(tree.key) self._inorder(tree.rightChild) def postorder(self): self._postorder(self.root) def _postorder(self, tree): if tree: self._postorder(tree.rightChild) self._postorder(tree.leftChild) print(tree.key) def preorder(self): self._preorder(self,self.root) def _preorder(self,tree): if tree: print(tree.key) self._preorder(tree.leftChild) self._preorder(tree.rightChild) class TreeNode: def __init__(self,key,val,left=None,right=None,parent=None): self.key = key self.payload = val self.leftChild = left self.rightChild = right self.parent = parent self.balanceFactor = 0 def hasLeftChild(self): return self.leftChild def hasRightChild(self): return self.rightChild def isLeftChild(self): return self.parent and self.parent.leftChild == self def isRightChild(self): return self.parent and self.parent.rightChild == self def isRoot(self): return not self.parent def isLeaf(self): return not (self.rightChild or self.leftChild) def hasAnyChildren(self): return self.rightChild or self.leftChild def hasBothChildren(self): return self.rightChild and self.leftChild def replaceNodeData(self,key,value,lc,rc): self.key = key self.payload = value self.leftChild = lc self.rightChild = rc if self.hasLeftChild(): self.leftChild.parent = self if self.hasRightChild(): self.rightChild.parent = self def findSuccessor(self): succ = None if self.hasRightChild(): succ = self.rightChild.findMin() else: if self.parent: if self.isLeftChild(): succ = self.parent else: self.parent.rightChild = None succ = self.parent.findSuccessor() self.parent.rightChild = self return succ def spliceOut(self): if self.isLeaf(): if self.isLeftChild(): self.parent.leftChild = None else: self.parent.rightChild = None elif self.hasAnyChildren(): if self.hasLeftChild(): if self.isLeftChild(): self.parent.leftChild = self.leftChild else: self.parent.rightChild = self.leftChild self.leftChild.parent = self.parent else: if self.isLeftChild(): self.parent.leftChild = self.rightChild else: self.parent.rightChild = self.rightChild self.rightChild.parent = self.parent def findMin(self): current = self while current.hasLeftChild(): current = current.leftChild return current def __iter__(self): """The standard inorder traversal of a binary tree.""" if self: if self.hasLeftChild(): for elem in self.leftChild: yield elem yield self.key if self.hasRightChild(): for elem in self.rightChild: yield elem ```
[ { "content": "Here is a code snippet:\n```python\n# -*- coding: utf-8 -*-\n'''\nThe function in `vdirsyncer.sync` can be called on two instances of `Storage`\nto synchronize them. Due to the abstract API storage classes are implementing,\nthe two given instances don't have to be of the same exact type. This all...
[ { "content": "Here is a code snippet:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n'''\nThe function in `vdirsyncer.sync` can be called on two instances of `Storage`\nto synchronize them. Due to the abstract API storage classes are implementing,\nthe two given instances don't have to be of the same exac...
```python # -*- coding: utf-8 -*- ''' The function in `vdirsyncer.sync` can be called on two instances of `Storage` to synchronize them. Due to the abstract API storage classes are implementing, the two given instances don't have to be of the same exact type. This allows us not only to synchronize a local vdir with a CalDAV server, but also synchronize two CalDAV servers or two local vdirs. The algorithm is based on the blogpost "How OfflineIMAP works" by Edward Z. Yang. http://blog.ezyang.com/2012/08/how-offlineimap-works/ ''' import itertools from . import exceptions, log from .utils import uniq from .utils.compat import iteritems, text_type sync_logger = log.get(__name__) class SyncError(exceptions.Error): '''Errors related to synchronization.''' class SyncConflict(SyncError): ''' Two items changed since the last sync, they now have different contents and no conflict resolution method was given. :param ident: The ident of the item. :param href_a: The item's href on side A. :param href_b: The item's href on side B. ''' ident = None href_a = None href_b = None class IdentConflict(SyncError): ''' Multiple items on the same storage have the same UID. :param storage: The affected storage. :param hrefs: List of affected hrefs on `storage`. ''' storage = None _hrefs = None @property def hrefs(self): return self._hrefs @hrefs.setter def hrefs(self, val): val = set(val) assert len(val) > 1 self._hrefs = val class StorageEmpty(SyncError): ''' One storage unexpectedly got completely empty between two synchronizations. The first argument is the empty storage. :param empty_storage: The empty :py:class:`vdirsyncer.storage.base.Storage`. ''' empty_storage = None class BothReadOnly(SyncError): ''' Both storages are marked as read-only. Synchronization is therefore not possible. ''' class StorageInfo(object): '''A wrapper class that holds prefetched items, the status and other things.''' def __init__(self, storage, status): ''' :param status: {ident: (href, etag)} ''' self.storage = storage self.status = status self.idents = None def prepare_idents(self, other_read_only): href_to_status = dict((href, (ident, etag)) for ident, (href, etag) in iteritems(self.status)) hrefs_to_download = [] self.idents = {} for href, etag in self.storage.list(): if href in href_to_status: ident, old_etag = href_to_status[href] self.idents[ident] = { 'etag': etag, 'href': href, 'ident': ident } if etag != old_etag and not other_read_only: hrefs_to_download.append(href) else: hrefs_to_download.append(href) # Prefetch items for href, item, etag in (self.storage.get_multi(hrefs_to_download) if hrefs_to_download else ()): props = self.idents.setdefault(item.ident, {}) props['item'] = item props['ident'] = item.ident if props.setdefault('href', href) != href: raise IdentConflict(storage=self.storage, hrefs=[props['href'], href]) if props.setdefault('etag', etag) != etag: raise SyncError('Etag changed during sync.') def sync(storage_a, storage_b, status, conflict_resolution=None, force_delete=False): '''Synchronizes two storages. :param storage_a: The first storage :type storage_a: :class:`vdirsyncer.storage.base.Storage` :param storage_b: The second storage :type storage_b: :class:`vdirsyncer.storage.base.Storage` :param status: {ident: (href_a, etag_a, href_b, etag_b)} metadata about the two storages for detection of changes. Will be modified by the function and should be passed to it at the next sync. If this is the first sync, an empty dictionary should be provided. :param conflict_resolution: Either 'a wins' or 'b wins'. If none is provided, the sync function will raise :py:exc:`SyncConflict`. :param force_delete: When one storage got completely emptied between two syncs, :py:exc:`StorageEmpty` is raised for safety. Setting this parameter to ``True`` disables this safety measure. ''' if storage_a.read_only and storage_b.read_only: raise BothReadOnly() a_info = StorageInfo(storage_a, dict( (ident, (href_a, etag_a)) for ident, (href_a, etag_a, href_b, etag_b) in iteritems(status) )) b_info = StorageInfo(storage_b, dict( (ident, (href_b, etag_b)) for ident, (href_a, etag_a, href_b, etag_b) in iteritems(status) )) a_info.prepare_idents(storage_b.read_only) b_info.prepare_idents(storage_a.read_only) if bool(a_info.idents) != bool(b_info.idents) \ and status and not force_delete: raise StorageEmpty( empty_storage=(storage_b if a_info.idents else storage_a)) actions = list(_get_actions(a_info, b_info)) with storage_a.at_once(): with storage_b.at_once(): for action in actions: action(a_info, b_info, conflict_resolution) status.clear() for ident in uniq(itertools.chain(a_info.status, b_info.status)): href_a, etag_a = a_info.status[ident] href_b, etag_b = b_info.status[ident] status[ident] = href_a, etag_a, href_b, etag_b def _action_upload(ident, source, dest): def inner(a, b, conflict_resolution): sync_logger.info('Copying (uploading) item {0} to {1}' .format(ident, dest.storage)) source_meta = source.idents[ident] if dest.storage.read_only: sync_logger.warning('{dest} is read-only. Skipping update...' .format(dest=dest.storage)) dest_href = dest_etag = None else: item = source_meta['item'] dest_href, dest_etag = dest.storage.upload(item) source.status[ident] = source_meta['href'], source_meta['etag'] dest.status[ident] = dest_href, dest_etag return inner def _action_update(ident, source, dest): def inner(a, b, conflict_resolution): sync_logger.info('Copying (updating) item {0} to {1}' .format(ident, dest.storage)) source_meta = source.idents[ident] if dest.storage.read_only: sync_logger.info('{dest} is read-only. Skipping update...' .format(dest=dest.storage)) dest_href = dest_etag = None else: dest_meta = dest.idents[ident] dest_href = dest_meta['href'] dest_etag = dest.storage.update(dest_href, source_meta['item'], dest_meta['etag']) assert isinstance(dest_etag, (bytes, text_type)) source.status[ident] = source_meta['href'], source_meta['etag'] dest.status[ident] = dest_href, dest_etag return inner def _action_delete(ident, info): storage = info.storage idents = info.idents def inner(a, b, conflict_resolution): sync_logger.info('Deleting item {0} from {1}'.format(ident, storage)) if storage.read_only: sync_logger.warning('{0} is read-only, skipping deletion...' .format(storage)) else: meta = idents[ident] etag = meta['etag'] href = meta['href'] storage.delete(href, etag) del a.status[ident] del b.status[ident] return inner def _action_delete_status(ident): def inner(a, b, conflict_resolution): sync_logger.info('Deleting status info for nonexisting item {0}' .format(ident)) del a.status[ident] del b.status[ident] return inner def _action_conflict_resolve(ident): def inner(a, b, conflict_resolution): sync_logger.info('Doing conflict resolution for item {0}...' .format(ident)) meta_a = a.idents[ident] meta_b = b.idents[ident] if meta_a['item'].raw == meta_b['item'].raw: sync_logger.info('...same content on both sides.') a.status[ident] = meta_a['href'], meta_a['etag'] b.status[ident] = meta_b['href'], meta_b['etag'] elif conflict_resolution is None: raise SyncConflict(ident=ident, href_a=meta_a['href'], href_b=meta_b['href']) elif conflict_resolution == 'a wins': sync_logger.info('...{0} wins.'.format(a.storage)) _action_update(ident, a, b)(a, b, conflict_resolution) elif conflict_resolution == 'b wins': sync_logger.info('...{0} wins.'.format(b.storage)) _action_update(ident, b, a)(a, b, conflict_resolution) else: raise ValueError('Invalid conflict resolution mode: {0}' .format(conflict_resolution)) return inner def _get_actions(a_info, b_info): for ident in uniq(itertools.chain(a_info.idents, b_info.idents, a_info.status)): a = a_info.idents.get(ident, None) b = b_info.idents.get(ident, None) assert not a or a['etag'] is not None assert not b or b['etag'] is not None _, status_etag_a = a_info.status.get(ident, (None, None)) _, status_etag_b = b_info.status.get(ident, (None, None)) if a and b: if a['etag'] != status_etag_a and b['etag'] != status_etag_b: # item was modified on both sides # OR: missing status yield _action_conflict_resolve(ident) elif a['etag'] != status_etag_a: # item was only modified in a yield _action_update(ident, a_info, b_info) elif b['etag'] != status_etag_b: # item was only modified in b yield _action_update(ident, b_info, a_info) elif a and not b: if a['etag'] != status_etag_a: # was deleted from b but modified on a # OR: new item was created in a yield _action_upload(ident, a_info, b_info) else: # was deleted from b and not modified on a yield _action_delete(ident, a_info) elif not a and b: if b['etag'] != status_etag_b: # was deleted from a but modified on b # OR: new item was created in b yield _action_upload(ident, b_info, a_info) else: # was deleted from a and not changed on b yield _action_delete(ident, b_info) elif not a and not b: # was deleted from a and b, clean up status yield _action_delete_status(ident) ```
[ { "content": "```python\nimport math, sys\nfrom SPARQLWrapper import SPARQLWrapper, JSON\nfrom queries.keywordSectionCount import keywordSectionCount\nfrom queries.createKeywordSectionQuery import createKeywordSectionQuery\n\n# get the year to query from the user\nyear = input(\"Enter year to query: \")\nyear =...
[ { "content": "<|memory_start|>```python\nimport math, sys\nfrom SPARQLWrapper import SPARQLWrapper, JSON\nfrom queries.keywordSectionCount import keywordSectionCount\nfrom queries.createKeywordSectionQuery import createKeywordSectionQuery\n\n# get the year to query from the user\nyear = input(\"Enter year to qu...
```python import math, sys from SPARQLWrapper import SPARQLWrapper, JSON from queries.keywordSectionCount import keywordSectionCount from queries.createKeywordSectionQuery import createKeywordSectionQuery # get the year to query from the user year = input("Enter year to query: ") year = str(year) # there are too many results to get all at once # here we ask the database how many results there # are for the year we are interested in. Given that # we can get 10,000 results per query, we do a little # math to compute how many times we need to query the database # to get all the results offset = 0 limit = float(keywordSectionCount(year)) numQueries = math.ceil(limit/10000) # setting up the query # specifying the web address of the database # setting the return format to JSON - JavaScript Object Notation sparql = SPARQLWrapper("http://abstractsearch.agu.org:8890/sparql") sparql.setReturnFormat(JSON) # keep looping and querying until we get all the results while (numQueries > 0): query = createKeywordSectionQuery(year,str(offset)) sparql.setQuery(query) offset = offset + 10000 results = sparql.query().convert() for result in results["results"]["bindings"]: print(result["keyword"]["value"] + " " + result["section"]["value"]) numQueries = numQueries - 1 ```
[ { "content": "Output the full code verbatim (no extra comments):\n```python\n# -*- coding: utf-8 -*-\n# Generated by Django 1.11.6 on 2017-10-25 01:44\nfrom __future__ import unicode_literals\n\nfrom decimal import Decimal\nfrom django.conf import settings\nfrom django.db import migrations, models\nimport djang...
[ { "content": "Output the full code verbatim (no extra comments):\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n# Generated by Django 1.11.6 on 2017-10-25 01:44\nfrom __future__ import unicode_literals\n\nfrom decimal import Decimal\nfrom django.conf import settings\nfrom django.db import migrations, mode...
```python # -*- coding: utf-8 -*- # Generated by Django 1.11.6 on 2017-10-25 01:44 from __future__ import unicode_literals from decimal import Decimal from django.conf import settings from django.db import migrations, models import django.db.models.deletion import djmoney.models.fields class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ('policy', '0006_auto_20171025_0144'), ] operations = [ migrations.CreateModel( name='Application', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('stage', models.CharField(choices=[('E', 'Early'), ('R', 'Regular'), ('L', 'Late')], max_length=1)), ('screening_result', models.CharField(choices=[('A', 'Accepted'), ('R', 'Rejected'), ('P', 'Pending')], default='P', max_length=1)), ('disclose_result', models.BooleanField(default=False)), ('essay_text', models.TextField(blank=True)), ('visa_letter', models.BooleanField(default=False)), ('financial_aid', models.BooleanField(default=False)), ('previous_participation', models.BooleanField(default=False)), ('last_update', models.DateTimeField(auto_now=True)), ('essay_topic', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='applications', to='policy.EssayTopic')), ], ), migrations.CreateModel( name='Group', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=50, unique=True)), ], options={ 'verbose_name': 'applicant group', }, ), migrations.CreateModel( name='Order', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('preferred_currency', models.CharField(choices=[('KRW', 'KRW'), ('USD', 'USD')], default='KRW', max_length=3)), ('paid_amount_currency', djmoney.models.fields.CurrencyField(choices=[('KRW', 'KRW'), ('USD', 'USD')], default='KRW', editable=False, max_length=3)), ('paid_amount', djmoney.models.fields.MoneyField(decimal_places=0, default=Decimal('0'), default_currency='KRW', max_digits=7)), ('dietary_preferences', models.CharField(blank=True, max_length=100, null=True)), ('breakfast_option', models.BooleanField(default=False)), ('pre_option', models.BooleanField(default=False, verbose_name='pre-conference banquet')), ('post_option', models.BooleanField(default=False, verbose_name='post-conference tour')), ('accommodation', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='orders', to='policy.AccommodationOption')), ('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), migrations.AddField( model_name='application', name='group', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='applications', to='registration.Group'), ), migrations.AddField( model_name='application', name='topic_preference', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='applications', to='policy.ProjectTopic'), ), migrations.AddField( model_name='application', name='user', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='applications', to=settings.AUTH_USER_MODEL), ), ] ```
[ { "content": "Replicate the code snippet exactly, without paraphrasing or reformatting:\n```python\n# -*- coding: utf-8 -*-\n'''\nLocal settings\n\n- Run in Debug mode\n- Use console backend for emails\n- Add Django Debug Toolbar\n- Add django-extensions as app\n'''\n\nfrom .common import * # noqa\n\n# DEBUG\n...
[ { "content": "Replicate the code snippet exactly, without paraphrasing or reformatting:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n'''\nLocal settings\n\n- Run in Debug mode\n- Use console backend for emails\n- Add Django Debug Toolbar\n- Add django-extensions as app\n'''\n\nfrom .common import * # n...
```python # -*- coding: utf-8 -*- ''' Local settings - Run in Debug mode - Use console backend for emails - Add Django Debug Toolbar - Add django-extensions as app ''' from .common import * # noqa # DEBUG # ------------------------------------------------------------------------------ DEBUG = env.bool('DJANGO_DEBUG', default=True) TEMPLATES[0]['OPTIONS']['debug'] = DEBUG # SECRET CONFIGURATION # ------------------------------------------------------------------------------ # See: https://docs.djangoproject.com/en/dev/ref/settings/#secret-key # Note: This key only used for development and testing. SECRET_KEY = env("DJANGO_SECRET_KEY", default='CHANGEME!!!') # Mail settings # ------------------------------------------------------------------------------ EMAIL_HOST = 'localhost' EMAIL_PORT = 1025 EMAIL_BACKEND = env('DJANGO_EMAIL_BACKEND', default='django.core.mail.backends.console.EmailBackend') # CACHING # ------------------------------------------------------------------------------ CACHES = { 'default': { 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache', 'LOCATION': '' } } # django-debug-toolbar # ------------------------------------------------------------------------------ MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',) INSTALLED_APPS += ('debug_toolbar', ) INTERNAL_IPS = ('127.0.0.1', '10.0.2.2',) DEBUG_TOOLBAR_CONFIG = { 'DISABLE_PANELS': [ 'debug_toolbar.panels.redirects.RedirectsPanel', ], 'SHOW_TEMPLATE_CONTEXT': True, } # django-extensions # ------------------------------------------------------------------------------ INSTALLED_APPS += ('django_extensions', ) # TESTING # ------------------------------------------------------------------------------ TEST_RUNNER = 'django.test.runner.DiscoverRunner' {%- if cookiecutter.use_celery == "y" -%} ########## CELERY # In development, all tasks will be executed locally by blocking until the task returns CELERY_ALWAYS_EAGER = True ########## END CELERY {%- endif %} # Your local stuff: Below this line define 3rd party library settings ```
[ { "content": "Repeat the following code:\n```python\n############################################################################\n# NHL GAMECENTER\n# XBMC ADD-ON\n############################################################################\nimport xbmcgui\nfrom resources.lib.userinterface import *\nfrom resour...
[ { "content": "Repeat the following code:\n<|memory_start|>```python\n############################################################################\n# NHL GAMECENTER\n# XBMC ADD-ON\n############################################################################\nimport xbmcgui\nfrom resources.lib.userinterface impor...
```python ############################################################################ # NHL GAMECENTER # XBMC ADD-ON ############################################################################ import xbmcgui from resources.lib.userinterface import * from resources.lib.scoring_updates import * params = get_params() print params try: url=urllib.unquote_plus(params["url"]) except: url=None try: mode = int(params['mode']) except: mode = None print "Mode: "+str(mode) print "URL: "+str(url) if mode == None or url==None or len(url)<1: CATEGORIES() xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 1: LIVE(url) #xbmcplugin.endOfDirectory(int(sys.argv[1])) #Do Not Cache Menu xbmcplugin.endOfDirectory(int(sys.argv[1]),1,0,0) elif mode == 2: LIVEQUALITY(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 3: LIVELINKS(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 4: ARCHIVE(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 5: ARCHIVEMONTH(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 6: ARCHIVEGAMES(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 7: ARCHIVEQUALITY(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 8: ARCHIVELINKS(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 9: LASTNIGHT(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 10: LASTNIGHTTYPE(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 11: LATESTGAMES(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 12: LATESTGQUALITY(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 13: LATESTGLINKS(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 14: LATESTGTYPE(url) xbmcplugin.endOfDirectory(int(sys.argv[1])) elif mode == 100: ADDON.setSetting(id='score_updates', value='true') s = ScoreThread() t = threading.Thread(target = s.Scoring_Updates) t.start() elif mode == 101: dialog = xbmcgui.Dialog() title = "Score Notifications" dialog.notification(title, 'Stopping...', ADDON_PATH+'/resources/images/nhl_logo.png', 5000, False) ADDON.setSetting(id='score_updates', value='false') ```
[ { "content": "Repeat the following code:\n```python\nfrom ems.qt import QtWidgets, QtCore, QtGui\nfrom ems.qt.richtext.char_format_proxy import CharFormatProxy\n\nQt = QtCore.Qt\nQObject = QtCore.QObject\nQColor = QtGui.QColor\nQAction = QtWidgets.QAction\nQKeySequence = QtGui.QKeySequence\nQFont = QtGui.QFont\...
[ { "content": "Repeat the following code:\n<|memory_start|>```python\nfrom ems.qt import QtWidgets, QtCore, QtGui\nfrom ems.qt.richtext.char_format_proxy import CharFormatProxy\n\nQt = QtCore.Qt\nQObject = QtCore.QObject\nQColor = QtGui.QColor\nQAction = QtWidgets.QAction\nQKeySequence = QtGui.QKeySequence\nQFon...
```python from ems.qt import QtWidgets, QtCore, QtGui from ems.qt.richtext.char_format_proxy import CharFormatProxy Qt = QtCore.Qt QObject = QtCore.QObject QColor = QtGui.QColor QAction = QtWidgets.QAction QKeySequence = QtGui.QKeySequence QFont = QtGui.QFont QIcon = QtGui.QIcon QPixmap = QtGui.QPixmap ThemeIcon = QIcon.fromTheme QApplication = QtWidgets.QApplication QColorDialog = QtWidgets.QColorDialog QFontComboBox = QtWidgets.QFontComboBox QComboBox = QtWidgets.QComboBox QFontDatabase = QtGui.QFontDatabase QTextDocument = QtGui.QTextDocument QTextCharFormat = QtGui.QTextCharFormat pyqtSignal = QtCore.pyqtSignal pyqtSlot = QtCore.pyqtSlot pyqtProperty = QtCore.pyqtProperty class CharFormatActions(QObject): documentChanged = pyqtSignal(QTextDocument) currentBlockFormatChanged = pyqtSignal(QTextCharFormat) def __init__(self, parentWidget, signalProxy=None, resourcePath=':/text-editor'): super(CharFormatActions, self).__init__(parentWidget) self.resourcePath = resourcePath self.actions = [] self.widgets = [] self.signals = CharFormatProxy(self) if signalProxy is None else signalProxy self._addActions(self.parent()) self._document = QTextDocument() self._lastBlockFormat = None def getDocument(self): return self._document @pyqtSlot(QTextDocument) def setDocument(self, document): if self._document is document: return if self._document: self._disconnectFromDocument(self._document) self._document = document self.documentChanged.emit(self._document) document = pyqtProperty(QTextDocument, getDocument, setDocument) def _disconnectFromDocument(self, document): return def _addActions(self, parent): self.actionTextBold = QAction( ThemeIcon('format-text-bold', self._icon('bold.png')), "&Bold", parent, priority=QAction.LowPriority, shortcut=Qt.CTRL + Qt.Key_B, triggered=self.signals.setBold, checkable=True) bold = QFont() bold.setBold(True) self.actionTextBold.setFont(bold) self.signals.boldChanged.connect(self.actionTextBold.setChecked) self.actions.append(self.actionTextBold) self.actionTextItalic = QAction( ThemeIcon('format-text-italic', self._icon('italic.png')), "&Italic", self, priority=QAction.LowPriority, shortcut=Qt.CTRL + Qt.Key_I, triggered=self.signals.setItalic, checkable=True) italic = QFont() italic.setItalic(True) self.actionTextItalic.setFont(italic) self.signals.italicChanged.connect(self.actionTextItalic.setChecked) self.actions.append(self.actionTextItalic) self.actionTextUnderline = QAction( ThemeIcon('format-text-underline', self._icon('underline.png')), "&Underline", self, priority=QAction.LowPriority, shortcut=Qt.CTRL + Qt.Key_U, triggered=self.signals.setUnderline, checkable=True) underline = QFont() underline.setUnderline(True) self.actionTextUnderline.setFont(underline) self.actions.append(self.actionTextUnderline) self.signals.underlineChanged.connect(self.actionTextUnderline.setChecked) pix = QPixmap(16, 16) pix.fill(Qt.black) self.actionTextColor = QAction(QIcon(pix), "&Color...", self, triggered=self._textColor) self.signals.foregroundColorChanged.connect(self._colorChanged) self.actions.append(self.actionTextColor) self.actionClearFormat = QAction(ThemeIcon('format-text-clear', self._icon('magic.png')), "&Remove Format", self, priority=QAction.LowPriority, shortcut=Qt.CTRL + Qt.Key_E, triggered=self.signals.clearFormat) self.actions.append(self.actionClearFormat) self.fontCombo = QFontComboBox() self.fontCombo.activated[str].connect(self.signals.setFontFamily) self.signals.fontFamilyChanged.connect(self.setFontFamily) self.widgets.append(self.fontCombo) self.sizeCombo = QComboBox() self.sizeCombo.setObjectName("sizeCombo") self.sizeCombo.setEditable(True) self.signals.pointSizeChanged.connect(self.setFontPointSize) self.widgets.append(self.sizeCombo) db = QFontDatabase() for size in db.standardSizes(): self.sizeCombo.addItem("{}".format(size)) self.sizeCombo.activated[str].connect(self._textSize) self.sizeCombo.setCurrentIndex( self.sizeCombo.findText( "{}".format(QApplication.font().pointSize()) ) ) def _textColor(self): color = self.signals.getForegroundColor() if not color: color = QColor(0,0,0) col = QColorDialog.getColor(color, self.parent()) if not col.isValid(): return self.signals.setForegroundColor(col) def _colorChanged(self, color): pix = QPixmap(16, 16) pix.fill(color) self.actionTextColor.setIcon(QIcon(pix)) def _textSize(self, pointSize): pointSize = float(pointSize) if pointSize < 0: return self.signals.setPointSize(pointSize) def addToToolbar(self, toolbar, addActions=True, addWidgets=True): if addActions: for action in self.actions: toolbar.addAction(action) if not addWidgets: return for widget in self.widgets: widget.setParent(toolbar) toolbar.addWidget(widget) def setFontFamily(self, family): self.fontCombo.setCurrentIndex(self.fontCombo.findText(family)) def setFontPointSize(self, pointSize): self.sizeCombo.setCurrentIndex(self.sizeCombo.findText("{}".format(int(pointSize)))) def iconPath(self, fileName): return self.resourcePath + '/' + fileName def _icon(self, fileName): return QIcon(self.iconPath(fileName)) ```
[ { "content": "```python\n# Secure Trading Python API\n# Authors: Secure Trading Ltd\n# Configuration variables\n\nfrom __future__ import unicode_literals\nfrom .requestobject import Request\nfrom .requestobject import Requests\nfrom .responseobject import Response\nfrom .exceptions import SecureTradingError\nfr...
[ { "content": "<|memory_start|>```python\n# Secure Trading Python API\n# Authors: Secure Trading Ltd\n# Configuration variables\n\nfrom __future__ import unicode_literals\nfrom .requestobject import Request\nfrom .requestobject import Requests\nfrom .responseobject import Response\nfrom .exceptions import Secure...
```python # Secure Trading Python API # Authors: Secure Trading Ltd # Configuration variables from __future__ import unicode_literals from .requestobject import Request from .requestobject import Requests from .responseobject import Response from .exceptions import SecureTradingError from .exceptions import ApiError from .exceptions import HttpError from .exceptions import ConnectionError from .exceptions import SendReceiveError from .converter import Converter from .config import Config from .api import Api from .phrasebook import PhraseBook import securetrading.util import pkgutil import platform dataFile = 'data/errormessages.json' data = pkgutil.get_data('securetrading', dataFile).decode("utf-8") error_messages = securetrading.util.json.loads(data) dataFile = 'data/phrasebook.json' data = pkgutil.get_data('securetrading', dataFile).decode("utf-8") phrase_book = securetrading.util.json.loads(data) __title__ = 'Secure Trading Python API' __version__ = "1.0.16" __author__ = 'Secure Trading Ltd' __license__ = 'MIT' __copyright__ = 'Copyright 2016 Secure Trading Ltd' version_information = ["Python", platform.python_version(), securetrading.__version__, platform.platform(), ] version_info = "::".join(version_information) ```
[ { "content": "Produce an exact reconstruction of the code:\n```python\nimport unittest\nfrom nqueens.chessboard import Chessboard\nfrom nqueens.solver import Solver\n\n\nclass TestSolver(unittest.TestCase):\n\n def setUp(self):\n self.standardBoard = Chessboard.create(8)\n\n def testSimpleSolution(...
[ { "content": "Produce an exact reconstruction of the code:\n<|memory_start|>```python\nimport unittest\nfrom nqueens.chessboard import Chessboard\nfrom nqueens.solver import Solver\n\n\nclass TestSolver(unittest.TestCase):\n\n def setUp(self):\n self.standardBoard = Chessboard.create(8)\n\n def tes...
```python import unittest from nqueens.chessboard import Chessboard from nqueens.solver import Solver class TestSolver(unittest.TestCase): def setUp(self): self.standardBoard = Chessboard.create(8) def testSimpleSolution(self): board = Chessboard.create(1) solver = Solver.create(board) solution = solver.solve() self.assertIsNotNone(solution) self.assertEqual(solution.getSize(), 1) self.assertTrue(solution.hasQueen(0, 0)) self.assertTrue(solution.isValid()) def testImpossible(self): board = Chessboard.create(2) solver = Solver.create(board) solution = solver.solve() self.assertIsNone(solution) def runTestFor(self, n): board = Chessboard.create(n) solver = Solver.create(board) solution = solver.solve() self.assertIsNotNone(solution) self.assertEqual(solution.getQueenCount(), n) self.assertTrue(solution.isValid()) # Useful to have a quick-running but still reasonably-sized test def testSmall(self): self.runTestFor(6) def testStandard(self): self.runTestFor(8) ```
[ { "content": "Replicate the source code:\n```python\n# -*- coding: utf-8 -*-\n# Copyright (c) 2015, Mayo Clinic\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without modification,\n# are permitted provided that the following conditions are met:\n#\n# Redistributions of...
[ { "content": "Replicate the source code:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n# Copyright (c) 2015, Mayo Clinic\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without modification,\n# are permitted provided that the following conditions are met:\n#\n# Re...
```python # -*- coding: utf-8 -*- # Copyright (c) 2015, Mayo Clinic # All rights reserved. # # Redistribution and use in source and binary forms, with or without modification, # are permitted provided that the following conditions are met: # # Redistributions of source code must retain the above copyright notice, this # list of conditions and the following disclaimer. # # Redistributions in binary form must reproduce the above copyright notice, # this list of conditions and the following disclaimer in the documentation # and/or other materials provided with the distribution. # # Neither the name of the <ORGANIZATION> nor the names of its contributors # may be used to endorse or promote products derived from this software # without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. # IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, # INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, # BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE # OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED # OF THE POSSIBILITY OF SUCH DAMAGE. import os import unittest import shutil from dbgap.file_downloader import FileDownloader single_file_template = 'dbgap/studies/%(study)s/%(fullname)s/GapExchange_%(fullname)s.xml' directory_template = '/dbgap/studies/%(study)s/%(fullname)s/pheno_variable_summaries' class FileDownloaderTestCase(unittest.TestCase): def test_dowload_single_file(self): study = 'phs001007' fullname = study + ".v1.p1" dld = FileDownloader('ftp.ncbi.nlm.nih.gov') self.assertEqual(open(os.path.join('data', 'phs001007.xml')).read(), dld.download_file(single_file_template % dict(study=study, fullname=fullname))) def test_dir_download(self): test_dir = os.path.join('data', 'dltest') shutil.rmtree(test_dir, ignore_errors=True) os.makedirs(test_dir) study = 'phs000722' fullname = study + ".v1.p1" dld = FileDownloader('ftp.ncbi.nlm.nih.gov') self.assertEqual(4, dld.download_dir(directory_template % dict(study=study, fullname=fullname), test_dir, name_map=lambda s: s.replace('.xml', '.tst'), file_filtr=lambda s: 'data_dict' in s)) if __name__ == '__main__': unittest.main() ```
[ { "content": "Here is the code content:\n```python\n##############################################################################\n#\n# OSIS stands for Open Student Information System. It's an application\n# designed to manage the core business of higher education institutions,\n# such as universities...
[ { "content": "Here is the code content:\n<|memory_start|>```python\n##############################################################################\n#\n# OSIS stands for Open Student Information System. It's an application\n# designed to manage the core business of higher education institutions,\n# such...
```python ############################################################################## # # OSIS stands for Open Student Information System. It's an application # designed to manage the core business of higher education institutions, # such as universities, faculties, institutes and professional schools. # The core business involves the administration of students, teachers, # courses, programs and so on. # # Copyright (C) 2015-2018 Université catholique de Louvain (http://www.uclouvain.be) # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # A copy of this license - GNU General Public License - is available # at the root of the source code of this program. If not, # see http://www.gnu.org/licenses/. # ############################################################################## from django.db import models from osis_common.models.serializable_model import SerializableModel, SerializableModelAdmin class OfferAdmin(SerializableModelAdmin): list_display = ('id', 'title', 'changed') search_fields = ['title'] class Offer(SerializableModel): external_id = models.CharField(max_length=100, blank=True, null=True, db_index=True) changed = models.DateTimeField(null=True, auto_now=True) title = models.CharField(max_length=255) def __str__(self): return "{} {}".format(self.id, self.title) class Meta: permissions = ( ("can_access_offer", "Can access offer"), ("can_access_catalog", "Can access catalog"), ) def find_by_id(offer_id): try: return Offer.objects.get(pk=offer_id) except Offer.DoesNotExist: return None ```
[ { "content": "Here is a code file:\n```python\nfrom os import path\n\nfrom enigma import iPlayableService, iServiceInformation, eTimer, eServiceCenter, eServiceReference, eDVBDB\n\nfrom Screens.Screen import Screen\nfrom Screens.ChannelSelection import FLAG_IS_DEDICATED_3D\nfrom Components.About import about\nf...
[ { "content": "Here is a code file:\n<|memory_start|>```python\nfrom os import path\n\nfrom enigma import iPlayableService, iServiceInformation, eTimer, eServiceCenter, eServiceReference, eDVBDB\n\nfrom Screens.Screen import Screen\nfrom Screens.ChannelSelection import FLAG_IS_DEDICATED_3D\nfrom Components.About...
```python from os import path from enigma import iPlayableService, iServiceInformation, eTimer, eServiceCenter, eServiceReference, eDVBDB from Screens.Screen import Screen from Screens.ChannelSelection import FLAG_IS_DEDICATED_3D from Components.About import about from Components.SystemInfo import SystemInfo from Components.ConfigList import ConfigListScreen from Components.config import config, configfile, getConfigListEntry from Components.Label import Label from Components.Sources.StaticText import StaticText from Components.Pixmap import Pixmap from Components.Sources.Boolean import Boolean from Components.ServiceEventTracker import ServiceEventTracker from Tools.Directories import resolveFilename, SCOPE_PLUGINS from Tools.HardwareInfo import HardwareInfo from Components.AVSwitch import iAVSwitch resolutionlabel = None class VideoSetup(Screen, ConfigListScreen): def __init__(self, session): Screen.__init__(self, session) # for the skin: first try VideoSetup, then Setup, this allows individual skinning self.skinName = ["VideoSetup", "Setup" ] self.setup_title = _("Video settings") self["HelpWindow"] = Pixmap() self["HelpWindow"].hide() self["VKeyIcon"] = Boolean(False) self['footnote'] = Label() self.hw = iAVSwitch self.onChangedEntry = [ ] # handle hotplug by re-creating setup self.onShow.append(self.startHotplug) self.onHide.append(self.stopHotplug) self.list = [ ] ConfigListScreen.__init__(self, self.list, session = session, on_change = self.changedEntry) from Components.ActionMap import ActionMap self["actions"] = ActionMap(["SetupActions", "MenuActions", "ColorActions"], { "cancel": self.keyCancel, "save": self.apply, "menu": self.closeRecursive, }, -2) self["key_red"] = StaticText(_("Cancel")) self["key_green"] = StaticText(_("OK")) self["description"] = Label("") self.createSetup() self.grabLastGoodMode() self.onLayoutFinish.append(self.layoutFinished) def layoutFinished(self): self.setTitle(self.setup_title) def startHotplug(self): self.hw.on_hotplug.append(self.createSetup) def stopHotplug(self): self.hw.on_hotplug.remove(self.createSetup) def createSetup(self): level = config.usage.setup_level.index self.list = [ getConfigListEntry(_("Video output"), config.av.videoport, _("Configures which video output connector will be used.")) ] if config.av.videoport.value in ('HDMI', 'YPbPr', 'Scart-YPbPr') and not path.exists(resolveFilename(SCOPE_PLUGINS)+'SystemPlugins/AutoResolution'): self.list.append(getConfigListEntry(_("Automatic resolution"), config.av.autores,_("If enabled the output resolution of the box will try to match the resolution of the video contents resolution"))) if config.av.autores.value in ('all', 'hd'): self.list.append(getConfigListEntry(_("Delay time"), config.av.autores_delay,_("Set the time before checking video source for resolution infomation."))) self.list.append(getConfigListEntry(_("Force de-interlace"), config.av.autores_deinterlace,_("If enabled the video will always be de-interlaced."))) self.list.append(getConfigListEntry(_("Automatic resolution label"), config.av.autores_label_timeout,_("Allows you to adjust the amount of time the resolution infomation display on screen."))) if config.av.autores.value in 'hd': self.list.append(getConfigListEntry(_("Show SD as"), config.av.autores_sd,_("This option allows you to choose how to display standard defintion video on your TV."))) self.list.append(getConfigListEntry(_("Show 480/576p 24fps as"), config.av.autores_480p24,_("This option allows you to choose how to display SD progressive 24Hz on your TV. (as not all TV's support these resolutions)"))) self.list.append(getConfigListEntry(_("Show 720p 24fps as"), config.av.autores_720p24,_("This option allows you to choose how to display 720p 24Hz on your TV. (as not all TV's support these resolutions)"))) self.list.append(getConfigListEntry(_("Show 1080p 24fps as"), config.av.autores_1080p24,_("This option allows you to choose how to display 1080p 24Hz on your TV. (as not all TV's support these resolutions)"))) self.list.append(getConfigListEntry(_("Show 1080p 25fps as"), config.av.autores_1080p25,_("This option allows you to choose how to display 1080p 25Hz on your TV. (as not all TV's support these resolutions)"))) self.list.append(getConfigListEntry(_("Show 1080p 30fps as"), config.av.autores_1080p30,_("This option allows you to choose how to display 1080p 30Hz on your TV. (as not all TV's support these resolutions)"))) self.list.append(getConfigListEntry(_('Always use smart1080p mode'), config.av.smart1080p, _("This option allows you to always use e.g. 1080p50 for TV/.ts, and 1080p24/p50/p60 for videos"))) # if we have modes for this port: if (config.av.videoport.value in config.av.videomode and config.av.autores.value == 'disabled') or config.av.videoport.value == 'Scart': # add mode- and rate-selection: self.list.append(getConfigListEntry(pgettext("Video output mode", "Mode"), config.av.videomode[config.av.videoport.value], _("This option configures the video output mode (or resolution)."))) if config.av.videomode[config.av.videoport.value].value == 'PC': self.list.append(getConfigListEntry(_("Resolution"), config.av.videorate[config.av.videomode[config.av.videoport.value].value], _("This option configures the screen resolution in PC output mode."))) elif config.av.videoport.value != 'Scart': self.list.append(getConfigListEntry(_("Refresh rate"), config.av.videorate[config.av.videomode[config.av.videoport.value].value], _("Configure the refresh rate of the screen."))) port = config.av.videoport.value if port not in config.av.videomode: mode = None else: mode = config.av.videomode[port].value # some modes (720p, 1080i) are always widescreen. Don't let the user select something here, "auto" is not what he wants. force_wide = self.hw.isWidescreenMode(port, mode) if not force_wide: self.list.append(getConfigListEntry(_("Aspect ratio"), config.av.aspect, _("Configure the aspect ratio of the screen."))) if force_wide or config.av.aspect.value in ("16:9", "16:10"): self.list.extend(( getConfigListEntry(_("Display 4:3 content as"), config.av.policy_43, _("When the content has an aspect ratio of 4:3, choose whether to scale/stretch the picture.")), getConfigListEntry(_("Display >16:9 content as"), config.av.policy_169, _("When the content has an aspect ratio of 16:9, choose whether to scale/stretch the picture.")) )) elif config.av.aspect.value == "4:3": self.list.append(getConfigListEntry(_("Display 16:9 content as"), config.av.policy_169, _("When the content has an aspect ratio of 16:9, choose whether to scale/stretch the picture."))) # if config.av.videoport.value == "HDMI": # self.list.append(getConfigListEntry(_("Allow unsupported modes"), config.av.edid_override)) if config.av.videoport.value == "Scart": self.list.append(getConfigListEntry(_("Color format"), config.av.colorformat, _("Configure which color format should be used on the SCART output."))) if level >= 1: self.list.append(getConfigListEntry(_("WSS on 4:3"), config.av.wss, _("When enabled, content with an aspect ratio of 4:3 will be stretched to fit the screen."))) if SystemInfo["ScartSwitch"]: self.list.append(getConfigListEntry(_("Auto scart switching"), config.av.vcrswitch, _("When enabled, your receiver will detect activity on the VCR SCART input."))) # if not isinstance(config.av.scaler_sharpness, ConfigNothing): # self.list.append(getConfigListEntry(_("Scaler sharpness"), config.av.scaler_sharpness, _("This option configures the picture sharpness."))) if SystemInfo["havecolorspace"]: self.list.append(getConfigListEntry(_("HDMI Colorspace"), config.av.hdmicolorspace,_("This option allows you can config the Colorspace from Auto to RGB"))) self["config"].list = self.list self["config"].l.setList(self.list) if config.usage.sort_settings.value: self["config"].list.sort() def keyLeft(self): ConfigListScreen.keyLeft(self) self.createSetup() def keyRight(self): ConfigListScreen.keyRight(self) self.createSetup() def confirm(self, confirmed): if not confirmed: config.av.videoport.setValue(self.last_good[0]) config.av.videomode[self.last_good[0]].setValue(self.last_good[1]) config.av.videorate[self.last_good[1]].setValue(self.last_good[2]) config.av.autores_sd.setValue(self.last_good_extra[0]) config.av.smart1080p.setValue(self.last_good_extra[1]) self.hw.setMode(*self.last_good) else: self.keySave() def grabLastGoodMode(self): port = config.av.videoport.value mode = config.av.videomode[port].value rate = config.av.videorate[mode].value self.last_good = (port, mode, rate) autores_sd = config.av.autores_sd.value smart1080p = config.av.smart1080p.value self.last_good_extra = (autores_sd, smart1080p) def saveAll(self): if config.av.videoport.value == 'Scart': config.av.autores.setValue('disabled') for x in self["config"].list: x[1].save() configfile.save() def apply(self): port = config.av.videoport.value mode = config.av.videomode[port].value rate = config.av.videorate[mode].value autores_sd = config.av.autores_sd.value smart1080p = config.av.smart1080p.value if ((port, mode, rate) != self.last_good) or (autores_sd, smart1080p) != self.last_good_extra: if autores_sd.find('1080') >= 0: self.hw.setMode(port, '1080p', '50Hz') elif (smart1080p == '1080p50') or (smart1080p == 'true'): # for compatibility with old ConfigEnableDisable self.hw.setMode(port, '1080p', '50Hz') elif smart1080p == '2160p50': self.hw.setMode(port, '2160p', '50Hz') elif smart1080p == '1080i50': self.hw.setMode(port, '1080i', '50Hz') elif smart1080p == '720p50': self.hw.setMode(port, '720p', '50Hz') else: self.hw.setMode(port, mode, rate) from Screens.MessageBox import MessageBox self.session.openWithCallback(self.confirm, MessageBox, _("Is this video mode ok?"), MessageBox.TYPE_YESNO, timeout = 20, default = False) else: self.keySave() # for summary: def changedEntry(self): for x in self.onChangedEntry: x() def getCurrentEntry(self): return self["config"].getCurrent()[0] def getCurrentValue(self): return str(self["config"].getCurrent()[1].getText()) def getCurrentDescription(self): return self["config"].getCurrent() and len(self["config"].getCurrent()) > 2 and self["config"].getCurrent()[2] or "" def createSummary(self): from Screens.Setup import SetupSummary return SetupSummary class AudioSetup(Screen, ConfigListScreen): def __init__(self, session): Screen.__init__(self, session) # for the skin: first try AudioSetup, then Setup, this allows individual skinning self.skinName = ["AudioSetup", "Setup" ] self.setup_title = _("Audio settings") self["HelpWindow"] = Pixmap() self["HelpWindow"].hide() self["VKeyIcon"] = Boolean(False) self['footnote'] = Label() self.hw = iAVSwitch self.onChangedEntry = [ ] # handle hotplug by re-creating setup self.onShow.append(self.startHotplug) self.onHide.append(self.stopHotplug) self.list = [ ] ConfigListScreen.__init__(self, self.list, session = session, on_change = self.changedEntry) from Components.ActionMap import ActionMap self["actions"] = ActionMap(["SetupActions", "MenuActions", "ColorActions"], { "cancel": self.keyCancel, "save": self.apply, "menu": self.closeRecursive, }, -2) self["key_red"] = StaticText(_("Cancel")) self["key_green"] = StaticText(_("OK")) self["description"] = Label("") self.createSetup() self.onLayoutFinish.append(self.layoutFinished) def layoutFinished(self): self.setTitle(self.setup_title) def startHotplug(self): self.hw.on_hotplug.append(self.createSetup) def stopHotplug(self): self.hw.on_hotplug.remove(self.createSetup) def createSetup(self): level = config.usage.setup_level.index self.list = [ ] if level >= 1: if SystemInfo["CanPcmMultichannel"]: self.list.append(getConfigListEntry(_("PCM Multichannel"), config.av.pcm_multichannel, _("Choose whether multi channel sound tracks should be output as PCM."))) if SystemInfo["CanDownmixAC3"]: self.list.append(getConfigListEntry(_("Dolby Digital / DTS downmix"), config.av.downmix_ac3, _("Choose whether multi channel sound tracks should be downmixed to stereo."))) if SystemInfo["CanDownmixAAC"]: self.list.append(getConfigListEntry(_("AAC downmix"), config.av.downmix_aac, _("Choose whether multi channel sound tracks should be downmixed to stereo."))) if SystemInfo["Canaudiosource"]: self.list.append(getConfigListEntry(_("Audio Source"), config.av.audio_source, _("Choose whether multi channel sound tracks should be convert to PCM or SPDIF."))) if SystemInfo["CanAACTranscode"]: self.list.append(getConfigListEntry(_("AAC transcoding"), config.av.transcodeaac, _("Choose whether AAC sound tracks should be transcoded."))) self.list.extend(( getConfigListEntry(_("General AC3 delay"), config.av.generalAC3delay, _("This option configures the general audio delay of Dolby Digital sound tracks.")), getConfigListEntry(_("General PCM delay"), config.av.generalPCMdelay, _("This option configures the general audio delay of stereo sound tracks.")) )) if SystemInfo["Can3DSurround"]: self.list.append(getConfigListEntry(_("3D Surround"), config.av.surround_3d,_("This option allows you to enable 3D Surround Sound."))) if SystemInfo["Can3DSpeaker"] and config.av.surround_3d.value != "none": self.list.append(getConfigListEntry(_("3D Surround Speaker Position"), config.av.surround_3d_speaker,_("This option allows you to change the virtuell loadspeaker position."))) if SystemInfo["CanAutoVolume"]: self.list.append(getConfigListEntry(_("Audio Auto Volume Level"), config.av.autovolume,_("This option configures you can set Auto Volume Level."))) if SystemInfo["Canedidchecking"]: self.list.append(getConfigListEntry(_("Bypass HDMI EDID Check"), config.av.bypass_edid_checking,_("This option allows you to bypass HDMI EDID check"))) self["config"].list = self.list self["config"].l.setList(self.list) if config.usage.sort_settings.value: self["config"].list.sort() def keyLeft(self): ConfigListScreen.keyLeft(self) self.createSetup() def keyRight(self): ConfigListScreen.keyRight(self) self.createSetup() def confirm(self, confirmed): self.keySave() def apply(self): self.keySave() # for summary: def changedEntry(self): for x in self.onChangedEntry: x() def getCurrentEntry(self): return self["config"].getCurrent()[0] def getCurrentValue(self): return str(self["config"].getCurrent()[1].getText()) def getCurrentDescription(self): return self["config"].getCurrent() and len(self["config"].getCurrent()) > 2 and self["config"].getCurrent()[2] or "" def createSummary(self): from Screens.Setup import SetupSummary return SetupSummary class AutoVideoModeLabel(Screen): def __init__(self, session): Screen.__init__(self, session) self["content"] = Label() self["restxt"] = Label() self.hideTimer = eTimer() self.hideTimer.callback.append(self.hide) self.onShow.append(self.hide_me) def hide_me(self): idx = config.av.autores_label_timeout.index if idx: idx += 4 self.hideTimer.start(idx*1000, True) previous = None isDedicated3D = False def applySettings(mode=config.osd.threeDmode.value, znorm=int(config.osd.threeDznorm.value)): global previous, isDedicated3D mode = isDedicated3D and mode == "auto" and "sidebyside" or mode if previous != (mode, znorm): try: previous = (mode, znorm) if SystemInfo["CanUse3DModeChoices"]: f = open("/proc/stb/fb/3dmode_choices", "r") choices = f.readlines()[0].split() f.close() if mode not in choices: if mode == "sidebyside": mode = "sbs" elif mode == "topandbottom": mode = "tab" elif mode == "auto": mode = "off" open(SystemInfo["3DMode"], "w").write(mode) open(SystemInfo["3DZNorm"], "w").write('%d' % znorm) except: return class AutoVideoMode(Screen): def __init__(self, session): Screen.__init__(self, session) self.__event_tracker = ServiceEventTracker(screen=self, eventmap= { iPlayableService.evStart: self.__evStart, iPlayableService.evVideoSizeChanged: self.VideoChanged, iPlayableService.evVideoProgressiveChanged: self.VideoChanged, iPlayableService.evVideoFramerateChanged: self.VideoChanged, iPlayableService.evBuffering: self.BufferInfo, iPlayableService.evStopped: self.BufferInfoStop }) self.delay = False self.bufferfull = True self.detecttimer = eTimer() self.detecttimer.callback.append(self.VideoChangeDetect) def checkIfDedicated3D(self): service = self.session.nav.getCurrentlyPlayingServiceReference() servicepath = service and service.getPath() if servicepath and servicepath.startswith("/"): if service.toString().startswith("1:"): info = eServiceCenter.getInstance().info(service) service = info and info.getInfoString(service, iServiceInformation.sServiceref) return service and eDVBDB.getInstance().getFlag(eServiceReference(service)) & FLAG_IS_DEDICATED_3D == FLAG_IS_DEDICATED_3D and "sidebyside" else: return ".3d." in servicepath.lower() and "sidebyside" or ".tab." in servicepath.lower() and "topandbottom" service = self.session.nav.getCurrentService() info = service and service.info() return info and info.getInfo(iServiceInformation.sIsDedicated3D) == 1 and "sidebyside" def __evStart(self): if config.osd.threeDmode.value == "auto": global isDedicated3D isDedicated3D = self.checkIfDedicated3D() if isDedicated3D: applySettings(isDedicated3D) else: applySettings() def BufferInfo(self): bufferInfo = self.session.nav.getCurrentService().streamed().getBufferCharge() if bufferInfo[0] > 98: self.bufferfull = True self.VideoChanged() else: self.bufferfull = False def BufferInfoStop(self): self.bufferfull = True def VideoChanged(self): if self.session.nav.getCurrentlyPlayingServiceReference() and not self.session.nav.getCurrentlyPlayingServiceReference().toString().startswith('4097:'): delay = config.av.autores_delay.value else: delay = config.av.autores_delay.value * 2 if not self.detecttimer.isActive() and not self.delay: self.delay = True self.detecttimer.start(delay) else: self.delay = True self.detecttimer.stop() self.detecttimer.start(delay) def VideoChangeDetect(self): global resolutionlabel config_port = config.av.videoport.value config_mode = str(config.av.videomode[config_port].value).replace('\n','') config_res = str(config.av.videomode[config_port].value[:-1]).replace('\n','') config_pol = str(config.av.videomode[config_port].value[-1:]).replace('\n','') config_rate = str(config.av.videorate[config_mode].value).replace('Hz','').replace('\n','') f = open("/proc/stb/video/videomode") current_mode = f.read()[:-1].replace('\n','') f.close() if current_mode.upper() in ('PAL', 'NTSC'): current_mode = current_mode.upper() current_pol = '' if 'i' in current_mode: current_pol = 'i' elif 'p' in current_mode: current_pol = 'p' current_res = current_pol and current_mode.split(current_pol)[0].replace('\n','') or "" current_rate = current_pol and current_mode.split(current_pol)[0].replace('\n','') and current_mode.split(current_pol)[1].replace('\n','') or "" video_height = None video_width = None video_pol = None video_rate = None if path.exists("/proc/stb/vmpeg/0/yres"): try: f = open("/proc/stb/vmpeg/0/yres", "r") video_height = int(f.read(),16) f.close() except: video_height = 0 if path.exists("/proc/stb/vmpeg/0/xres"): try: f = open("/proc/stb/vmpeg/0/xres", "r") video_width = int(f.read(),16) f.close() except: video_width = 0 if path.exists("/proc/stb/vmpeg/0/progressive"): try: f = open("/proc/stb/vmpeg/0/progressive", "r") video_pol = "p" if int(f.read(),16) else "i" f.close() except: video_pol = "i" if path.exists("/proc/stb/vmpeg/0/framerate"): f = open("/proc/stb/vmpeg/0/framerate", "r") try: video_rate = int(f.read()) except: video_rate = 50 f.close() if not video_height or not video_width or not video_pol or not video_rate: service = self.session.nav.getCurrentService() if service is not None: info = service.info() else: info = None if info: video_height = int(info.getInfo(iServiceInformation.sVideoHeight)) video_width = int(info.getInfo(iServiceInformation.sVideoWidth)) video_pol = ("i", "p")[info.getInfo(iServiceInformation.sProgressive)] video_rate = int(info.getInfo(iServiceInformation.sFrameRate)) if (video_height and video_width and video_pol and video_rate) or (config.av.smart1080p.value != 'false'): resolutionlabel["content"].setText(_("Video content: %ix%i%s %iHz") % (video_width, video_height, video_pol, (video_rate + 500) / 1000)) if video_height != -1: if video_height > 720 or video_width > 1280: new_res = "1080" elif (576 < video_height <= 720) or video_width > 1024: new_res = "720" elif (480 < video_height <= 576) or video_width > 720 or video_rate in (25000, 23976, 24000): new_res = "576" else: new_res = "480" else: new_res = config_res if video_rate != -1: if video_rate == 25000 and video_pol == 'i': new_rate = 50000 elif video_rate == 59940 or (video_rate == 29970 and video_pol == 'i') or (video_rate == 29970 and video_pol == 'p' and config.av.autores.value == 'disabled'): new_rate = 60000 elif video_rate == 23976: new_rate = 24000 elif video_rate == 29970: new_rate = 30000 else: new_rate = video_rate new_rate = str((new_rate + 500) / 1000) else: new_rate = config_rate if video_pol != -1: new_pol = str(video_pol) else: new_pol = config_pol write_mode = None new_mode = None if config_mode in ('PAL', 'NTSC'): write_mode = config_mode elif config.av.autores.value == 'all' or (config.av.autores.value == 'hd' and int(new_res) >= 720): if (config.av.autores_deinterlace.value and HardwareInfo().is_nextgen()) or (config.av.autores_deinterlace.value and not HardwareInfo().is_nextgen() and int(new_res) <= 720): new_pol = new_pol.replace('i','p') if new_res+new_pol+new_rate in iAVSwitch.modes_available: new_mode = new_res+new_pol+new_rate if new_mode == '480p24' or new_mode == '576p24': new_mode = config.av.autores_480p24.value if new_mode == '720p24': new_mode = config.av.autores_720p24.value if new_mode == '1080p24': new_mode = config.av.autores_1080p24.value if new_mode == '1080p25': new_mode = config.av.autores_1080p25.value if new_mode == '1080p30': new_mode = config.av.autores_1080p30.value elif new_res+new_pol in iAVSwitch.modes_available: new_mode = new_res+new_pol else: new_mode = config_mode+new_rate write_mode = new_mode elif config.av.autores.value == 'hd' and int(new_res) <= 576: if (config.av.autores_deinterlace.value and HardwareInfo().is_nextgen()) or (config.av.autores_deinterlace.value and not HardwareInfo().is_nextgen() and not config.av.autores_sd.value == '1080i'): new_mode = config.av.autores_sd.value.replace('i','p')+new_rate else: if new_pol in 'p': new_mode = config.av.autores_sd.value.replace('i','p')+new_rate else: new_mode = config.av.autores_sd.value+new_rate if new_mode == '720p24': new_mode = config.av.autores_720p24.value if new_mode == '1080p24': new_mode = config.av.autores_1080p24.value if new_mode == '1080p25': new_mode = config.av.autores_1080p25.value if new_mode == '1080p30': new_mode = config.av.autores_1080p30.value write_mode = new_mode else: if path.exists('/proc/stb/video/videomode_%shz' % new_rate) and config_rate == 'multi': f = open("/proc/stb/video/videomode_%shz" % new_rate, "r") multi_videomode = f.read().replace('\n','') f.close() if multi_videomode and (current_mode != multi_videomode): write_mode = multi_videomode else: write_mode = config_mode+new_rate # workaround for bug, see http://www.droidsat.org/forum/showthread.php?1642-Autoresolution-Plugin&p=38836&viewfull=1#post38836 # always use a fixed resolution and frame rate (e.g. 1080p50 if supported) for TV or .ts files # always use a fixed resolution and correct rate (e.g. 1080p24/p50/p60 for all other videos if config.av.smart1080p.value != 'false': ref = self.session.nav.getCurrentlyPlayingServiceReference() if ref is not None: try: mypath = ref.getPath() except: mypath = '' else: mypath = '' # no frame rate information available, check if filename (or directory name) contains a hint # (allow user to force a frame rate this way): if (mypath.find('p24.') >= 0) or (mypath.find('24p.') >= 0): new_rate = '24' elif (mypath.find('p25.') >= 0) or (mypath.find('25p.') >= 0): new_rate = '25' elif (mypath.find('p30.') >= 0) or (mypath.find('30p.') >= 0): new_rate = '30' elif (mypath.find('p50.') >= 0) or (mypath.find('50p.') >= 0): new_rate = '50' elif (mypath.find('p60.') >= 0) or (mypath.find('60p.') >= 0): new_rate = '60' elif new_rate == 'multi': new_rate = '' # omit frame rate specifier, e.g. '1080p' instead of '1080p50' if there is no clue if mypath != '': if mypath.endswith('.ts'): print "DEBUG VIDEOMODE/ playing .ts file" new_rate = '50' # for .ts files else: print "DEBUG VIDEOMODE/ playing other (non .ts) file" # new_rate from above for all other videos else: print "DEBUG VIDEOMODE/ no path or no service reference, presumably live TV" new_rate = '50' # for TV / or no service reference, then stay at 1080p50 new_rate = new_rate.replace('25', '50') new_rate = new_rate.replace('30', '60') if (config.av.smart1080p.value == '1080p50') or (config.av.smart1080p.value == 'true'): # for compatibility with old ConfigEnableDisable write_mode = '1080p' + new_rate elif config.av.smart1080p.value == '2160p50': write_mode = '2160p' + new_rate elif config.av.smart1080p.value == '1080i50': if new_rate == '24': write_mode = '1080p24' # instead of 1080i24 else: write_mode = '1080i' + new_rate elif config.av.smart1080p.value == '720p50': write_mode = '720p' + new_rate print "[VideoMode] smart1080p mode, selecting ",write_mode if write_mode and current_mode != write_mode and self.bufferfull: # first we read now the real available values for every stb, # before we try to write the new mode changeResolution = False try: if path.exists("/proc/stb/video/videomode_choices"): vf = open("/proc/stb/video/videomode_choices") values = vf.readline().replace("\n", "").split(" ", -1) for x in values: if x == write_mode: try: f = open("/proc/stb/video/videomode", "w") f.write(write_mode) f.close() changeResolution = True except Exception, e: print("[VideoMode] write_mode exception:" + str(e)) if not changeResolution: print "[VideoMode] setMode - port: %s, mode: %s is not available" % (config_port, write_mode) resolutionlabel["restxt"].setText(_("Video mode: %s not available") % write_mode) # we try to go for not available 1080p24/1080p30/1080p60 to change to 1080p from 60hz_choices if available # TODO: can we make it easier, or more important --> smaller ? # should we outsourced that way, like two new "def ..." # or some other stuff, not like this? if (write_mode == "1080p24") or (write_mode == "1080p30") or (write_mode == "1080p60"): for x in values: if x == "1080p": try: f = open("/proc/stb/video/videomode", "w") f.write(x) f.close() changeResolution = True except Exception, e: print("[VideoMode] write_mode exception:" + str(e)) if not changeResolution: print "[VideoMode] setMode - port: %s, mode: 1080p is also not available" % config_port resolutionlabel["restxt"].setText(_("Video mode: 1080p also not available")) else: print "[VideoMode] setMode - port: %s, mode: %s" % (config_port, x) resolutionlabel["restxt"].setText(_("Video mode: %s") % x) if (write_mode == "2160p24") or (write_mode == "2160p30") or (write_mode == "2160p60"): for x in values: if x == "2160p": try: f = open("/proc/stb/video/videomode", "w") f.write(x) f.close() changeResolution = True except Exception, e: print("[VideoMode] write_mode exception:" + str(e)) if not changeResolution: print "[VideoMode] setMode - port: %s, mode: 2160p is also not available" % config_port resolutionlabel["restxt"].setText(_("Video mode: 2160p also not available")) else: print "[VideoMode] setMode - port: %s, mode: %s" % (config_port, x) resolutionlabel["restxt"].setText(_("Video mode: %s") % x) else: resolutionlabel["restxt"].setText(_("Video mode: %s") % write_mode) print "[VideoMode] setMode - port: %s, mode: %s" % (config_port, write_mode) if config.av.autores.value != "disabled" and config.av.autores_label_timeout.value != '0': resolutionlabel.show() vf.close() except Exception, e: print("[VideoMode] read videomode_choices exception:" + str(e)) elif write_mode and current_mode != write_mode: # the resolution remained stuck at a wrong setting after streaming when self.bufferfull was False (should be fixed now after adding BufferInfoStop) print "[VideoMode] not changing from",current_mode,"to",write_mode,"as self.bufferfull is",self.bufferfull iAVSwitch.setAspect(config.av.aspect) iAVSwitch.setWss(config.av.wss) iAVSwitch.setPolicy43(config.av.policy_43) iAVSwitch.setPolicy169(config.av.policy_169) self.delay = False self.detecttimer.stop() def autostart(session): global resolutionlabel if not path.exists(resolveFilename(SCOPE_PLUGINS)+'SystemPlugins/AutoResolution'): if resolutionlabel is None: resolutionlabel = session.instantiateDialog(AutoVideoModeLabel) AutoVideoMode(session) else: config.av.autores.setValue(False) config.av.autores.save() configfile.save() ```
[ { "content": "Repeat the code exactly:\n```python\n'''Simple classifier model\n\n'''\n\nfrom cortex.main import run\nfrom cortex.plugins import ModelPlugin\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom cortex.built_ins.models.utils import update_encoder_args\n\n\nclass MyClassi...
[ { "content": "Repeat the code exactly:\n<|memory_start|>```python\n'''Simple classifier model\n\n'''\n\nfrom cortex.main import run\nfrom cortex.plugins import ModelPlugin\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom cortex.built_ins.models.utils import update_encoder_args\n\n...
```python '''Simple classifier model ''' from cortex.main import run from cortex.plugins import ModelPlugin import torch import torch.nn as nn import torch.nn.functional as F from cortex.built_ins.models.utils import update_encoder_args class MyClassifier(ModelPlugin): '''Basic image classifier. Classifies images using standard convnets. ''' defaults = dict( data=dict(batch_size=128, inputs=dict(inputs='images')), optimizer=dict(optimizer='Adam', learning_rate=1e-3), train=dict(epochs=200, save_on_best='losses.classifier')) def build(self, classifier_type='convnet', classifier_args=dict(dropout=0.2)): '''Builds a simple image classifier. Args: classifier_type (str): Network type for the classifier. classifier_args: Classifier arguments. Can include dropout, batch_norm, layer_norm, etc. ''' classifier_args = classifier_args or {} shape = self.get_dims('x', 'y', 'c') dim_l = self.get_dims('labels') Encoder, args = update_encoder_args( shape, model_type=classifier_type, encoder_args=classifier_args) args.update(**classifier_args) classifier = Encoder(shape, dim_out=dim_l, **args) self.nets.classifier = classifier def routine(self, inputs, targets, criterion=nn.CrossEntropyLoss()): ''' Args: criterion: Classifier criterion. ''' classifier = self.nets.classifier outputs = classifier(inputs) predicted = torch.max(F.log_softmax(outputs, dim=1).data, 1)[1] loss = criterion(outputs, targets) correct = 100. * predicted.eq( targets.data).cpu().sum() / targets.size(0) self.losses.classifier = loss self.results.accuracy = correct def predict(self, inputs): classifier = self.nets.classifier outputs = classifier(inputs) predicted = torch.max(F.log_softmax(outputs, dim=1).data, 1)[1] return predicted def visualize(self, images, inputs, targets): predicted = self.predict(inputs) self.add_image(images.data, labels=(targets.data, predicted.data), name='gt_pred') if __name__ == '__main__': classifier = MyClassifier() run(model=classifier) ```
[ { "content": "Write out the code verbatim, preserving indentation and whitespace:\n```python\nimport re\n\nfrom streamlink.plugin import Plugin, pluginmatcher\nfrom streamlink.plugin.api import validate\nfrom streamlink.stream import HLSStream\nfrom streamlink.utils import parse_json\n\n\n@pluginmatcher(re.comp...
[ { "content": "Write out the code verbatim, preserving indentation and whitespace:\n<|memory_start|>```python\nimport re\n\nfrom streamlink.plugin import Plugin, pluginmatcher\nfrom streamlink.plugin.api import validate\nfrom streamlink.stream import HLSStream\nfrom streamlink.utils import parse_json\n\n\n@plugi...
```python import re from streamlink.plugin import Plugin, pluginmatcher from streamlink.plugin.api import validate from streamlink.stream import HLSStream from streamlink.utils import parse_json @pluginmatcher(re.compile( r"https?://www\.cbsnews\.com/live/" )) class CBSNews(Plugin): _re_default_payload = re.compile(r"CBSNEWS.defaultPayload = (\{.*)") _schema_items = validate.Schema( validate.transform(_re_default_payload.search), validate.any(None, validate.all( validate.get(1), validate.transform(parse_json), {"items": [validate.all({ "video": validate.url(), "format": "application/x-mpegURL" }, validate.get("video"))]}, validate.get("items") )) ) def _get_streams(self): items = self.session.http.get(self.url, schema=self._schema_items) if items: for hls_url in items: yield from HLSStream.parse_variant_playlist(self.session, hls_url).items() __plugin__ = CBSNews ```
[ { "content": "Produce an exact reconstruction of the code:\n```python\n# Copyright (c) 2012 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\nimport logging\nimport multiprocessing\n\nfrom pylib import android_...
[ { "content": "Produce an exact reconstruction of the code:\n<|memory_start|>```python\n# Copyright (c) 2012 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\nimport logging\nimport multiprocessing\n\nfrom pylib...
```python # Copyright (c) 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. import logging import multiprocessing from pylib import android_commands from pylib.base.test_result import TestResults from pylib.forwarder import Forwarder def _ShardedTestRunnable(test): """Standalone function needed by multiprocessing.Pool.""" log_format = '[' + test.device + '] # %(asctime)-15s: %(message)s' if logging.getLogger().handlers: logging.getLogger().handlers[0].setFormatter(logging.Formatter(log_format)) else: logging.basicConfig(format=log_format) # Handle SystemExit here since python has a bug to exit current process try: return test.Run() except SystemExit: return TestResults() def SetTestsContainer(tests_container): """Sets tests container. multiprocessing.Queue can't be pickled across processes, so we need to set this as a 'global', per process, via multiprocessing.Pool. """ BaseTestSharder.tests_container = tests_container class BaseTestSharder(object): """Base class for sharding tests across multiple devices. Args: attached_devices: A list of attached devices. """ # See more in SetTestsContainer. tests_container = None def __init__(self, attached_devices, build_type='Debug'): self.attached_devices = attached_devices # Worst case scenario: a device will drop offline per run, so we need # to retry until we're out of devices. # TODO(frankf): There are two sources of flakiness: # 1. Device flakiness # 2. Test/product flakiness # We should differentiate between these. Otherwise, blindly retrying tests # might mask test/product flakiness. For type 2, we should follow the # general chrome best practices. self.retries = len(self.attached_devices) self.tests = [] self.build_type = build_type def CreateShardedTestRunner(self, device, index): """Factory function to create a suite-specific test runner. Args: device: Device serial where this shard will run index: Index of this device in the pool. Returns: An object of BaseTestRunner type (that can provide a "Run()" method). """ pass def SetupSharding(self, tests): """Called before starting the shards.""" pass def _KillHostForwarder(self): Forwarder.KillHost(self.build_type) def RunShardedTests(self): """Runs the tests in all connected devices. Returns: A TestResults object. """ logging.warning('*' * 80) logging.warning('Sharding in ' + str(len(self.attached_devices)) + ' devices.') logging.warning('Note that the output is not synchronized.') logging.warning('Look for the "Final result" banner in the end.') logging.warning('*' * 80) final_results = TestResults() self._KillHostForwarder() for retry in xrange(self.retries): logging.warning('Try %d of %d', retry + 1, self.retries) logging.warning('Attempting to run %d tests: %s' % (len(self.tests), self.tests)) self.SetupSharding(self.tests) test_runners = [] # Try to create N shards, and retrying on failure. try: for index, device in enumerate(self.attached_devices): logging.warning('*' * 80) logging.warning('Creating shard %d for %s', index, device) logging.warning('*' * 80) test_runner = self.CreateShardedTestRunner(device, index) test_runners += [test_runner] except android_commands.errors.DeviceUnresponsiveError as e: logging.critical('****Failed to create a shard: [%s]', e) self.attached_devices.remove(device) continue logging.warning('Starting...') pool = multiprocessing.Pool(len(self.attached_devices), SetTestsContainer, [BaseTestSharder.tests_container]) # map can't handle KeyboardInterrupt exception. It's a python bug. # So use map_async instead. async_results = pool.map_async(_ShardedTestRunnable, test_runners) try: results_lists = async_results.get(999999) except android_commands.errors.DeviceUnresponsiveError as e: logging.critical('****Failed to run test: [%s]', e) self.attached_devices = android_commands.GetAttachedDevices() continue test_results = TestResults.FromTestResults(results_lists) # Re-check the attached devices for some devices may # become offline retry_devices = set(android_commands.GetAttachedDevices()) # Remove devices that had exceptions. retry_devices -= TestResults.DeviceExceptions(results_lists) # Retry on devices that didn't have any exception. self.attached_devices = list(retry_devices) # TODO(frankf): Do not break TestResults encapsulation. if (retry == self.retries - 1 or len(self.attached_devices) == 0): all_passed = final_results.ok + test_results.ok final_results = test_results final_results.ok = all_passed break else: final_results.ok += test_results.ok self.tests = [] for t in test_results.GetAllBroken(): self.tests += [t.name] if not self.tests: break else: # We ran out retries, possibly out of healthy devices. # There's no recovery at this point. raise Exception('Unrecoverable error while retrying test runs.') self._KillHostForwarder() return final_results ```
[ { "content": "```python\n#\n# \n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE F...
[ { "content": "<|memory_start|>```python\n#\n# \n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOL...
```python # # # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. # """ This package provides some bit twiddling functions by: Carl J. Nobile email: carl.nobile@gmail.com """ __docformat__ = "restructuredtext en" class Utilities(object): def reverseByteOrder(self, data): """ Reverses the byte order of an int (16-bit) or long (32-bit) value. """ # Courtesy Vishal Sapre byteCount = len(hex(data)[2:].replace('L','')[::2]) val = 0 for i in range(byteCount): val = (val << 8) | (data & 0xff) data >>= 8 return val ```
[ { "content": "Produce an exact reconstruction of the code:\n```python\n# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license i...
[ { "content": "Produce an exact reconstruction of the code:\n<|memory_start|>```python\n# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root ...
```python # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- import uuid from msrest.pipeline import ClientRawResponse from .. import models class Operations(object): """Operations operations. :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An objec model deserializer. :ivar api_version: Client API version. Constant value: "2017-04-01". """ models = models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self.api_version = "2017-04-01" self.config = config def list( self, custom_headers=None, raw=False, **operation_config): """Lists all of the available ServiceBus REST API operations. :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: An iterator like instance of Operation :rtype: ~azure.mgmt.servicebus.models.OperationPaged[~azure.mgmt.servicebus.models.Operation] :raises: :class:`ErrorResponseException<azure.mgmt.servicebus.models.ErrorResponseException>` """ def internal_paging(next_link=None, raw=False): if not next_link: # Construct URL url = '/providers/Microsoft.ServiceBus/operations' # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') else: url = next_link query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send( request, header_parameters, **operation_config) if response.status_code not in [200]: raise models.ErrorResponseException(self._deserialize, response) return response # Deserialize response deserialized = models.OperationPaged(internal_paging, self._deserialize.dependencies) if raw: header_dict = {} client_raw_response = models.OperationPaged(internal_paging, self._deserialize.dependencies, header_dict) return client_raw_response return deserialized ```
[ { "content": "Provide an exact copy of the source code:\n```python\n''' tools for working with .sas7bdat files '''\r\n\r\nimport os\r\nfrom collections import OrderedDict\r\n\r\nimport pandas as pd\r\nfrom sas7bdat import SAS7BDAT\r\n\r\nfrom .knowledge.questionnaires import map_ph4, map_ph4_ssaga\r\n\r\nmap_su...
[ { "content": "Provide an exact copy of the source code:\n<|memory_start|>```python\n''' tools for working with .sas7bdat files '''\r\n\r\nimport os\r\nfrom collections import OrderedDict\r\n\r\nimport pandas as pd\r\nfrom sas7bdat import SAS7BDAT\r\n\r\nfrom .knowledge.questionnaires import map_ph4, map_ph4_ssa...
```python ''' tools for working with .sas7bdat files ''' import os from collections import OrderedDict import pandas as pd from sas7bdat import SAS7BDAT from .knowledge.questionnaires import map_ph4, map_ph4_ssaga map_subject = {'core': {'file_pfixes': []}} parent_dir = '/processed_data/zork/zork-phase4-69/session/' n_header_lines = 30 def extract_descriptions(path): ''' given path to .sas7bdat file, returns dictionary mapping column labels to their verbose descriptions in the SAS header. dictionary will only contain an entry if there was new information present (if there was a description, and it was different from the label) ''' f = SAS7BDAT(path) kmap = OrderedDict() for line in str(f.header).splitlines()[n_header_lines + 1:]: line_parts = line.split(maxsplit=4) label = line_parts[1] try: description = line_parts[4].rstrip() if description == label or description[0] == '$': continue else: kmap[label] = description except IndexError: pass return kmap def exemplary_files(kdict): ''' given a questionnaire knowledge map, return a new dictionary mapping questionnaire names to the filepath of an exemplary .sas7bdat file for each file prefix ''' exemplars = {} for test, tdict in kdict.items(): for fpx in tdict['file_pfixes']: fd = parent_dir + test fn = fpx + '.sas7bdat' fp = os.path.join(fd, fn) if os.path.exists(fp): exemplars[test] = fp else: print(fp, 'did not exist') return exemplars def build_labelmaps(): ''' return a dict in which keys are questionnaires names and values are dictionaries mapping column labels to descriptions ''' comb_dict = map_ph4.copy() comb_dict.update(map_ph4_ssaga) exemplars = exemplary_files(comb_dict) big_kmap = {} for test, fp in exemplars.items(): kmap = extract_descriptions(fp) big_kmap[test] = kmap return big_kmap def df_fromsas(fullpath, id_lbl='ind_id'): ''' convert .sas7bdat to dataframe. unused because fails on incorrectly formatted files. ''' # read csv in as dataframe df = pd.read_sas(fullpath, format='sas7bdat') # convert id to str and save as new column df[id_lbl] = df[id_lbl].apply(int).apply(str) df['ID'] = df[id_lbl] return df ```
[ { "content": "Output the full code verbatim (no extra comments):\n```python\n\"\"\"\nStartup script for the pyTanks server\n\nRequirements:\n Python 3.5 or newer\n websockets 7.0 (pip install websockets==7.0)\n\nUsage:\n python start.py\n\n The pyTanks server uses the settings found in config.py to ...
[ { "content": "Output the full code verbatim (no extra comments):\n<|memory_start|>```python\n\"\"\"\nStartup script for the pyTanks server\n\nRequirements:\n Python 3.5 or newer\n websockets 7.0 (pip install websockets==7.0)\n\nUsage:\n python start.py\n\n The pyTanks server uses the settings found ...
```python """ Startup script for the pyTanks server Requirements: Python 3.5 or newer websockets 7.0 (pip install websockets==7.0) Usage: python start.py The pyTanks server uses the settings found in config.py to control how the server works. Those values can be changed directly or be overridden by appending one or more of these command line args: log=n - Overrides the default logging level. (See the usage section of the readme.) ip:port - Overrides the ip and port used to host the server. """ import sys import config def main(): """ Check the environment, apply any command line args to config.py, and start wsServer.py """ # Check Python version if sys.version_info[0] < 3 or sys.version_info[1] < 5: print("Python 3.5 or newer is required to run the pyTanks server") return # Check for websockets from importlib import util if util.find_spec("websockets") is None: print("The websockets module is required to run the pyTanks server") return # Import the code that requires the above things from serverLogic.wsServer import runServer # Parse and apply the args for arg in sys.argv: if arg == sys.argv[0] or arg == "": continue elif arg.startswith("log="): try: config.server.logLevel = int(arg[-1:]) except ValueError: print("Invalid log level") return elif arg.startswith("minPlayers="): try: num = int(arg[-1:]) if num <= 1: print("minPlayers must be greater than 1") return config.server.minPlayers = num except ValueError: print("Invalid min player count") return elif ":" in arg: config.server.ipAndPort = arg else: print(__doc__[__doc__.index("Usage:"):].strip()) return # Start the server runServer() if __name__ == "__main__": main() ```
[ { "content": "```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Author: Janice Cheng\n\n\"\"\"\n把诘个服务器端改成在一个类中\n\n\"\"\"\n\nimport pika\nimport subprocess\n\ncredentials = pika.PlainCredentials('janice', 'janice123')\nparameters = pika.ConnectionParameters('172.16.201.134', 5672, '/', credentials)\nc...
[ { "content": "<|memory_start|>```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Author: Janice Cheng\n\n\"\"\"\n把诘个服务器端改成在一个类中\n\n\"\"\"\n\nimport pika\nimport subprocess\n\ncredentials = pika.PlainCredentials('janice', 'janice123')\nparameters = pika.ConnectionParameters('172.16.201.134', 5672, '/',...
```python #!/usr/bin/env python # -*- coding: utf-8 -*- # Author: Janice Cheng """ 把诘个服务器端改成在一个类中 """ import pika import subprocess credentials = pika.PlainCredentials('janice', 'janice123') parameters = pika.ConnectionParameters('172.16.201.134', 5672, '/', credentials) connection = pika.BlockingConnection(parameters) channel = connection.channel() channel.queue_declare(queue='rpc_queue') # 把这里改成 subprocess 来执行这条命令 def cmd_func(cmd): cmd_data = subprocess.Popen(cmd, shell=True, stdin=subprocess.PIPE, stdout=subprocess.PIPE) return cmd_data.stdout.read() def on_request(ch, method, props, body): n = body.decode() print(" [.] Calling (%s)" % n) response = cmd_func(n) response = str(response,encoding='utf-8') ch.basic_publish(exchange='', routing_key=props.reply_to, properties=pika.BasicProperties(correlation_id = \ props.correlation_id), body=response) ch.basic_ack(delivery_tag = method.delivery_tag) channel.basic_qos(prefetch_count=1) channel.basic_consume(on_request, queue='rpc_queue') #在对列里获取数据 print(" [x] Awaiting RPC requests") channel.start_consuming() ```
[ { "content": "```python\nimport unittest\nimport transaction\n\nfrom pyramid import testing\n\nfrom .models import DBSession\n\n\nclass TestMyViewSuccessCondition(unittest.TestCase):\n def setUp(self):\n self.config = testing.setUp()\n from sqlalchemy import create_engine\n engine = crea...
[ { "content": "<|memory_start|>```python\nimport unittest\nimport transaction\n\nfrom pyramid import testing\n\nfrom .models import DBSession\n\n\nclass TestMyViewSuccessCondition(unittest.TestCase):\n def setUp(self):\n self.config = testing.setUp()\n from sqlalchemy import create_engine\n ...
```python import unittest import transaction from pyramid import testing from .models import DBSession class TestMyViewSuccessCondition(unittest.TestCase): def setUp(self): self.config = testing.setUp() from sqlalchemy import create_engine engine = create_engine('sqlite://') from .models import ( Base, MyModel, ) DBSession.configure(bind=engine) Base.metadata.create_all(engine) with transaction.manager: model = MyModel(name='one', value=55) DBSession.add(model) def tearDown(self): DBSession.remove() testing.tearDown() def test_passing_view(self): from .views import my_view request = testing.DummyRequest() info = my_view(request) self.assertEqual(info['one'].name, 'one') self.assertEqual(info['project'], 'pyramid_pycharm') class TestMyViewFailureCondition(unittest.TestCase): def setUp(self): self.config = testing.setUp() from sqlalchemy import create_engine engine = create_engine('sqlite://') from .models import ( Base, MyModel, ) DBSession.configure(bind=engine) def tearDown(self): DBSession.remove() testing.tearDown() def test_failing_view(self): from .views import my_view request = testing.DummyRequest() info = my_view(request) self.assertEqual(info.status_int, 500) ```
[ { "content": "Replicate the source code:\n```python\n#!/usr/bin/env python\n\nfrom gi.repository import Gtk, Gdk\n\nimport pairdefs\n\nclass Picker:\n '''RandomSK UI and related functions.'''\n def pick_rand(self, widget=None, data=None):\n '''Pick a random sequence from our current pairdict and se...
[ { "content": "Replicate the source code:\n<|memory_start|>```python\n#!/usr/bin/env python\n\nfrom gi.repository import Gtk, Gdk\n\nimport pairdefs\n\nclass Picker:\n '''RandomSK UI and related functions.'''\n def pick_rand(self, widget=None, data=None):\n '''Pick a random sequence from our current...
```python #!/usr/bin/env python from gi.repository import Gtk, Gdk import pairdefs class Picker: '''RandomSK UI and related functions.''' def pick_rand(self, widget=None, data=None): '''Pick a random sequence from our current pairdict and set output text.''' self.disp.set_text(self.pairdict.pick()) def copy(self, widget, data=None): '''Copy displayed text to the system clipboard.''' self.clipboard.set_text(self.disp.get_text(), -1) def destroy(self, widget, data=None): '''Quit function.''' Gtk.main_quit() def update_title(self): '''Update title string to reflect current pairdict name.''' self.title = "RandomSK - %s" % self.key self.window.set_title(self.title) def choose_pair(self, widget=None, data=None): '''Store selected pairdict name and pairdict.''' self.key = self.pairs_store[self.picker.get_active_iter()][0] self.pairdict = self.pairs[self.key] self.update_title() def __init__(self): '''Set up UI and internal vars.''' self.builder = Gtk.Builder() self.builder.add_from_file("randomsk.ui") self.title = "RandomSK" #initial title string self.pairs = pairdefs.pairs # get refs to every part we'll be manipulating self.disp = self.builder.get_object("output_text") #text field for output self.picker = self.builder.get_object("pair_select") #combobox for choosing pairdict # set up the picker's label store and attach self.pairs_store = Gtk.ListStore(str) for k in sorted(self.pairs.keys()): self.pairs_store.append([k]) self.picker.set_model(self.pairs_store) renderer_text = Gtk.CellRendererText() self.picker.pack_start(renderer_text, True) self.picker.add_attribute(renderer_text, "text", 0) # set up signal handlers handlers_main = { "app.quit": self.destroy, "app.generate": self.pick_rand, "app.do_copy": self.copy, "app.pick": self.choose_pair } self.builder.connect_signals(handlers_main) # create a clipboard for easy copying self.clipboard = Gtk.Clipboard.get(Gdk.SELECTION_CLIPBOARD) # set our initial title and display the lot self.window = self.builder.get_object("main_window") self.window.set_title(self.title) self.window.show_all() # set the initial value for our dropdown, activating update triggers self.picker.set_active_iter(self.pairs_store[0].iter) self.pick_rand() #pick immediately def main(): Gtk.main() return # If the program is run directly or passed as an argument to the python # interpreter then create a Picker instance and show it if __name__ == "__main__": rsk = Picker() main() ```
[ { "content": "```python\n# Copyright 2014-2017 Camptocamp SA\n# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).\nfrom odoo.tests import common\n\n\nclass FixedFeesTester(common.TransactionCase):\n\n def setUp(self):\n \"\"\"Initialize credit control level mock to test fees computations\"...
[ { "content": "<|memory_start|>```python\n# Copyright 2014-2017 Camptocamp SA\n# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).\nfrom odoo.tests import common\n\n\nclass FixedFeesTester(common.TransactionCase):\n\n def setUp(self):\n \"\"\"Initialize credit control level mock to test fee...
```python # Copyright 2014-2017 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo.tests import common class FixedFeesTester(common.TransactionCase): def setUp(self): """Initialize credit control level mock to test fees computations""" super(FixedFeesTester, self).setUp() self.currency_model = self.env['res.currency'] self.euro = self.currency_model.search([('name', '=', 'EUR')]) self.assertTrue(self.euro) self.usd = self.currency_model.search([('name', '=', 'USD')]) self.assertTrue(self.usd) self.company = self.browse_ref('base.main_company') self.env.cr.execute( """UPDATE res_company SET currency_id = %s WHERE id = %s""", (self.company.id, self.euro.id), ) level_obj = self.env['credit.control.policy.level'] self.euro_level = level_obj.new({ 'name': 'Euro Level', 'dunning_fixed_amount': 5.0, 'dunning_currency_id': self.euro, 'dunning_type': 'fixed', }) self.usd_level = level_obj.new({ 'name': 'USD Level', 'dunning_fixed_amount': 5.0, 'dunning_currency_id': self.usd, 'dunning_type': 'fixed', }) self.dunning_model = self.env['credit.control.dunning.fees.computer'] self.line_model = self.env['credit.control.line'] def test_type_getter(self): """Test that correct compute function is returned for "fixed" type""" c_fun = self.dunning_model._get_compute_fun('fixed') self.assertEqual(c_fun, self.dunning_model.compute_fixed_fees) def test_unknow_type(self): """Test that non implemented error is raised if invalide fees type""" with self.assertRaises(NotImplementedError): self.dunning_model._get_compute_fun('bang') def test_computation_same_currency(self): """Test that fees are correctly computed with same currency""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': self.euro, 'company_id': self.company, }) fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertEqual(fees, self.euro_level.dunning_fixed_amount) def test_computation_different_currency(self): """Test that fees are correctly computed with different currency""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': self.usd.id, 'company_id': self.company, }) fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertNotEqual(fees, self.euro_level.dunning_fixed_amount) def test_computation_credit_currency_empty(self): """Test that fees are correctly computed with empty credit currency""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': False, 'company_id': self.company, }) fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertEqual(fees, self.euro_level.dunning_fixed_amount) def test_computation_level_currency_empty(self): """Test that fees are correctly computed with empty level currency""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': self.euro, 'company_id': self.company, }) self.euro_level.currency_id = False fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertEqual(fees, self.euro_level.dunning_fixed_amount) def test_computation_all_currency_empty(self): """Test that fees are correctly computed with empty currencies""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': False, 'company_id': self.company, }) self.euro_level.currency_id = False fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertEqual(fees, self.euro_level.dunning_fixed_amount) def test_no_fees(self): """Test that fees are not generated if no amount defined on level""" credit_line = self.line_model.new({ 'policy_level_id': self.euro_level, 'currency_id': self.usd, 'company_id': self.company, }) self.euro_level.dunning_fixed_amount = 0.0 fees = self.dunning_model.compute_fixed_fees(credit_line) self.assertEqual(fees, 0.0) ```
[ { "content": "Here is the source code:\n```python\n\"\"\"k5lib public functions.\n\nk5lib is a collection of functions and utilities to communicate with Fujits K5 cloud REST API.\n\n\"\"\"\nfrom .authenticate import get_global_token\nfrom .authenticate import get_region_token\nfrom .authenticate import get_proj...
[ { "content": "Here is the source code:\n<|memory_start|>```python\n\"\"\"k5lib public functions.\n\nk5lib is a collection of functions and utilities to communicate with Fujits K5 cloud REST API.\n\n\"\"\"\nfrom .authenticate import get_global_token\nfrom .authenticate import get_region_token\nfrom .authenticate...
```python """k5lib public functions. k5lib is a collection of functions and utilities to communicate with Fujits K5 cloud REST API. """ from .authenticate import get_global_token from .authenticate import get_region_token from .authenticate import get_project_token from .authenticate import get_domain_id from .authenticate import get_defaultproject_id from .authenticate import get_project_id from .authenticate import get_project_info from .contract import list_regions from .contract import get_region_info from .contract import activate_region from .contract import create_project from .contract import list_projects from .contract import delete_project from .orchestration import create_stack from .orchestration import delete_stack from .orchestration import get_stack_info from .orchestration import list_stacks from .orchestration import get_stack_id from .image import clone_vm from .image import get_volume_info from .image import list_images from .image import get_image_id from .image import get_image_info from .image import export_image from .image import share_image from .image import accept_image_share from .image import get_export_status from .image import get_image_import_queue_status from .compute import get_vnc_console_url from .compute import create_keypair from .compute import import_keypair from .compute import list_keypairs from .compute import create_server from .compute import create_server_with_ip from .compute import create_server_from_volume from .compute import delete_server from .compute import list_servers from .compute import get_server_password from .compute import get_server_name from .compute import get_server_id from .compute import get_server_info from .compute import add_server_interface from .compute import list_server_interfaces from .compute import get_server_interface_info from .compute import detach_server_interface from .compute import list_flavors from .compute import get_flavor_id from .network import create_network_connector from .network import list_network_connectors from .network import get_network_connector_id from .network import delete_network_connector from .network import create_network_connector_endpoint from .network import list_network_connector_endpoints from .network import list_network_connector_endpoint_interfaces from .network import get_network_connector_endpoint_id from .network import get_network_connector_endpoint_info from .network import connect_network_connector_endpoint from .network import disconnect_network_connector_endpoint from .network import delete_network_connector_endpoint from .network import create_port_on_network from .network import create_inter_project_connection from .network import delete_inter_project_connection from .network import update_inter_project_connection from .network import create_network from .network import delete_network from .network import list_networks from .network import get_network_id from .network import create_subnet from .network import delete_subnet from .network import list_subnets from .network import get_subnet_id from .network import find_first_free_ip from .network import list_ports from .network import get_port_id from .network import attach_floating_ip_to_port from .network import delete_port from .network import create_security_group from .network import _rest_delete_security_group from .network import list_security_groups from .network import get_security_group_id from .network import create_security_group_rule from .network import create_router from .network import delete_router from .network import list_routers from .network import get_router_id from .network import get_router_info from .network import get_router_ip from .network import update_router from .network import add_router_interface from .network import remove_router_interface from .network import list_floating_ips from .fw import list_firewall_rules from .fw import create_firewall_rule from .fw import create_firewall_policy from .fw import create_firewall from .lb import create_lb from .utils import create_logfile from .utils import gen_passwd from .vpn import create_ipsec_vpn_service from .vpn import list_ipsec_vpn_services from .vpn import get_ipsec_vpn_service_info from .vpn import get_ipsec_vpn_service_id from .vpn import update_ipsec_vpn_service from .vpn import delete_ipsec_vpn_service from .vpn import create_ipsec_policy from .vpn import list_ipsec_policies from .vpn import get_ipsec_policy_info from .vpn import get_ipsec_policy_id from .vpn import update_ipsec_policy from .vpn import delete_ipsec_policy from .vpn import create_ike_policy from .vpn import list_ike_policies from .vpn import get_ike_policy_info from .vpn import get_ike_policy_id from .vpn import update_ike_policy from .vpn import delete_ike_policy from .vpn import create_ipsec_vpn_connection from .vpn import list_ipsec_vpn_connections from .vpn import get_ipsec_vpn_connection_info from .vpn import get_ipsec_vpn_connection_id from .vpn import update_ipsec_vpn_connection from .vpn import delete_ipsec_vpn_connection from .vpn import create_ssl_vpn_service from .vpn import create_ssl_vpn_connection from .vpn import list_ssl_vpn_connections from .vpn import get_ssl_vpn_connection_id from .vpn import delete_ssl_vpn_connection from .key import create_key from .key import create_key_container from .key import list_keys from .key import list_key_containers ```
[ { "content": "Return the code exactly, with no changes:\n```python\n#!/usr/bin/env python\n# *********************************************************************\n# * Copyright (C) 2014 Luca Baldini (luca.baldini@pi.infn.it) *\n# * *\n# ...
[ { "content": "Return the code exactly, with no changes:\n<|memory_start|>```python\n#!/usr/bin/env python\n# *********************************************************************\n# * Copyright (C) 2014 Luca Baldini (luca.baldini@pi.infn.it) *\n# * ...
```python #!/usr/bin/env python # ********************************************************************* # * Copyright (C) 2014 Luca Baldini (luca.baldini@pi.infn.it) * # * * # * For the license terms see the file LICENSE, distributed * # * along with this software. * # ********************************************************************* # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. class E3GeometryRecord: """ Utility class encapsulating a geometry record. Note we only parse a few fields, for the time being. """ def __init__(self, *args): """ """ self.__Angle = args[0]/100. self.__Dist12 = float(args[1]) self.__Dist23 = float(args[2]) def angle(self): """ """ return self.__Angle def d12(self): """ """ return self.__Dist12 def d23(self): """ """ return self.__Dist23 def __str__(self): """ String formatting """ return 'Angle to north: %.1f deg, d12 = %.1f cm, d23 = %.1f cm.' %\ (self.angle(), self.d12(), self.d23()) ```
[ { "content": "```python\n# Copyright (c) 2015 VMware, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/l...
[ { "content": "<|memory_start|>```python\n# Copyright (c) 2015 VMware, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://...
```python # Copyright (c) 2015 VMware, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """ Exception definitions. """ from oslo_vmware import exceptions from cinder.i18n import _ class InvalidAdapterTypeException(exceptions.VMwareDriverException): """Thrown when the disk adapter type is invalid.""" msg_fmt = _("Invalid disk adapter type: %(invalid_type)s.") class InvalidDiskTypeException(exceptions.VMwareDriverException): """Thrown when the disk type is invalid.""" msg_fmt = _("Invalid disk type: %(disk_type)s.") class VirtualDiskNotFoundException(exceptions.VMwareDriverException): """Thrown when virtual disk is not found.""" msg_fmt = _("There is no virtual disk device.") class ProfileNotFoundException(exceptions.VMwareDriverException): """Thrown when the given storage profile cannot be found.""" msg_fmt = _("Storage profile: %(storage_profile)s not found.") class NoValidDatastoreException(exceptions.VMwareDriverException): """Thrown when there are no valid datastores.""" msg_fmt = _("There are no valid datastores.") class ClusterNotFoundException(exceptions.VMwareDriverException): """Thrown when the given cluster cannot be found.""" msg_fmt = _("Compute cluster: %(cluster)s not found.") class NoValidHostException(exceptions.VMwareDriverException): """Thrown when there are no valid ESX hosts.""" msg_fmt = _("There are no valid ESX hosts.") class TemplateNotFoundException(exceptions.VMwareDriverException): """Thrown when template cannot be found.""" msg_fmt = _("Template cannot be found at path: %(path)s.") ```
[ { "content": "Here is the source code:\n```python\n# -*- coding: utf-8 -*-\n#\n# Inqbus.installer documentation build configuration file, created by\n# sphinx-quickstart on Tue Feb 4 10:28:33 2014.\n#\n# This file is execfile()d with the current directory set to its containing dir.\n#\n# Note that not all poss...
[ { "content": "Here is the source code:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n#\n# Inqbus.installer documentation build configuration file, created by\n# sphinx-quickstart on Tue Feb 4 10:28:33 2014.\n#\n# This file is execfile()d with the current directory set to its containing dir.\n#\n# Note t...
```python # -*- coding: utf-8 -*- # # Inqbus.installer documentation build configuration file, created by # sphinx-quickstart on Tue Feb 4 10:28:33 2014. # # This file is execfile()d with the current directory set to its containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys, os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ----------------------------------------------------- # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be extensions # coming with Sphinx (named 'sphinx.ext.*') or your custom ones. extensions = [] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'Inqbus.installer' copyright = u'2014, Sandra Rum, Dr. Volker Jaenisch' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '1.0' # The full version, including alpha/beta/rc tags. release = '1.0' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # -- Options for HTML output --------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'default' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. #html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'Inqbusinstallerdoc' # -- Options for LaTeX output -------------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, author, documentclass [howto/manual]). latex_documents = [ ('index', 'Inqbusinstaller.tex', u'Inqbus.installer Documentation', u'Sandra Rum, Dr. Volker Jaenisch', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output -------------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'inqbusinstaller', u'Inqbus.installer Documentation', [u'Sandra Rum, Dr. Volker Jaenisch'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------------ # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'Inqbusinstaller', u'Inqbus.installer Documentation', u'Sandra Rum, Dr. Volker Jaenisch', 'Inqbusinstaller', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote' # -- Options for Epub output --------------------------------------------------- # Bibliographic Dublin Core info. epub_title = u'Inqbus.installer' epub_author = u'Sandra Rum, Dr. Volker Jaenisch' epub_publisher = u'Sandra Rum, Dr. Volker Jaenisch' epub_copyright = u'2014, Sandra Rum, Dr. Volker Jaenisch' # The language of the text. It defaults to the language option # or en if the language is not set. #epub_language = '' # The scheme of the identifier. Typical schemes are ISBN or URL. #epub_scheme = '' # The unique identifier of the text. This can be a ISBN number # or the project homepage. #epub_identifier = '' # A unique identification for the text. #epub_uid = '' # A tuple containing the cover image and cover page html template filenames. #epub_cover = () # HTML files that should be inserted before the pages created by sphinx. # The format is a list of tuples containing the path and title. #epub_pre_files = [] # HTML files shat should be inserted after the pages created by sphinx. # The format is a list of tuples containing the path and title. #epub_post_files = [] # A list of files that should not be packed into the epub file. #epub_exclude_files = [] # The depth of the table of contents in toc.ncx. #epub_tocdepth = 3 # Allow duplicate toc entries. #epub_tocdup = True ```
[ { "content": "Replicate the code snippet exactly, without paraphrasing or reformatting:\n```python\n##############################################################################\n#\n# OSIS stands for Open Student Information System. It's an application\n# designed to manage the core business of higher ed...
[ { "content": "Replicate the code snippet exactly, without paraphrasing or reformatting:\n<|memory_start|>```python\n##############################################################################\n#\n# OSIS stands for Open Student Information System. It's an application\n# designed to manage the core busin...
```python ############################################################################## # # OSIS stands for Open Student Information System. It's an application # designed to manage the core business of higher education institutions, # such as universities, faculties, institutes and professional schools. # The core business involves the administration of students, teachers, # courses, programs and so on. # # Copyright (C) 2015-2019 Université catholique de Louvain (http://www.uclouvain.be) # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # A copy of this license - GNU General Public License - is available # at the root of the source code of this program. If not, # see http://www.gnu.org/licenses/. # ############################################################################## from django.utils.translation import gettext_lazy as _ BASE = 'BASE' DISSERTATION = 'DISSERTATION' CHOICES = ( (BASE, _("Base")), (DISSERTATION, _("Dissertation")) ) ```
[ { "content": "```python\nfrom .core import ( # noqa\n AssetID,\n AssetIDPlusDay,\n EPOCH,\n ExplodingObject,\n FakeDataPortal,\n FetcherDataPortal,\n MockDailyBarReader,\n OpenPrice,\n add_security_data,\n all_pairs_matching_predicate,\n all_subindices,\n assert_single_posit...
[ { "content": "<|memory_start|>```python\nfrom .core import ( # noqa\n AssetID,\n AssetIDPlusDay,\n EPOCH,\n ExplodingObject,\n FakeDataPortal,\n FetcherDataPortal,\n MockDailyBarReader,\n OpenPrice,\n add_security_data,\n all_pairs_matching_predicate,\n all_subindices,\n ass...
```python from .core import ( # noqa AssetID, AssetIDPlusDay, EPOCH, ExplodingObject, FakeDataPortal, FetcherDataPortal, MockDailyBarReader, OpenPrice, add_security_data, all_pairs_matching_predicate, all_subindices, assert_single_position, assert_timestamp_equal, check_allclose, check_arrays, chrange, create_daily_df_for_asset, create_data_portal, create_data_portal_from_trade_history, create_empty_splits_mergers_frame, create_minute_bar_data, create_minute_df_for_asset, drain_zipline, empty_asset_finder, empty_assets_db, empty_trading_env, make_alternating_boolean_array, make_cascading_boolean_array, make_test_handler, make_trade_data_for_asset_info, parameter_space, patch_os_environment, patch_read_csv, permute_rows, powerset, product_upper_triangle, read_compressed, seconds_to_timestamp, security_list_copy, str_to_seconds, subtest, temp_pipeline_engine, test_resource_path, tmp_asset_finder, tmp_assets_db, tmp_bcolz_equity_minute_bar_reader, tmp_dir, tmp_trading_env, to_series, to_utc, trades_by_sid_to_dfs, write_bcolz_minute_data, write_compressed, ) from .fixtures import ZiplineTestCase # noqa ```
[ { "content": "Here is a code file:\n```python\n#!/usr/bin/env python\n\n'''\nLicensed to the Apache Software Foundation (ASF) under one\nor more contributor license agreements. See the NOTICE file\ndistributed with this work for additional information\nregarding copyright ownership. The ASF licenses this file...
[ { "content": "Here is a code file:\n<|memory_start|>```python\n#!/usr/bin/env python\n\n'''\nLicensed to the Apache Software Foundation (ASF) under one\nor more contributor license agreements. See the NOTICE file\ndistributed with this work for additional information\nregarding copyright ownership. The ASF li...
```python #!/usr/bin/env python ''' Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ''' import json from mock.mock import MagicMock, patch from stacks.utils.RMFTestCase import * from resource_management.core.exceptions import Fail @patch("os.path.isfile", new = MagicMock(return_value=True)) @patch("glob.glob", new = MagicMock(return_value=["one", "two"])) class TestWebHCatServer(RMFTestCase): COMMON_SERVICES_PACKAGE_DIR = "HIVE/0.12.0.2.0/package" STACK_VERSION = "2.0.6" def test_configure_default(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "configure", config_file="default.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assert_configure_default() self.assertNoMoreResources() def test_start_default(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "start", config_file="default.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assert_configure_default() self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/hdp/current/hive-webhcat/sbin/webhcat_server.sh start', environment = {'HADOOP_HOME': '/usr/hdp/current/hadoop-client'}, not_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", user = 'hcat', ) self.assertNoMoreResources() def test_stop_default(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "stop", config_file="default.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assertResourceCalled('Execute', '/usr/hdp/current/hive-webhcat/sbin/webhcat_server.sh stop', user = 'hcat', environment = {'HADOOP_HOME': '/usr/hdp/current/hadoop-client' } ) self.assertResourceCalled('Execute', 'ambari-sudo.sh kill -9 `cat /var/run/webhcat/webhcat.pid`', only_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", ignore_failures = True ) self.assertResourceCalled('Execute', "! (ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1)") self.assertResourceCalled('File', '/var/run/webhcat/webhcat.pid', action = ['delete'], ) self.assertNoMoreResources() def test_configure_secured(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "configure", config_file="secured.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assert_configure_secured() self.assertNoMoreResources() @patch("webhcat_service.graceful_stop", new = MagicMock(side_effect=Fail)) def test_stop_graceful_stop_failed(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "stop", config_file="default.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assertResourceCalled('Execute', "find /var/log/webhcat -maxdepth 1 -type f -name '*' -exec echo '==> {} <==' \\; -exec tail -n 40 {} \\;", logoutput = True, ignore_failures = True, user = 'hcat', ) self.assertResourceCalled('Execute', 'ambari-sudo.sh kill -9 `cat /var/run/webhcat/webhcat.pid`', only_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", ignore_failures = True ) self.assertResourceCalled('Execute', "! (ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1)") self.assertResourceCalled('File', '/var/run/webhcat/webhcat.pid', action = ['delete'], ) self.assertNoMoreResources() def test_start_secured(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "start", config_file="secured.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assert_configure_secured() self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/hdp/current/hive-webhcat/sbin/webhcat_server.sh start', environment = {'HADOOP_HOME': '/usr/hdp/current/hadoop-client'}, not_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", user = 'hcat', ) self.assertNoMoreResources() def test_stop_secured(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "stop", config_file="secured.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assertResourceCalled('Execute', '/usr/hdp/current/hive-webhcat/sbin/webhcat_server.sh stop', user = 'hcat', environment = {'HADOOP_HOME': '/usr/hdp/current/hadoop-client' } ) self.assertResourceCalled('Execute', 'ambari-sudo.sh kill -9 `cat /var/run/webhcat/webhcat.pid`', only_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", ignore_failures = True ) self.assertResourceCalled('Execute', "! (ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1)") self.assertResourceCalled('File', '/var/run/webhcat/webhcat.pid', action = ['delete'], ) self.assertNoMoreResources() @patch("webhcat_service.graceful_stop", new = MagicMock(side_effect=Fail)) def test_stop_secured_graceful_stop_failed(self): self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "stop", config_file="secured.json", stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES ) self.assertResourceCalled('Execute', "find /var/log/webhcat -maxdepth 1 -type f -name '*' -exec echo '==> {} <==' \\; -exec tail -n 40 {} \\;", logoutput = True, ignore_failures = True, user = 'hcat', ) self.assertResourceCalled('Execute', 'ambari-sudo.sh kill -9 `cat /var/run/webhcat/webhcat.pid`', only_if = "ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1", ignore_failures = True ) self.assertResourceCalled('Execute', "! (ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1)") self.assertResourceCalled('File', '/var/run/webhcat/webhcat.pid', action = ['delete'], ) self.assertNoMoreResources() def assert_configure_default(self): self.assertResourceCalled('Directory', '/var/run/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755, ) self.assertResourceCalled('Directory', '/var/log/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755, ) self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf', owner = 'hcat', group = 'hadoop', create_parents = True, cd_access = 'a' ) self.assertResourceCalled('XmlConfig', 'webhcat-site.xml', owner = 'hcat', group = 'hadoop', conf_dir = '/etc/hive-webhcat/conf', configurations = self.getConfig()['configurations']['webhcat-site'], configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site'] ) self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']), owner = 'hcat', group = 'hadoop', ) self.assertResourceCalled('Directory', '/usr/hdp/current/hive-webhcat/conf', cd_access = 'a', create_parents = True ) self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties', content = InlineTemplate('log4jproperties\nline2'), owner = 'hcat', group = 'hadoop', mode = 0644, ) def assert_configure_secured(self): self.assertResourceCalled('Directory', '/var/run/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755, ) self.assertResourceCalled('Directory', '/var/log/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755, ) self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf', owner = 'hcat', group = 'hadoop', create_parents = True, cd_access = 'a' ) self.assertResourceCalled('XmlConfig', 'webhcat-site.xml', owner = 'hcat', group = 'hadoop', conf_dir = '/etc/hive-webhcat/conf', configurations = self.getConfig()['configurations']['webhcat-site'], configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site'] ) self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']), owner = 'hcat', group = 'hadoop', ) self.assertResourceCalled('Directory', '/usr/hdp/current/hive-webhcat/conf', cd_access = 'a', create_parents = True ) self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties', content = InlineTemplate('log4jproperties\nline2'), owner = 'hcat', group = 'hadoop', mode = 0644, ) def test_pre_upgrade_restart(self): config_file = self.get_src_folder()+"/test/python/stacks/2.0.6/configs/default.json" with open(config_file, "r") as f: json_content = json.load(f) version = '2.2.1.0-3242' json_content['commandParams']['version'] = version self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "pre_upgrade_restart", config_dict = json_content, stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES) self.assertResourceCalled('Execute', ('ambari-python-wrap', '/usr/bin/hdp-select', 'set', 'hive-webhcat', version), sudo=True,) self.assertNoMoreResources() @patch("resource_management.core.shell.call") def test_pre_upgrade_restart_23(self, call_mock): import sys config_file = self.get_src_folder()+"/test/python/stacks/2.0.6/configs/default.json" with open(config_file, "r") as f: json_content = json.load(f) version = '2.3.0.0-1234' json_content['commandParams']['version'] = version json_content['hostLevelParams']['stack_version'] = "2.3" mocks_dict = {} self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "pre_upgrade_restart", config_dict = json_content, stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES, call_mocks = [(0, None, ''), (0, None, '')], mocks_dict = mocks_dict) self.assertTrue("params" in sys.modules) self.assertTrue(sys.modules["params"].webhcat_conf_dir is not None) self.assertTrue("/usr/hdp/current/hive-webhcat/etc/webhcat" == sys.modules["params"].webhcat_conf_dir) self.assertResourceCalledIgnoreEarlier('Execute', ('ambari-python-wrap', '/usr/bin/hdp-select', 'set', 'hive-webhcat', version), sudo=True,) self.assertNoMoreResources() self.assertEquals(2, mocks_dict['call'].call_count) self.assertEquals(2, mocks_dict['checked_call'].call_count) self.assertEquals( ('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hive-hcatalog', '--stack-version', '2.3.0.0-1234', '--conf-version', '0'), mocks_dict['checked_call'].call_args_list[0][0][0]) self.assertEquals( ('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hive-hcatalog', '--stack-version', '2.3.0.0-1234', '--conf-version', '0'), mocks_dict['call'].call_args_list[0][0][0]) self.assertEquals( ('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.3.0.0-1234', '--conf-version', '0'), mocks_dict['checked_call'].call_args_list[1][0][0]) self.assertEquals( ('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.3.0.0-1234', '--conf-version', '0'), mocks_dict['call'].call_args_list[1][0][0]) @patch("resource_management.core.shell.call") def test_rolling_restart_configure(self, call_mock): import sys config_file = self.get_src_folder()+"/test/python/stacks/2.0.6/configs/default.json" with open(config_file, "r") as f: json_content = json.load(f) version = '2.3.0.0-1234' json_content['commandParams']['version'] = version json_content['hostLevelParams']['stack_version'] = "2.3" mocks_dict = {} self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/webhcat_server.py", classname = "WebHCatServer", command = "configure", config_dict = json_content, stack_version = self.STACK_VERSION, target = RMFTestCase.TARGET_COMMON_SERVICES, call_mocks = [(0, None), (0, None)], mocks_dict = mocks_dict) self.assertResourceCalled('Directory', '/var/run/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755) self.assertResourceCalled('Directory', '/var/log/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, mode = 0755) self.assertResourceCalled('Directory', '/usr/hdp/current/hive-webhcat/etc/webhcat', owner = 'hcat', group = 'hadoop', create_parents = True, cd_access = 'a',) self.assertResourceCalled('XmlConfig', 'webhcat-site.xml', owner = 'hcat', group = 'hadoop', conf_dir = '/usr/hdp/current/hive-webhcat/etc/webhcat', configurations = self.getConfig()['configurations']['webhcat-site'], configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site']) self.assertResourceCalled('XmlConfig', 'hive-site.xml', owner = 'hive', group = 'hadoop', conf_dir = '/usr/hdp/2.3.0.0-1234/hive/conf', configuration_attributes = {u'final': {u'hive.optimize.bucketmapjoin.sortedmerge': u'true', u'javax.jdo.option.ConnectionDriverName': u'true', u'javax.jdo.option.ConnectionPassword': u'true'}}, configurations = self.getConfig()['configurations']['hive-site'], ) self.assertResourceCalled('XmlConfig', 'yarn-site.xml', owner = 'yarn', group = 'hadoop', conf_dir = '/usr/hdp/2.3.0.0-1234/hadoop/conf', configuration_attributes = {u'final': {u'yarn.nodemanager.container-executor.class': u'true', u'yarn.nodemanager.disk-health-checker.min-healthy-disks': u'true', u'yarn.nodemanager.local-dirs': u'true'}}, configurations = self.getConfig()['configurations']['yarn-site'], ) self.assertResourceCalled('File', '/usr/hdp/current/hive-webhcat/etc/webhcat/webhcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']), owner = 'hcat', group = 'hadoop') self.assertResourceCalled('Directory', '/usr/hdp/current/hive-webhcat/etc/webhcat', cd_access = 'a', create_parents = True) self.assertResourceCalled('File', '/usr/hdp/current/hive-webhcat/etc/webhcat/webhcat-log4j.properties', content = InlineTemplate('log4jproperties\nline2'), owner = 'hcat', group = 'hadoop', mode = 0644) self.assertNoMoreResources() ```
[ { "content": "Repeat the code exactly as the original, including blank lines:\n```python\n#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n##################################################\n# GNU Radio Python Flow Graph\n# Title: Amrx\n# Generated: Tue Aug 8 20:51:18 2017\n####################################...
[ { "content": "Repeat the code exactly as the original, including blank lines:\n<|memory_start|>```python\n#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n##################################################\n# GNU Radio Python Flow Graph\n# Title: Amrx\n# Generated: Tue Aug 8 20:51:18 2017\n####################...
```python #!/usr/bin/env python2 # -*- coding: utf-8 -*- ################################################## # GNU Radio Python Flow Graph # Title: Amrx # Generated: Tue Aug 8 20:51:18 2017 ################################################## from gnuradio import analog from gnuradio import blocks from gnuradio import eng_notation from gnuradio import filter from gnuradio import gr from gnuradio import uhd from gnuradio.eng_option import eng_option from gnuradio.filter import firdes from optparse import OptionParser import time, sys class AMrx(gr.top_block): def __init__(self): gr.top_block.__init__(self, "Amrx") ################################################## # Variables ################################################## self.samp_rate = samp_rate = 2500000 ################################################## # Blocks ################################################## self.uhd_usrp_source_0 = uhd.usrp_source( ",".join(("", "")), uhd.stream_args( cpu_format="fc32", channels=range(1), ), ) self.uhd_usrp_source_0.set_samp_rate(samp_rate) self.uhd_usrp_source_0.set_center_freq(435000000, 0) self.uhd_usrp_source_0.set_gain(80, 0) self.uhd_usrp_source_0.set_antenna('TX/RX', 0) self.uhd_usrp_source_0.set_bandwidth(100000, 0) self.rational_resampler_xxx_0 = filter.rational_resampler_ccc( interpolation=44100, decimation=2500000, taps=None, fractional_bw=None, ) self.blocks_wavfile_sink_0 = blocks.wavfile_sink(sys.argv[1], 1, 44100, 8) self.analog_am_demod_cf_0 = analog.am_demod_cf( channel_rate=44100, audio_decim=1, audio_pass=20000, audio_stop=21000, ) self.analog_agc2_xx_0 = analog.agc2_cc(.1, 1e-6, 1.0, 0) self.analog_agc2_xx_0.set_max_gain(5) ################################################## # Connections ################################################## self.connect((self.analog_agc2_xx_0, 0), (self.rational_resampler_xxx_0, 0)) self.connect((self.analog_am_demod_cf_0, 0), (self.blocks_wavfile_sink_0, 0)) self.connect((self.rational_resampler_xxx_0, 0), (self.analog_am_demod_cf_0, 0)) self.connect((self.uhd_usrp_source_0, 0), (self.analog_agc2_xx_0, 0)) def get_samp_rate(self): return self.samp_rate def set_samp_rate(self, samp_rate): self.samp_rate = samp_rate self.uhd_usrp_source_0.set_samp_rate(self.samp_rate) def main(top_block_cls=AMrx, options=None): tb = top_block_cls() tb.start() print('Receiving on ' + str(tb.uhd_usrp_source_0.get_center_freq()) + 'Hz with a channel bandwidth of ' + str(tb.uhd_usrp_source_0.get_bandwidth()) + 'Hz') try: raw_input('Press Enter to quit: ') except EOFError: pass tb.stop() tb.wait() print('.wav file generated') if __name__ == '__main__': main() ```
[ { "content": "Reproduce the code exactly as provided (keep formatting):\n```python\n# ex:ts=4:sw=4:sts=4:et\n# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-\n\nr\"\"\"\nThe base class for QuickRelease steps, along with error reporting classes, \ncustom subclasses of steps, and classes to run th...
[ { "content": "Reproduce the code exactly as provided (keep formatting):\n<|memory_start|>```python\n# ex:ts=4:sw=4:sts=4:et\n# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-\n\nr\"\"\"\nThe base class for QuickRelease steps, along with error reporting classes, \ncustom subclasses of steps, and c...
```python # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- r""" The base class for QuickRelease steps, along with error reporting classes, custom subclasses of steps, and classes to run the individual parts of the steps. L{Step}s are usually searched for in the C{quickrelease/steps} directory. This behavior can be modified by setting the B{C{QUICKRELEASE_DEFINITIONS_PATH}} to the name of a directory containing both a "processes" and "steps" directory. To exclude the inclusion of any L{Process}es or L{Step<quickrelease.step.Step>}s in the standard QuickRelease directories, set B{C{QUICKRELEASE_OVERRIDE_DEFAULT_DEFINITIONS}} in the environment. """ import os from quickrelease.exception import ReleaseFrameworkError, ReleaseFrameworkErrorCollection from quickrelease.utils import GetActivePartners, PrintReleaseFrameworkError class StepError(ReleaseFrameworkError): """ An exception subclassed from L{ReleaseFrameworkError<quickrelease.exception.ReleaseFrameworkError>} which provides a more useful error message about the specific L{Step} the error occured in. """ def __init__(self, stepObj, errStr, *args, **kwargs): ReleaseFrameworkError.__init__(self, errStr, stepObj) assert isinstance(stepObj, Step), "StepErrors require a Step object" self._partnerStr = "" if isinstance(stepObj, PartnerStep): self._partnerStr = " (partner: %s)" % (stepObj.activePartner) def _GetErroredStep(self): return self.details erroredStep = property(_GetErroredStep) def __str__(self): return "Error in step %s%s: %s" % (self.erroredStep, self._partnerStr, ReleaseFrameworkError.__str__(self)) class _StandardStepRunner(object): def __init__(self, *args, **kwargs): object.__init__(self) def DoPreflight(self, stepObj): stepObj.Preflight() def DoExecute(self, stepObj): stepObj.Execute() def DoVerify(self, stepObj): stepObj.Verify() def DoNotify(self, stepObj): stepObj.Notify() class Step(object): """ An object representing a single step of a defined L{Process<quickrelease.process.Process>} """ def __init__(self, *args, **kwargs): """ Construct a L{Step} object. @param process: The parent-process this L{Step} belongs to. @type process: L{Process<quickrelease.process.Process>} @param runner: The L{Step}-runner to use for this L{Step}. This allows different types of L{Step}'s to modify the logic of what it means to "run" a step if they so choose (e.g. a L{PartnerStep}). @type runner: object """ object.__init__(self) self._parentProcess = None self._runner = _StandardStepRunner() if kwargs.has_key('process'): self._parentProcess = kwargs['process'] if kwargs.has_key('runner'): self._runner = kwargs['runner'] def __str__(self): """The L{Step}'s name.""" return self.__class__.__name__ def _GetName(self): return str(self) def _GetRunner(self): return self._runner def _GetParentProcess(self): return self._parentProcess def _GetConfig(self): if self.process is None: raise self.SimpleStepError("%s has no associated process to " "obtain a ConfigSpec." % (self)) elif self.process.config is None: raise self.SimpleStepError("Process %s has no associated " "ConfigSpec." % (self.process)) else: return self.process.config def _GetLogger(self): if self.process is None: raise self.SimpleStepError("%s has no associated process to " "obtain a Logger." % (self)) elif self.process.logger is None: raise self.SimpleStepError("Process %s has no associated Logger." % (self.process)) else: return self.process.logger def Log(self, msg): return self.logger.Log(msg, step=self) def LogErr(self, msg): return self.logger.LogErr(msg, step=self) def LogDebug(self, msg): return self.logger.LogDebug(msg, step=self) def ShellCommandLog(self, combined=True): # TODO: handle to a named file for a log command pass name = property(_GetName) runner = property(_GetRunner) """Return the runner object responsible for running the parts of the step. Read-only. @type: Runner object""" config = property(_GetConfig) """The config associated with the L{Step}'s parent process, if any. Read-only. @type: L{ConfigSpec<quickrelease.config.ConfigSpec>} or C{None}.""" process = property(_GetParentProcess) """The process this step is a part of, if any. Read-only. @type: L{Process<quickrelease.process.Process>} or C{None}""" logger = property(_GetLogger) """The logger associated with the L{Step}'s parent process, if any. Read-only. @type: L{Logger<quickrelease.logger.Logger>} or C{None}.""" def Preflight(self): """ A method intended for L{Step}s to override with any activities which must be executed before I{either} the L{Execute<quickrelease.step.Step.Execute>} or L{Verify<quickrelease.step.Step.Verify>} methods need to be executed, if any such activities exist. """ pass def Execute(self): """ A method intended for dervied L{Step}s to override with the execution logic of the particular L{Process} step. B{Note}: This method B{must} be redefined by the dervied L{Step}. @raise NotImplementedError: If the derived L{Step} does not define an C{Execute} method. """ raise NotImplementedError("Need implementation for %s::Execute()" % (str(self))) def Verify(self): """ A method intended for dervied L{Step}s to override with the unit test-like verification logic of the particular L{Process} step. B{Note}: This method B{must} be redefined by the dervied L{Step}. @raise NotImplementedError: If the derived L{Step} does not define an C{Verify} method. """ raise NotImplementedError("Need implementation for %s::Verify()" % (str(self))) def Notify(self): """ A method intended for L{Step}s to override with any notifications that should occur after a step has successful been executed and/or verified. B{Note}: Currently, these notifications will fire even if only the verification-steps are run. """ pass # We're kinda cheating here; when using it, it looks like SimpleStepError # is an exception type, not a function; it's mostly a convenience function # for creating a StepError Exception with a simple message, so we don't # have to pass the step object the StepError expects explicitly. def SimpleStepError(self, errStr, details=None): """ A convenience method for creating a L{StepError} with a simple message, so users don't have to pass the L{Step} object the L{StepError} expects explicitly. @param errStr: the error string @type errStr: C{str} @param details: Extra details about the error condition. @type details: Variable @return: An initialized L{StepError} with the current step associated to it. @rtype: L{StepError} """ return StepError(self, errStr, details=details) class _PartnerStepRunner(object): def __init__(self, *args, **kwargs): object.__init__(self) def _RunPartnerStepMethod(self, stepObj, methodName): conf = stepObj.config rootDir = conf.rootDir errors = [] for p in GetActivePartners(conf): try: os.chdir(rootDir) stepObj.activePartner = p stepMethod = getattr(stepObj, methodName) stepMethod() except ReleaseFrameworkError, ex: if stepObj._haltOnFirstError: raise ex else: # Unless we're in quiet mode... PrintReleaseFrameworkError(ex) errors.append(ex) if len(errors) != 0: raise ReleaseFrameworkErrorCollection(errors) def DoPreflight(self, stepObj): self._RunPartnerStepMethod(stepObj, "Preflight") def DoExecute(self, stepObj): self._RunPartnerStepMethod(stepObj, "Execute") def DoVerify(self, stepObj): self._RunPartnerStepMethod(stepObj, "Verify") def DoNotify(self, stepObj): self._RunPartnerStepMethod(stepObj, "Notify") class PartnerStep(Step): """ A special type of L{Step} which will perform the requested C{Execute} and C{Verify} methods for all active partners (as determined by L{GetActivePartners<quickrelease.utils.GetActivePartners>}). Subclasses may call the the constructor of C{PartnerStep} with the following keywords to modify its behavior: 1. C{auto_set_partner_config}: By default, when the L{PartnerStep} sets the next partner to execute or verify the portion of the current step, it will also set the section of the associated L{ConfigSpec<quickrelease.config.ConfigSpec>} to the active partner section (via a call to L{SetPartnerSection<quickrelease.config.ConfigSpec.SetPartnerSection>}. Setting this to C{False} will disable that bahavior and make the subclassed L{PartnerStep}s responsible for managing the state of their L{ConfigSpec<quickrelease.config.ConfigSpec>}. 2. C{halt_on_first_error}: By default, if an error is encountered during the execution or verification portion of a L{PartnerStep}, the error will be reported and noted, but the L{Step} will continue for each active partner. Once each active partner's step has been called, I{then} the L{PartnerStep} will halt. For example, say there exist two partners, "Acme" and "Biffco" and a three-step process, consisting of L{PartnerStep}s named C{WillBeOK}, C{WillFailForAcme}, and C{WillNotRun}. By default, C{WillBeOK} will run for Acme and Biffco; C{WillFailForAcme} will run for Acme and fail, and will then run for Biffco, and succeed. At this point, the L{PartnerStep} will halt with the errors, and the last step will not run. If this is set to C{True}, the L{PartnerStep} would immediately halt when it encountered the Acme-error. """ def __init__(self, *args, **kwargs): Step.__init__(self, *args, **kwargs) self._runner = _PartnerStepRunner() self._activePartner = None self._partnerData = {} self._autoSetPartnerConfig = True self._haltOnFirstError = False if kwargs.has_key('auto_set_partner_config'): self._autoSetPartnerConfig = kwargs['auto_set_partner_config'] if kwargs.has_key('halt_on_first_error'): self._haltOnFirstError = kwargs['halt_on_first_error'] def _GetActivePartner(self): return self._activePartner def _SetActivePartner(self, partner): if partner not in GetActivePartners(self.config): raise self.SimpleStepError("Unknown partner '%s'" % (partner)) self._activePartner = partner if self._autoSetPartnerConfig: self.config.SetPartnerSection(partner) if partner not in self._partnerData.keys(): self._partnerData[partner] = {} activePartner = property(_GetActivePartner, _SetActivePartner) def Save(self, key, data): """ Store partner-specific data that may need to persist across a set of C{Execute}/C{Verify} calls. @param key: Key to retrieve the data. @type key: C{str} @param data: The data to store. @type data: Variable """ self._partnerData[self.activePartner][key] = data def Load(self, key): """ Retrieve partner-specific data that may need to persist across a set of C{Execute}/C{Verify} calls. @param key: Key of data to retrieve @type key: C{str} @raise KeyError: If the data described by the specified key does not exist. """ return self._partnerData[self.activePartner][key] ```
[ { "content": "```python\n'''\n2DPCA for feature extraction of MNIST digits dataset \nAuthor : Akhil P M\n'''\n\n\nfrom settings import *\nfrom sklearn.ensemble import RandomForestClassifier\nimport utils\n\n\ndef compute_covariance_matrix(A):\n\t\"\"\" compute the 2D covariance matrix in image space\"\"\"\n\n\t...
[ { "content": "<|memory_start|>```python\n'''\n2DPCA for feature extraction of MNIST digits dataset \nAuthor : Akhil P M\n'''\n\n\nfrom settings import *\nfrom sklearn.ensemble import RandomForestClassifier\nimport utils\n\n\ndef compute_covariance_matrix(A):\n\t\"\"\" compute the 2D covariance matrix in image s...
```python ''' 2DPCA for feature extraction of MNIST digits dataset Author : Akhil P M ''' from settings import * from sklearn.ensemble import RandomForestClassifier import utils def compute_covariance_matrix(A): """ compute the 2D covariance matrix in image space""" no_of_images = len(A) cov = np.zeros((A.shape[2], A.shape[2])) for i in xrange(no_of_images): cov = cov + np.dot(np.transpose(A[i]), A[i]) cov = cov / no_of_images return cov def extract_feature(A, x): """ compute y[i] = A[i]*x for all images """ no_of_images = len(A) features = np.zeros((no_of_images, A.shape[1])) for i in xrange(no_of_images): features[i] = np.ravel(np.dot(A[i], x)) return features def main(): """ the main function""" #set the timer start = time.time() #load the data trainX = np.load('trainX.npy') testX = np.load('testX.npy') trainY = np.load('trainY.npy') testY = np.load('testY.npy') print('\n!!! Data Loading Completed !!!\n') #generate 2D data data_train = utils.generate_2D(trainX) data_test = utils.generate_2D(testX) ncol = data_train.shape[2] features_train = np.zeros((len(data_train), data_train.shape[1])) features_test = np.zeros((len(data_test), data_test.shape[1])) #get the mean image mean_image = utils.get_mean_image(data_train) #substract the mean image from all images & center them normalized_data = utils.substract_mean(data_train, mean_image) data_train = utils.substract_mean(data_train, mean_image) data_test = utils.substract_mean(data_test, mean_image) #compute the covariance matrix in 2D space SA = compute_covariance_matrix(normalized_data) #find eigen values & eigen vectors of covariance matrix U, s, _ = np.linalg.svd(SA) #extract features using 2DPCA selected = [] clf = RandomForestClassifier(n_estimators=300, n_jobs=-1) max_acc = 0.0 for i in xrange(ncol): proj_dir = U[:, i].reshape(ncol, 1) tempTrainX = extract_feature(data_train, proj_dir) tempTestX = extract_feature(data_test, proj_dir) clf.fit(tempTrainX, trainY) pred = clf.predict(tempTestX) acc = accuracy_score(testY, pred) print('PC vector %d gives accuracy : %f\n' %(i+1, acc)) #if acc >=0.1: # selected.append(i) # features_train = features_train + s[i] * tempTrainX # features_test = features_test + s[i] * tempTestX if acc > max_acc: max_acc = acc features_train = np.copy(tempTrainX) features_test = np.copy(tempTestX) print features_train.shape np.save('trainX_feat', features_train) np.save('testX_feat', features_test) clf.fit(features_train, trainY) pred = clf.predict(features_test) print('accuracy : %f\n' %accuracy_score(testY, pred)) #print selected print('Test Time : %f Minutes\n' %((time.time()-start)/60)) if __name__ == '__main__': main() ```
[ { "content": "```python\n#!/usr/bin/env python\n#-*- coding:utf-8 -*-\n\"\"\"\nThis module provide configure file management service in i18n environment.\n\"\"\"\nimport os\nimport logging\nimport logging.handlers\n\n\n_LOG_FORMAT = \"%(levelname)s: %(asctime)s: %(filename)s:%(lineno)d * %(thread)d %(message)s\...
[ { "content": "<|memory_start|>```python\n#!/usr/bin/env python\n#-*- coding:utf-8 -*-\n\"\"\"\nThis module provide configure file management service in i18n environment.\n\"\"\"\nimport os\nimport logging\nimport logging.handlers\n\n\n_LOG_FORMAT = \"%(levelname)s: %(asctime)s: %(filename)s:%(lineno)d * %(threa...
```python #!/usr/bin/env python #-*- coding:utf-8 -*- """ This module provide configure file management service in i18n environment. """ import os import logging import logging.handlers _LOG_FORMAT = "%(levelname)s: %(asctime)s: %(filename)s:%(lineno)d * %(thread)d %(message)s" _LOG_DATEFMT = "%m-%d %H:%M:%S" def init_log(log_path, level=logging.INFO, when="D", backup=7, format=_LOG_FORMAT, datefmt=_LOG_DATEFMT): """ init_log - initialize log module Args: log_path - Log file path prefix. Log data will go to two files: log_path.log and log_path.log.wf Any non-exist parent directories will be created automatically level - msg above the level will be displayed DEBUG < INFO < WARNING < ERROR < CRITICAL the default value is logging.INFO when - how to split the log file by time interval 'S' : Seconds 'M' : Minutes 'H' : Hours 'D' : Days 'W' : Week day default value: 'D' format - format of the log default format: %(levelname)s: %(asctime)s: %(filename)s:%(lineno)d * %(thread)d %(message)s INFO: 12-09 18:02:42: log.py:40 * 139814749787872 HELLO WORLD backup - how many backup file to keep default value: 7 Raises: OSError: fail to create log directories IOError: fail to open log file """ formatter = logging.Formatter(format, datefmt) logger = logging.getLogger() logger.setLevel(level) dir = os.path.dirname(log_path) if not os.path.isdir(dir): os.makedirs(dir) handler = logging.handlers.TimedRotatingFileHandler(log_path + ".log", when=when, backupCount=backup) handler.setLevel(level) handler.setFormatter(formatter) logger.addHandler(handler) handler = logging.handlers.TimedRotatingFileHandler(log_path + ".log.wf", when=when, backupCount=backup) handler.setLevel(logging.WARNING) handler.setFormatter(formatter) logger.addHandler(handler) if __name__ == '__main__': init_log('./log') ```
[ { "content": "Here is the code content:\n```python\nfrom setuptools import setup\n\nsetup(\n name='pureples',\n version='0.0',\n author='adrian, simon',\n author_email='mail@adrianwesth.dk',\n maintainer='simon, adrian',\n maintainer_email='mail@adrianwesth.dk',\n url='https://github.com/uk...
[ { "content": "Here is the code content:\n<|memory_start|>```python\nfrom setuptools import setup\n\nsetup(\n name='pureples',\n version='0.0',\n author='adrian, simon',\n author_email='mail@adrianwesth.dk',\n maintainer='simon, adrian',\n maintainer_email='mail@adrianwesth.dk',\n url='https...
```python from setuptools import setup setup( name='pureples', version='0.0', author='adrian, simon', author_email='mail@adrianwesth.dk', maintainer='simon, adrian', maintainer_email='mail@adrianwesth.dk', url='https://github.com/ukuleleplayer/pureples', license="MIT", description='HyperNEAT and ES-HyperNEAT implemented in pure Python', long_description='Python implementation of HyperNEAT and ES-HyperNEAT ' + 'developed by Adrian Westh and Simon Krabbe Munck for evolving arbitrary neural networks. ' + 'HyperNEAT and ES-HyperNEAT is originally developed by Kenneth O. Stanley and Sebastian Risi', packages=['pureples', 'pureples/hyperneat', 'pureples/es_hyperneat', 'pureples/shared'], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Education', 'Intended Audience :: Science/Research', 'License :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3.x', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Scientific/Engineering' ], install_requires=['numpy', 'neat-python', 'graphviz', 'matplotlib', 'gym'] ) ```
[ { "content": "Output the full code verbatim (no extra comments):\n```python\n# Copyright (C) 2011 by David Tomaschik <david@systemoverlord.com>\n# \n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n...
[ { "content": "Output the full code verbatim (no extra comments):\n<|memory_start|>```python\n# Copyright (C) 2011 by David Tomaschik <david@systemoverlord.com>\n# \n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Softwa...
```python # Copyright (C) 2011 by David Tomaschik <david@systemoverlord.com> # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN # THE SOFTWARE. import sys from core import config def print_verbose(msg): if config.config.get('verbose',False): print_error(msg) def print_error(msg): sys.stderr.write(str(msg)) sys.stderr.write('\n') sys.stderr.flush() # vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4 ```
[ { "content": "Repeat the code exactly as the original, including blank lines:\n```python\nfrom django.conf.urls import url\n\nfrom rest_framework import routers, viewsets\n\nfrom rest_framework_nested import routers as nested_routers\n\n\nclass HybridRoutingMixin(object):\n \"\"\"\n Extends functionality ...
[ { "content": "Repeat the code exactly as the original, including blank lines:\n<|memory_start|>```python\nfrom django.conf.urls import url\n\nfrom rest_framework import routers, viewsets\n\nfrom rest_framework_nested import routers as nested_routers\n\n\nclass HybridRoutingMixin(object):\n \"\"\"\n Extend...
```python from django.conf.urls import url from rest_framework import routers, viewsets from rest_framework_nested import routers as nested_routers class HybridRoutingMixin(object): """ Extends functionality of DefaultRouter adding possibility to register simple API views, not just Viewsets. Based on: http://stackoverflow.com/questions/18818179/routing-api-views-in-django-rest-framework http://stackoverflow.com/questions/18817988/using-django-rest-frameworks-browsable-api-with-apiviews """ def get_routes(self, viewset): """ Checks if the viewset is an instance of ViewSet, otherwise assumes it's a simple view and does not run original `get_routes` code. """ if issubclass(viewset, viewsets.ViewSetMixin): return super(HybridRoutingMixin, self).get_routes(viewset) return [] def get_urls(self): """ Append non-viewset views to the urls generated by the original `get_urls` method. """ # URLs for viewsets ret = super(HybridRoutingMixin, self).get_urls() # URLs for simple views for prefix, viewset, basename in self.registry: # Skip viewsets if issubclass(viewset, viewsets.ViewSetMixin): continue # URL regex regex = '{prefix}{trailing_slash}$'.format( prefix=prefix, trailing_slash=self.trailing_slash ) # The view name has to have suffix "-list" due to specifics # of the DefaultRouter implementation. ret.append(url(regex, viewset.as_view(), name='{0}-list'.format(basename))) return ret class HybridDefaultRouter(HybridRoutingMixin, routers.DefaultRouter): pass class HybridSimpleRouter(HybridRoutingMixin, routers.SimpleRouter): pass class HybridNestedRouter(HybridRoutingMixin, nested_routers.NestedSimpleRouter): pass ```
[ { "content": "Here is a code snippet:\n```python\n# encoding: utf-8\n\nimport torch\nimport torch.autograd as autograd\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\nclass Softmax(nn.Module):\n\n def __init__(self, num_labels, feature_dim):\n super(Softmax, sel...
[ { "content": "Here is a code snippet:\n<|memory_start|>```python\n# encoding: utf-8\n\nimport torch\nimport torch.autograd as autograd\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\nclass Softmax(nn.Module):\n\n def __init__(self, num_labels, feature_dim):\n su...
```python # encoding: utf-8 import torch import torch.autograd as autograd import torch.nn as nn import torch.nn.functional as F import torch.optim as optim class Softmax(nn.Module): def __init__(self, num_labels, feature_dim): super(Softmax, self).__init__() self.linear = nn.Linear(feature_dim, num_labels) def forward(self, x): p = F.softmax(self.linear(x)) log_p = F.log_softmax(self.linear(x)) return p, log_p data = [("me gusta comer en la cafeteria".split(), "SPANISH"), ("Give it to me".split(), "ENGLISH"), ("No creo que sea una buena idea".split(), "SPANISH"), ("No it is not a good idea to get lost at sea".split(), "ENGLISH")] test_data = [("Yo creo que si".split(), "SPANISH"), ("it is lost on me".split(), "ENGLISH")] word_to_ix = {} for sent, _ in data + test_data: for word in sent: if word not in word_to_ix: word_to_ix[word] = len(word_to_ix) print(word_to_ix) label_to_ix = {"SPANISH": 0, "ENGLISH": 1} VOCAB_SIZE = len(word_to_ix) NUM_LABELS = 2 def make_bow_vector(sentence, word_to_ix): vec = torch.zeros(len(word_to_ix)) for word in sentence: vec[word_to_ix[word]] += 1 return vec.view(1, -1) def make_target(label, label_to_ix): return torch.LongTensor([label_to_ix[label]]) model = Softmax(NUM_LABELS, VOCAB_SIZE) loss_function = nn.NLLLoss() optimizer = optim.SGD(model.parameters(), lr=0.1) for epoch in range(100): for instance, label in data: # Step 1. Remember that Pytorch accumulates gradients. # We need to clear them out before each instance model.zero_grad() # Step 2. Make our BOW vector and also we must wrap the target in a # Variable as an integer. For example, if the target is SPANISH, then # we wrap the integer 0. The loss function then knows that the 0th # element of the log probabilities is the log probability # corresponding to SPANISH bow_vec = autograd.Variable(make_bow_vector(instance, word_to_ix)) target = autograd.Variable(make_target(label, label_to_ix)) # Step 3. Run our forward pass. _, log_probs = model(bow_vec) # Step 4. Compute the loss, gradients, and update the parameters by # calling optimizer.step() loss = loss_function(log_probs, target) loss.backward() optimizer.step() for instance, label in test_data: bow_vec = autograd.Variable(make_bow_vector(instance, word_to_ix)) probs, log_probs = model(bow_vec) print(probs) # Index corresponding to Spanish goes up, English goes down! print(next(model.parameters())[:, word_to_ix["creo"]]) ```
[ { "content": "Provide a verbatim copy of the code:\n```python\n# TODO: add Group and/or Selection\n\nclass Region(object):\n \"\"\" Base class for a spatial Region container\n\n A Region can contain simple and advanced geometry.\n\n A Region can be thought of as a 3D analogy to a sheet of\n paper in...
[ { "content": "Provide a verbatim copy of the code:\n<|memory_start|>```python\n# TODO: add Group and/or Selection\n\nclass Region(object):\n \"\"\" Base class for a spatial Region container\n\n A Region can contain simple and advanced geometry.\n\n A Region can be thought of as a 3D analogy to a sheet ...
```python # TODO: add Group and/or Selection class Region(object): """ Base class for a spatial Region container A Region can contain simple and advanced geometry. A Region can be thought of as a 3D analogy to a sheet of paper in 2D. A Region defines its own coordinate system, dimension and resolution. """ def __init__(self, **attr): """ Initialize a region Parameters ---------- dimension ; array-like, integer The number of units in each spatial direction e.g. [1000,2000,400] resolution : array-like, double The resolution in each spatial direction e.g. [2.5, 3.0, 10.23] resolution_unit : array-like The unit of resolution in each spatial direction e.g. ['nm', 'nm', 'nm'] origo : array-like The locus of the origo of the coordinate system XXX: in relation to some global coordinate system (?) this could be implied in the affine axes_orientation : array-like The orthogonal orientation of the x-, y- and z-axes of a cartesian coordinate system as a 3x3 array. coordinate_system : ['left-handed', 'right-handed'] (default='left-handed') affine : array-like origo and axes_orientation can be grouped in a 4x4 affine array. TODO: correct? extension : ['bounded', 'unbounded'] Defining a dimension implies a bounded Region name : string, optional (default='') An optional name for the Region. attr : keyword arguments, optional (default=no attributes) Attributes to add to a graph as key-value pairs. """ pass def __str__(self): """Return the Region name. Returns ------- name : string The name of the Region. """ return self.name ```
[ { "content": "Repeat the code precisely as written (spacing intact):\n```python\n__author__ = 'yinjun'\n\n# Definition for singly-linked list.\n# class ListNode:\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution:\n # @param l1: the first list\n # @param l2...
[ { "content": "Repeat the code precisely as written (spacing intact):\n<|memory_start|>```python\n__author__ = 'yinjun'\n\n# Definition for singly-linked list.\n# class ListNode:\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution:\n # @param l1: the first list\...
```python __author__ = 'yinjun' # Definition for singly-linked list. # class ListNode: # def __init__(self, x): # self.val = x # self.next = None class Solution: # @param l1: the first list # @param l2: the second list # @return: the sum list of l1 and l2 def addLists(self, l1, l2): # write your code here h = ListNode(0) l = h add = 0 while l1!=None and l2!=None: l.next = ListNode(l1.val + l2.val + add) if l.next.val >= 10: add = 1 l.next.val -=10 else: add = 0 l = l.next l1 = l1.next l2 = l2.next while l1 != None: l.next = ListNode(l1.val + add) if l.next.val >= 10: add = 1 l.next.val -= 10 else: add = 0 l = l.next l1 = l1.next while l2 != None: l.next = ListNode(l2.val + add) if l.next.val >= 10: add = 1 l.next.val -= 10 else: add = 0 l = l.next l2 = l2.next if add > 0: l.next = ListNode(add) return h.next ```
[ { "content": "```python\n# Copyright 2021 The Brax Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless req...
[ { "content": "<|memory_start|>```python\n# Copyright 2021 The Brax Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\...
```python # Copyright 2021 The Brax Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for Mujoco converter.""" from absl.testing import absltest from brax.tools import mujoco _TEST_XML = """ <mujoco model="mjmodel"> <compiler angle="degree" coordinate="local" inertiafromgeom="true"/> <option integrator="RK4" timestep="0.01"/> <default> <joint armature="1" damping="1" limited="true"/> <geom conaffinity="0" condim="3" density="5.0" friction="1 0.5 0.5" margin="0.01" rgba="0.8 0.6 0.4 1"/> </default> <worldbody> <light cutoff="100" diffuse="1 1 1" dir="-0 0 -1.3" directional="true" exponent="1" pos="0 0 1.3" specular=".1 .1 .1"/> <geom conaffinity="1" condim="3" material="MatPlane" name="floor" pos="0 0 0" rgba="0.8 0.9 0.8 1" size="40 40 40" type="plane"/> <body name="parent" pos="0 0 0.75"> <camera name="track" mode="trackcom" pos="0 -3 0.3" xyaxes="1 0 0 0 0 1"/> <geom name="parent_geom" pos="0 0 0" size="0.25" type="sphere"/> <joint armature="0" damping="0" limited="false" margin="0.01" name="root" pos="0 0 0" type="free"/> <body name="child_1" pos="0 0 0"> <geom fromto="0.0 0.0 0.0 0.2 0.2 0.0" name="aux_1_geom" size="0.08" type="capsule"/> <body name="child_2" pos="0.2 0.2 0"> <joint axis="0 0 1" name="child_2_joint" pos="0.0 0.0 0.0" range="-30 30" type="hinge"/> <geom fromto="0.0 0.0 0.0 0.2 0.2 0.0" name="child_2_geom" size="0.08" type="capsule"/> <body pos="0.2 0.2 0"> <joint axis="-1 1 0" name="anon_joint" pos="0.0 0.0 0.0" range="30 70" type="hinge"/> <geom fromto="0.0 0.0 0.0 0.4 0.4 0.0" name="left_ankle_geom" size="0.08" type="capsule"/> </body> </body> </body> </body> </worldbody> <actuator> <motor ctrllimited="true" ctrlrange="-1.0 1.0" joint="child_2_joint" gear="150"/> <motor ctrllimited="true" ctrlrange="-1.0 1.0" joint="anon_joint" gear="150"/> </actuator> </mujoco> """ class MujocoTest(absltest.TestCase): def test_build(self): m = mujoco.MujocoConverter(_TEST_XML, add_collision_pairs=True) # Sanity check. config = m.config self.assertTrue(config.bodies) self.assertTrue(config.joints) self.assertTrue(config.actuators) self.assertTrue(config.collide_include) if __name__ == '__main__': absltest.main() ```
[ { "content": "```python\nimport django.core.handlers.wsgi\nimport logging\nfrom django.template import loader,Context,RequestContext\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render_to_response\nfrom ebaysdk import finding\nfrom ebaysdk.exception import Connection...
[ { "content": "<|memory_start|>```python\nimport django.core.handlers.wsgi\nimport logging\nfrom django.template import loader,Context,RequestContext\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render_to_response\nfrom ebaysdk import finding\nfrom ebaysdk.exception i...
```python import django.core.handlers.wsgi import logging from django.template import loader,Context,RequestContext from django.http import HttpResponse, HttpResponseRedirect from django.shortcuts import render_to_response from ebaysdk import finding from ebaysdk.exception import ConnectionError from ebayapi.api import * from ebayapi import api as ebayapi from retail import Supplier,ShopInfo,getSupplierFromEbayInfo import retailtype from error import * from lxml import etree import urllib, random, json, datetime import zuser from StringIO import StringIO def getEbayInfo(request): token = getToken(request) if ('ebayinfo' in request.session): return request.session.get('ebayinfo',{}) else: return {} def getTokenFromEbayInfo(ebayinfo): return ebayinfo['token'] def ebay_ajax_prefix(handler): def rst_handler(request,*args,**kargs): token = getToken(request) if token: ebayinfo = getEbayInfo(request) return handler(request,ebayinfo,*args,**kargs) else: return returnError("Not authorised") return rst_handler def ebay_view_prefix(handler): def rst_handler(request,*args,**kargs): token = getToken(request) if token: ebayinfo = getEbayInfo(request) return handler(request,*args,**kargs) else: context = Context({}) return (render_to_response("ebaylogin.html",context,context_instance=RequestContext(request))) return rst_handler def GetXSLT(xslt_context,xslt_template): xslt_template = loader.get_template(xslt_template) xslt_str = xslt_template.render(xslt_context) xslt_doc = etree.parse(StringIO(xslt_str)) xslt = etree.XSLT(xslt_doc) return xslt def checkSuccess(doc): ack = doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): return True; else: return False; def getItem(itemid,token): item = GetItem(itemid,token) xml_doc = etree.parse(StringIO(item)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): title = xml_doc.xpath("//xs:Title", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text price = xml_doc.xpath("//xs:ConvertedCurrentPrice", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text return {'label':title,'value':price} else: return None return None # We will save the referrer so that we can route back def auth(request): # We first need to check whether this session is already linked to some ebay shop or not. token = getToken(request) if token: return HttpResponseRedirect('/admin/') else: if ('HTTP_REFERER' in request.META): request.session['continue'] = request.META['HTTP_REFERER'] sessionid = GetSessionID(request) xml_doc = etree.parse(StringIO(sessionid)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): session = xml_doc.xpath("//xs:SessionID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] ebayinfo = request.session.get('ebayinfo',{}) ebayinfo['session'] = session.text request.session['ebayinfo'] = ebayinfo args = urllib.quote_plus("zre="+request.META['HTTP_HOST']) token = GetToken(args,session.text) return token else: return HttpResponse(ack.text) def logoutebay(request): request.session['ebayinfo'] = None return HttpResponseRedirect('/admin/') def authsuccess(request): return HttpResponseRedirect('http://' + request.GET['zre']) def authfail(request): return HttpResponseRedirect('http://' + request.GET['zre']) # This private function gets the ebay token if it exists in the current session. It will try fetch one if ebay is connected. It returns None if failed to get a token. def getToken(request): if (not 'ebayinfo' in request.session) or (not request.session['ebayinfo']): request.session['ebayinfo'] = {} ebayinfo = request.session.get('ebayinfo',{}) user = zuser.getCurrentUser(request) # we are going to fetch the token if it does not exist yet token = "" if (('token' in ebayinfo) and (ebayinfo['token'])): token = ebayinfo['token'] else: if ('session' in ebayinfo): token = FetchToken(request,ebayinfo['session']) xml_doc = etree.parse(StringIO(token)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): token = xml_doc.xpath("//xs:eBayAuthToken", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] token = token.text else: msg = xml_doc.xpath("//xs:LongMessage", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] ebayerror = msg.text ebayinfo['error'] = ebayerror # should not update ebayinfo in request.session # request.session['ebayinfo'] = ebayinfo logging.info("Can not get token from ebay id:" + token) if (not token): # can not get token from session if user: usr = user if (usr and usr.ebaytoken): token = usr.ebaytoken # By the above computation we have tried to get the token if (token): ebayinfo['token'] = token else: logging.info("Can not get session for ebay auth") return None # so far we might need to update the token of the current user if user: usr = user if (usr): usr.ebaytoken = token usr.put() logging.info("ebayinfo:" + json.dumps(ebayinfo)) if ('token' in ebayinfo) and ebayinfo['token']: request.session['ebayinfo'] = ebayinfo # here we try to get as much info as possible from a ebay token if((not 'id' in ebayinfo) or (not 'email' in ebayinfo)): user = GetUserInfo(token) user_doc = etree.parse(StringIO(user)) ack = user_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if ('Success' in ack.text): email = user_doc.xpath("//xs:Email", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] ebayinfo['email'] = email.text uid = user_doc.xpath("//xs:UserID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] ebayinfo['id'] = uid.text else: request.session['ebayinfo'] = {} logging.info("Can not find email address in ebayinfo") return None if((not 'store' in ebayinfo) or (not 'logo' in ebayinfo) or (not 'category' in ebayinfo)): store = GetStore(token) store_doc = etree.parse(StringIO(store)) ack = store_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if ('Success' in ack.text): name = store_doc.xpath("//xs:Store/xs:Name", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] ebayinfo['store'] = name.text logo = store_doc.xpath("//xs:Store/xs:Logo/xs:URL", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) if logo: ebayinfo['logo'] = logo[0].text else: ebayinfo['logo'] = None cgs = {} categories = store_doc.xpath("//xs:Store/xs:CustomCategories/xs:CustomCategory", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) a = etree.tostring(categories[0]) for category in categories: name = category.xpath("./xs:Name", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text id = category.xpath("./xs:CategoryID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text cgs[id] = {'name':name,'children':{}} childcategories = category.xpath("./xs:ChildCategory", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) for child in childcategories: name = child.xpath("./xs:Name", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text cid = child.xpath("./xs:CategoryID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text cgs[id]['children'][cid] = {'name':name} ebayinfo['categories'] = cgs else: request.session['ebayinfo'] = {} logging.info("Can not find shopinfo in ebayinfo:" + store) return None request.session['ebayinfo'] = ebayinfo currentSite().setebayinfo(json.dumps(ebayinfo)) return ebayinfo['token'] else: return None #### # This function will append general infomation after item description # It will replace everything after <!-- below is embed code --> tag #### @ebay_view_prefix def ebayorders(request): tt = datetime.datetime.utcnow() context = Context({"ORDER_GROUP":[tt]}) return (render_to_response("ebayorders.html",context,context_instance=RequestContext(request))) @ebay_ajax_prefix def ebayordersajax(request,ebayinfo): token = getTokenFromEbayInfo(ebayinfo) year = request.GET['year'] month = request.GET['month'] day = request.GET['day'] tt = datetime.datetime(year=int(year),month=int(month),day=int(day)) ft = tt - datetime.timedelta(hours=120) tt = tt.strftime("%Y-%m-%dT%H:%M:%S.000Z") ft = ft.strftime("%Y-%m-%dT%H:%M:%S.000Z") xml_doc_str = GetOrders(token,ft,tt) xml_doc = etree.parse(StringIO(xml_doc_str)) xslt = GetXSLT(Context({}),'xslt/EbayOrdersJSON.xslt') xrst = xslt(xml_doc) rst = unicode(xrst) return HttpResponse(rst) def relist(ebayinfo,item): token = ebayinfo['token'] config = {'SELLER_ID':ebayinfo['id']} config['INITIAL'] = item.description config['ITEM'] = item config['EXTRA'] = ShopInfo.all().filter("type =","ebay").order("name") format = loader.get_template("ebay/format.html") content = format.render(Context(config)) ebayitem = GetItem(item.ebayid,token) xml_doc = etree.parse(StringIO(ebayitem)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): sellingstatus = xml_doc.xpath("//xs:SellingStatus/xs:ListingStatus", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text if (sellingstatus == "Completed"): revise = RelistItemSimple(item,token,content) xml_doc = etree.parse(StringIO(revise)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): ebayid = xml_doc.xpath("//xs:ItemID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text item.ebayid = refid item.put() return (HttpResponse(revise,mimetype = "text/xml"),item) else: return (returnError("Related ebay item is still active"),item) else: return (HttpResponse(ebayitem,mimetype = "text/xml"),item) #### # This function will append general infomation after item description # It will replace everything after <!-- below is embed code --> tag #### def format(ebayinfo,itemid): token = ebayinfo['token'] id = ebayinfo['id'] config = {'SELLER_ID':id} item = GetItem(itemid,token) xml_doc = etree.parse(StringIO(item)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): description = xml_doc.xpath("//xs:Description", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] refid = xml_doc.xpath("//xs:SKU", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) if (not refid): return returnError('SKU Not Provided') else: refid = refid[0].text # refid = xml_doc.xpath("//xs:ItemID", # namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text name = xml_doc.xpath("//xs:Title", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text price = xml_doc.xpath("//xs:ConvertedCurrentPrice", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text galleryurl = xml_doc.xpath("//xs:GalleryURL", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text infourl = xml_doc.xpath("//xs:ViewItemURL", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text ebaycategory = xml_doc.xpath("//xs:PrimaryCategory/xs:CategoryID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text category = xml_doc.xpath("//xs:StoreCategoryID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text sndcategory = xml_doc.xpath("//xs:StoreCategory2ID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text sellingstatus = xml_doc.xpath("//xs:SellingStatus/xs:ListingStatus", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text topd = description.text.split("<!-- below is embeded code -->") config['INITIAL'] = topd[0] config['EXTRA'] = ShopInfo.all().filter("type =","ebay").order("name") # save the item iteminfo = {'refid':refid,'name':name ,'price':float(price),'cost':float(price),'galleryurl':galleryurl ,'infourl':infourl,'category':category,'sndcategory':sndcategory ,'description':topd[0],'ebayid':itemid,'ebaycategory':ebaycategory ,'specification':"{}"} item = retailtype.getItem(refid) supplier = getSupplierFromEbayInfo(ebayinfo) if item: iteminfo['specification'] = item.specification # FIXME: We do not update galleryurl back to ebay gallery url at the moment. # iteminfo['galleryurl'] = item.galleryurl item.ebayid = itemid supplier = item.parent() zitem = supplier.saveItem(iteminfo) config['ITEM'] = zitem format = loader.get_template("ebay/format.html") content = format.render(Context(config)) if (sellingstatus != "Completed"): revise = ReviseItemSimple(item,token,content) xml_doc = etree.parse(StringIO(revise)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): return (HttpResponse(revise,mimetype = "text/xml"),item) else: return (HttpResponse(revise,mimetype = "text/xml"),None) else: revise = RelistItemSimple(item,token,content) xml_doc = etree.parse(StringIO(revise)) ack = xml_doc.xpath("//xs:Ack", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] if('Success' in ack.text): ebayid = xml_doc.xpath("//xs:ItemID", namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text zitem.ebayid = refid zitem.put() return (HttpResponse(revise,mimetype = "text/xml"),item) else: return (HttpResponse(revise,mimetype = "text/xml"),None) else: return (HttpResponse(item,mimetype = "text/xml"),None) #### # This function will append general infomation after item description # It will replace everything after <!-- below is embed code --> tag #### def sync(ebayinfo,item): token = ebayinfo['token'] id = ebayinfo['id'] config = {'SELLER_ID':id} description = item.description name = item.name config['INITIAL'] = description config['ITEM'] = item config['EXTRA'] = ShopInfo.all().filter("type =","ebay").order("name") format = loader.get_template("ebay/format.html") content = format.render(Context(config)) if (not item.ebayid): revise = ReviseItemBySKU(item.refid,name,token,content) else: revise = ReviseItem(item,token,content) return HttpResponse(revise,mimetype = "text/xml") def getactivelist(request): token = getToken(request) page = 1 if ("page" in request.GET): page = int(request.GET['page']) xml_doc = None if token: if 'itemid' in request.GET: rid = request.GET['itemid'] iteminfo = GetItem(rid,token) xml_doc = etree.parse(StringIO(iteminfo)) xslt = GetXSLT(Context({}),'xslt/MyeBaySelling.xslt') list_content = etree.tostring(xslt(xml_doc.getroot())) else: my_ebay_selling = GetMyeBaySelling(token,page) xml_doc = etree.parse(StringIO(my_ebay_selling)) total = xml_doc.xpath("//xs:TotalNumberOfPages",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0] total = int(total.text); xslt = GetXSLT(Context({'pages':range(total+1)[1:]}),'xslt/MyeBaySelling2.xslt') list_content = etree.tostring(xslt(xml_doc.getroot())) return list_content else: return None def getinactivelist(request): token = getToken(request) page = 1 if ("page" in request.GET): page = int(request.GET['page']) xml_doc = None if token: if 'itemid' in request.GET: rid = request.GET['itemid'] iteminfo = GetItem(rid,token) xml_doc = etree.parse(StringIO(iteminfo)) xslt = GetXSLT(Context({}),'xslt/MyeBaySelling.xslt') list_content = etree.tostring(xslt(xml_doc.getroot())) else: my_ebay_selling = GetMyeBaySellingInactive(token,page) xml_doc = etree.parse(StringIO(my_ebay_selling)) total = xml_doc.xpath("//xs:TotalNumberOfPages",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) list_content = "" #none item if there is no TotalNumberOfPages provided if(total): total = int(total[0].text); xslt = GetXSLT(Context({'pages':range(total+1)[1:]}),'xslt/MyeBaySelling2.xslt') list_content = etree.tostring(xslt(xml_doc.getroot())) return list_content else: return None @ebay_view_prefix def fetchcategory(request): query = request.GET['term'] token = getToken(request) rslt = GetCategories(request,token,query) xml_doc = etree.parse(StringIO(rslt)) suggests = xml_doc.xpath("//xs:SuggestedCategory",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) items = [] for suggest in suggests: id = suggest.xpath("./xs:Category/xs:CategoryID",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text label = suggest.xpath("./xs:Category/xs:CategoryName",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"})[0].text label = [label] parents = suggest.xpath("./xs:Category/xs:CategoryParentName",namespaces={'xs':"urn:ebay:apis:eBLBaseComponents"}) for parent in parents: label.append(parent.text) label = "->".join(label) items.append({'label':label,'value':id}) return HttpResponse(json.dumps(items),mimetype="text/plain") ```
[ { "content": "```python\nimport subprocess\nclass CmdRunException(Exception):\n pass\n\nclass Util(object):\n verbose = False\n @classmethod\n def cmdOutput(cls, cmd, havepassword = False):\n if cls.verbose and not havepassword:\n print(cmd)\n proc = subprocess.Popen(cmd, sh...
[ { "content": "<|memory_start|>```python\nimport subprocess\nclass CmdRunException(Exception):\n pass\n\nclass Util(object):\n verbose = False\n @classmethod\n def cmdOutput(cls, cmd, havepassword = False):\n if cls.verbose and not havepassword:\n print(cmd)\n proc = subproce...
```python import subprocess class CmdRunException(Exception): pass class Util(object): verbose = False @classmethod def cmdOutput(cls, cmd, havepassword = False): if cls.verbose and not havepassword: print(cmd) proc = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) proc.wait() if (0 != proc.returncode): if havepassword: cmd = "<command containing password>" raise CmdRunException((cmd,proc.stdout.read(),proc.stderr.read())) output = proc.stdout.read().strip() return output @classmethod def runCmd(cls, cmd, havepassword = False): if cls.verbose and not havepassword: print(cmd) p = subprocess.call(cmd, shell=True) if (0 != p): if havepassword: cmd = "<command containing password>" raise CmdRunException(cmd) @classmethod def setVerbose(cls): cls.verbose = True ```
[ { "content": "Return the code unaltered:\n```python\n# Copyright 2016 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# ht...
[ { "content": "Return the code unaltered:\n<|memory_start|>```python\n# Copyright 2016 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the Licen...
```python # Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Example of DNNClassifier for Iris plant dataset.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from sklearn import metrics, cross_validation import tensorflow as tf from tensorflow.contrib import learn def main(unused_argv): # Load dataset. iris = learn.datasets.load_dataset('iris') x_train, x_test, y_train, y_test = cross_validation.train_test_split( iris.data, iris.target, test_size=0.2, random_state=42) # Build 3 layer DNN with 10, 20, 10 units respectively. classifier = learn.DNNClassifier(hidden_units=[10, 20, 10], n_classes=3) # Fit and predict. classifier.fit(x_train, y_train, steps=200) score = metrics.accuracy_score(y_test, classifier.predict(x_test)) print('Accuracy: {0:f}'.format(score)) if __name__ == '__main__': tf.app.run() ```
[ { "content": "Here is the snippet:\n```python\n# coding: utf-8\n# Copyright (C) 2016 Bruno Abude Cardoso\n#\n# Imagem Cinemática is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of th...
[ { "content": "Here is the snippet:\n<|memory_start|>```python\n# coding: utf-8\n# Copyright (C) 2016 Bruno Abude Cardoso\n#\n# Imagem Cinemática is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either...
```python # coding: utf-8 # Copyright (C) 2016 Bruno Abude Cardoso # # Imagem Cinemática is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Imagem Cinemática is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import cv2 class ICGRAY2BGR(object): def __init__(self, plugin_path): self.plugin_path = plugin_path self.parameters = [] def parameter_changed(self, param_name, value): return None def apply_filter(self, frame): colorspace, data, pos, timestamp = frame data = cv2.cvtColor(data, cv2.COLOR_GRAY2BGR) return ("BGR", data) def release_plugin(self, error_level=0): pass ```
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n```python\nimport os\nimport time\nimport errno\n\n\nclass LockFileError(RuntimeError):\n pass\n\n\nclass LockFile(object):\n \"\"\"\n Lock file (Unix-only), implemented via symlinks.\n \"\"\"\n\n def __init__(self, filename, tim...
[ { "content": "Reconstruct the code file line-for-line, unmodified:\n<|memory_start|>```python\nimport os\nimport time\nimport errno\n\n\nclass LockFileError(RuntimeError):\n pass\n\n\nclass LockFile(object):\n \"\"\"\n Lock file (Unix-only), implemented via symlinks.\n \"\"\"\n\n def __init__(sel...
```python import os import time import errno class LockFileError(RuntimeError): pass class LockFile(object): """ Lock file (Unix-only), implemented via symlinks. """ def __init__(self, filename, timeout=0.05, fail_if_active=False): self.filename = filename self.timeout = timeout self.fd = None self.fail_if_active = fail_if_active @classmethod def check(self, filename): # Check if the lockfile exists try: pid = int(os.readlink(filename)) except OSError as err: if err.errno == errno.ENOENT: return False raise # Check if the process is still alive try: os.kill(pid, 0) except OSError as err: if err.errno == errno.ESRCH: # no such process return False raise # Seems to be still around return True def __enter__(self): tries = 0 while True: if tries > 0: if self.fail_if_active: raise LockFileError("Process is already running") else: time.sleep(self.timeout) try: os.symlink(str(os.getpid()), self.filename) break except OSError as err: if err.errno != errno.EEXIST: raise try: pid = int(os.readlink(self.filename)) except OSError as err: if err.errno == errno.ENOENT: continue raise # Check if it's still alive try: os.kill(pid, 0) except OSError as err: if err.errno == errno.ESRCH: # no such process os.unlink(self.filename) continue raise tries += 1 def __exit__(self, type, value, traceback): os.unlink(self.filename) ```
[ { "content": "Here is the code block:\n```python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2009 Rene Liebscher\n#\n# This program is free software; you can redistribute it and/or modify it under\n# the terms of the GNU Lesser General Public License as published by the Free \n# Software Foundation; either ve...
[ { "content": "Here is the code block:\n<|memory_start|>```python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2009 Rene Liebscher\n#\n# This program is free software; you can redistribute it and/or modify it under\n# the terms of the GNU Lesser General Public License as published by the Free \n# Software Found...
```python # -*- coding: utf-8 -*- # # Copyright (C) 2009 Rene Liebscher # # This program is free software; you can redistribute it and/or modify it under # the terms of the GNU Lesser General Public License as published by the Free # Software Foundation; either version 3 of the License, or (at your option) any # later version. # # This program is distributed in the hope that it will be useful, but WITHOUT # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS # FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more # details. # # You should have received a copy of the GNU Lesser General Public License # along with this program; if not, see <http://www.gnu.org/licenses/>. # __revision__ = "$Id: DualOfHarmonicMean.py,v 1.7 2009-10-27 20:06:27 rliebscher Exp $" from fuzzy.norm.Norm import Norm, product, sum class DualOfHarmonicMean(Norm): def __init__(self): super(DualOfHarmonicMean, self).__init__(Norm.UNKNOWN) #XXX def __call__(self, *args): args = self.checkArgsN(args) sum_ = sum(*args) if sum_ == len(args): return 1.0 product_ = product(*args) count_ = float(len(args)) return (sum_-count_*product_)/(count_-sum_) ```
[ { "content": "```python\n# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport django\nfrom django.contrib.admin.sites import AdminSite\nfrom django.utils.functional import SimpleLazyObject\n\nfrom djadmin import settings\nfrom .models import DjadminModelSetting\nfrom .signals import get_re...
[ { "content": "<|memory_start|>```python\n# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport django\nfrom django.contrib.admin.sites import AdminSite\nfrom django.utils.functional import SimpleLazyObject\n\nfrom djadmin import settings\nfrom .models import DjadminModelSetting\nfrom .signa...
```python # -*- coding: utf-8 -*- from __future__ import unicode_literals import django from django.contrib.admin.sites import AdminSite from django.utils.functional import SimpleLazyObject from djadmin import settings from .models import DjadminModelSetting from .signals import get_register_model_with_mixin, handle_djadmin_field_data, add_visitor from .util import (get_user_agent, get_admin_color_theme, get_admin_color_theme_hex_code, is_session_exist, create_new_session, is_admin_url) if django.VERSION >= (1, 10): from django.utils.deprecation import MiddlewareMixin else: MiddlewareMixin = object class DJMiddleware(MiddlewareMixin): def process_request(self, request): # Set user_agent of user in request request.user_agent = SimpleLazyObject(lambda: get_user_agent(request)) # Check user session if not is_session_exist(request): # Add as a visitor session = create_new_session(request) add_visitor(request) if is_admin_url(request): admin_color_theme = get_admin_color_theme(settings.ADMIN_COLOR_THEME) admin_color_theme_code = get_admin_color_theme_hex_code(admin_color_theme) allow_forget_password_admin = settings.ALLOW_FORGET_PASSWORD_ADMIN AdminSite.site_header = settings.ADMIN_HEADER_TITLE request.ADMIN_COLOR_THEME = admin_color_theme request.ALLOW_FORGET_PASSWORD_ADMIN = allow_forget_password_admin request.ADMIN_COLOR_THEME_CODE = admin_color_theme_code if request.user.is_superuser and settings.DJADMIN_DYNAMIC_FIELD_DISPLAY: register_model_object_list = get_register_model_with_mixin() exist_model_object_list = DjadminModelSetting.objects.all() register_model_list = [model.__name__ for model in register_model_object_list] exist_model_list = [str(model.model) for model in exist_model_object_list] create_model_name = [model for model in register_model_list if model not in exist_model_list] delete_model_name = [model for model in exist_model_list if model not in register_model_list] if len(create_model_name): handle_djadmin_field_data(register_model_object_list, True) if len(delete_model_name): if settings.DJADMIN_DYNAMIC_DELETE_UNREGISTER_FIELD: handle_djadmin_field_data(register_model_object_list, False) ```
[ { "content": "Here is a code file:\n```python\nimport pickle\n\nPA = 'A6'\n\ndef load(partId, caseId=1):\n\t\"\"\"\n\tThis function returns the example test-cases for a specific part of an assignment.\n\tInput:\n\t\tpartId (int) = part number of the assignment (1 for A*Part1, 2 for A*Part2 and so on)\n\t\tcaseI...
[ { "content": "Here is a code file:\n<|memory_start|>```python\nimport pickle\n\nPA = 'A6'\n\ndef load(partId, caseId=1):\n\t\"\"\"\n\tThis function returns the example test-cases for a specific part of an assignment.\n\tInput:\n\t\tpartId (int) = part number of the assignment (1 for A*Part1, 2 for A*Part2 and s...
```python import pickle PA = 'A6' def load(partId, caseId=1): """ This function returns the example test-cases for a specific part of an assignment. Input: partId (int) = part number of the assignment (1 for A*Part1, 2 for A*Part2 and so on) caseId (int) = caseId = k to return the kth test case. Typically there are two per part. Output: testcase (dict) = {'input': <input test case>, 'output': <expected output for the input test case>} """ data = pickle.load(open('testInput%s.pkl'%PA,'r')) part = u'%s-part-%d'%(PA, partId) if not data['exampleInputs'].has_key(part): print "There are no example test cases required for this part. You can keep on improving the pitch track and submit once you are satisfied. Plots provide you feedback on the accuracy of the pitch track." return None if caseId > len(data['exampleInputs'][part]) or caseId <=0: print "Please provide a valid caseId (>=1), number of test cases in this assignment are %d"%(len(data['exampleInputs'][part])) return None return {'input': data['exampleInputs'][part][caseId-1], 'output': data['exampleOutputs'][part][caseId-1]} ```
[ { "content": "Here is the snippet:\n```python\nfrom sympy import S, Symbol, I, Rational, PurePoly\nfrom sympy.matrices import Matrix, SparseMatrix, eye, zeros, ShapeError\nfrom sympy.utilities.pytest import raises, XFAIL\n\n\ndef test_sparse_matrix():\n def sparse_eye(n):\n return SparseMatrix.eye(n)\...
[ { "content": "Here is the snippet:\n<|memory_start|>```python\nfrom sympy import S, Symbol, I, Rational, PurePoly\nfrom sympy.matrices import Matrix, SparseMatrix, eye, zeros, ShapeError\nfrom sympy.utilities.pytest import raises, XFAIL\n\n\ndef test_sparse_matrix():\n def sparse_eye(n):\n return Spar...
```python from sympy import S, Symbol, I, Rational, PurePoly from sympy.matrices import Matrix, SparseMatrix, eye, zeros, ShapeError from sympy.utilities.pytest import raises, XFAIL def test_sparse_matrix(): def sparse_eye(n): return SparseMatrix.eye(n) def sparse_zeros(n): return SparseMatrix.zeros(n) # creation args raises(TypeError, lambda: SparseMatrix(1, 2)) a = SparseMatrix(( (1, 0), (0, 1) )) assert SparseMatrix(a) == a # test element assignment a = SparseMatrix(( (1, 0), (0, 1) )) a[3] = 4 assert a[1, 1] == 4 a[3] = 1 a[0, 0] = 2 assert a == SparseMatrix(( (2, 0), (0, 1) )) a[1, 0] = 5 assert a == SparseMatrix(( (2, 0), (5, 1) )) a[1, 1] = 0 assert a == SparseMatrix(( (2, 0), (5, 0) )) assert a._smat == {(0, 0): 2, (1, 0): 5} # test_multiplication a = SparseMatrix(( (1, 2), (3, 1), (0, 6), )) b = SparseMatrix(( (1, 2), (3, 0), )) c = a*b assert c[0, 0] == 7 assert c[0, 1] == 2 assert c[1, 0] == 6 assert c[1, 1] == 6 assert c[2, 0] == 18 assert c[2, 1] == 0 x = Symbol("x") c = b * Symbol("x") assert isinstance(c, SparseMatrix) assert c[0, 0] == x assert c[0, 1] == 2*x assert c[1, 0] == 3*x assert c[1, 1] == 0 c = 5 * b assert isinstance(c, SparseMatrix) assert c[0, 0] == 5 assert c[0, 1] == 2*5 assert c[1, 0] == 3*5 assert c[1, 1] == 0 #test_power A = SparseMatrix([[2, 3], [4, 5]]) assert (A**5)[:] == [6140, 8097, 10796, 14237] A = SparseMatrix([[2, 1, 3], [4, 2, 4], [6, 12, 1]]) assert (A**3)[:] == [290, 262, 251, 448, 440, 368, 702, 954, 433] # test_creation x = Symbol("x") a = SparseMatrix([[x, 0], [0, 0]]) m = a assert m.cols == m.rows assert m.cols == 2 assert m[:] == [x, 0, 0, 0] b = SparseMatrix(2, 2, [x, 0, 0, 0]) m = b assert m.cols == m.rows assert m.cols == 2 assert m[:] == [x, 0, 0, 0] assert a == b S = sparse_eye(3) S.row_del(1) assert S == SparseMatrix([ [1, 0, 0], [0, 0, 1]]) S = sparse_eye(3) S.col_del(1) assert S == SparseMatrix([ [1, 0], [0, 0], [0, 1]]) S = SparseMatrix.eye(3) S[2, 1] = 2 S.col_swap(1, 0) assert S == SparseMatrix([ [0, 1, 0], [1, 0, 0], [2, 0, 1]]) a = SparseMatrix(1, 2, [1, 2]) b = a.copy() c = a.copy() assert a[0] == 1 a.row_del(0) assert a == SparseMatrix(0, 2, []) b.col_del(1) assert b == SparseMatrix(1, 1, [1]) # test_determinant x, y = Symbol('x'), Symbol('y') assert SparseMatrix(1, 1, [0]).det() == 0 assert SparseMatrix([[1]]).det() == 1 assert SparseMatrix(((-3, 2), (8, -5))).det() == -1 assert SparseMatrix(((x, 1), (y, 2*y))).det() == 2*x*y - y assert SparseMatrix(( (1, 1, 1), (1, 2, 3), (1, 3, 6) )).det() == 1 assert SparseMatrix(( ( 3, -2, 0, 5), (-2, 1, -2, 2), ( 0, -2, 5, 0), ( 5, 0, 3, 4) )).det() == -289 assert SparseMatrix(( ( 1, 2, 3, 4), ( 5, 6, 7, 8), ( 9, 10, 11, 12), (13, 14, 15, 16) )).det() == 0 assert SparseMatrix(( (3, 2, 0, 0, 0), (0, 3, 2, 0, 0), (0, 0, 3, 2, 0), (0, 0, 0, 3, 2), (2, 0, 0, 0, 3) )).det() == 275 assert SparseMatrix(( (1, 0, 1, 2, 12), (2, 0, 1, 1, 4), (2, 1, 1, -1, 3), (3, 2, -1, 1, 8), (1, 1, 1, 0, 6) )).det() == -55 assert SparseMatrix(( (-5, 2, 3, 4, 5), ( 1, -4, 3, 4, 5), ( 1, 2, -3, 4, 5), ( 1, 2, 3, -2, 5), ( 1, 2, 3, 4, -1) )).det() == 11664 assert SparseMatrix(( ( 2, 7, -1, 3, 2), ( 0, 0, 1, 0, 1), (-2, 0, 7, 0, 2), (-3, -2, 4, 5, 3), ( 1, 0, 0, 0, 1) )).det() == 123 # test_submatrix m0 = sparse_eye(4) assert m0[:3, :3] == sparse_eye(3) assert m0[2:4, 0:2] == sparse_zeros(2) m1 = SparseMatrix(3, 3, lambda i, j: i + j) assert m1[0, :] == SparseMatrix(1, 3, (0, 1, 2)) assert m1[1:3, 1] == SparseMatrix(2, 1, (2, 3)) m2 = SparseMatrix( [[0, 1, 2, 3], [4, 5, 6, 7], [8, 9, 10, 11], [12, 13, 14, 15]]) assert m2[:, -1] == SparseMatrix(4, 1, [3, 7, 11, 15]) assert m2[-2:, :] == SparseMatrix([[8, 9, 10, 11], [12, 13, 14, 15]]) assert SparseMatrix([[1, 2], [3, 4]]).submatrix([1, 1]) == Matrix([[4]]) # test_submatrix_assignment m = sparse_zeros(4) m[2:4, 2:4] = sparse_eye(2) assert m == SparseMatrix([(0, 0, 0, 0), (0, 0, 0, 0), (0, 0, 1, 0), (0, 0, 0, 1)]) assert len(m._smat) == 2 m[:2, :2] = sparse_eye(2) assert m == sparse_eye(4) m[:, 0] = SparseMatrix(4, 1, (1, 2, 3, 4)) assert m == SparseMatrix([(1, 0, 0, 0), (2, 1, 0, 0), (3, 0, 1, 0), (4, 0, 0, 1)]) m[:, :] = sparse_zeros(4) assert m == sparse_zeros(4) m[:, :] = ((1, 2, 3, 4), (5, 6, 7, 8), (9, 10, 11, 12), (13, 14, 15, 16)) assert m == SparseMatrix((( 1, 2, 3, 4), ( 5, 6, 7, 8), ( 9, 10, 11, 12), (13, 14, 15, 16))) m[:2, 0] = [0, 0] assert m == SparseMatrix((( 0, 2, 3, 4), ( 0, 6, 7, 8), ( 9, 10, 11, 12), (13, 14, 15, 16))) # test_reshape m0 = sparse_eye(3) assert m0.reshape(1, 9) == SparseMatrix(1, 9, (1, 0, 0, 0, 1, 0, 0, 0, 1)) m1 = SparseMatrix(3, 4, lambda i, j: i + j) assert m1.reshape(4, 3) == \ SparseMatrix([(0, 1, 2), (3, 1, 2), (3, 4, 2), (3, 4, 5)]) assert m1.reshape(2, 6) == \ SparseMatrix([(0, 1, 2, 3, 1, 2), (3, 4, 2, 3, 4, 5)]) # test_applyfunc m0 = sparse_eye(3) assert m0.applyfunc(lambda x: 2*x) == sparse_eye(3)*2 assert m0.applyfunc(lambda x: 0 ) == sparse_zeros(3) # test_LUdecomp testmat = SparseMatrix([[ 0, 2, 5, 3], [ 3, 3, 7, 4], [ 8, 4, 0, 2], [-2, 6, 3, 4]]) L, U, p = testmat.LUdecomposition() assert L.is_lower assert U.is_upper assert (L*U).permuteBkwd(p) - testmat == sparse_zeros(4) testmat = SparseMatrix([[ 6, -2, 7, 4], [ 0, 3, 6, 7], [ 1, -2, 7, 4], [-9, 2, 6, 3]]) L, U, p = testmat.LUdecomposition() assert L.is_lower assert U.is_upper assert (L*U).permuteBkwd(p) - testmat == sparse_zeros(4) x, y, z = Symbol('x'), Symbol('y'), Symbol('z') M = Matrix(((1, x, 1), (2, y, 0), (y, 0, z))) L, U, p = M.LUdecomposition() assert L.is_lower assert U.is_upper assert (L*U).permuteBkwd(p) - M == sparse_zeros(3) # test_LUsolve A = SparseMatrix([[2, 3, 5], [3, 6, 2], [8, 3, 6]]) x = SparseMatrix(3, 1, [3, 7, 5]) b = A*x soln = A.LUsolve(b) assert soln == x A = SparseMatrix([[0, -1, 2], [5, 10, 7], [8, 3, 4]]) x = SparseMatrix(3, 1, [-1, 2, 5]) b = A*x soln = A.LUsolve(b) assert soln == x # test_inverse A = sparse_eye(4) assert A.inv() == sparse_eye(4) assert A.inv(method="CH") == sparse_eye(4) assert A.inv(method="LDL") == sparse_eye(4) A = SparseMatrix([[2, 3, 5], [3, 6, 2], [7, 2, 6]]) Ainv = SparseMatrix(Matrix(A).inv()) assert A*Ainv == sparse_eye(3) assert A.inv(method="CH") == Ainv assert A.inv(method="LDL") == Ainv A = SparseMatrix([[2, 3, 5], [3, 6, 2], [5, 2, 6]]) Ainv = SparseMatrix(Matrix(A).inv()) assert A*Ainv == sparse_eye(3) assert A.inv(method="CH") == Ainv assert A.inv(method="LDL") == Ainv # test_cross v1 = Matrix(1, 3, [1, 2, 3]) v2 = Matrix(1, 3, [3, 4, 5]) assert v1.cross(v2) == Matrix(1, 3, [-2, 4, -2]) assert v1.norm(2)**2 == 14 # conjugate a = SparseMatrix(((1, 2 + I), (3, 4))) assert a.C == SparseMatrix([ [1, 2 - I], [3, 4] ]) # mul assert a*Matrix(2, 2, [1, 0, 0, 1]) == a assert a + Matrix(2, 2, [1, 1, 1, 1]) == SparseMatrix([ [2, 3 + I], [4, 5] ]) # col join assert a.col_join(sparse_eye(2)) == SparseMatrix([ [1, 2 + I], [3, 4], [1, 0], [0, 1] ]) # symmetric assert not a.is_symmetric(simplify=False) # test_cofactor assert sparse_eye(3) == sparse_eye(3).cofactorMatrix() test = SparseMatrix([[1, 3, 2], [2, 6, 3], [2, 3, 6]]) assert test.cofactorMatrix() == \ SparseMatrix([[27, -6, -6], [-12, 2, 3], [-3, 1, 0]]) test = SparseMatrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]]) assert test.cofactorMatrix() == \ SparseMatrix([[-3, 6, -3], [6, -12, 6], [-3, 6, -3]]) # test_jacobian x = Symbol('x') y = Symbol('y') L = SparseMatrix(1, 2, [x**2*y, 2*y**2 + x*y]) syms = [x, y] assert L.jacobian(syms) == Matrix([[2*x*y, x**2], [y, 4*y + x]]) L = SparseMatrix(1, 2, [x, x**2*y**3]) assert L.jacobian(syms) == SparseMatrix([[1, 0], [2*x*y**3, x**2*3*y**2]]) # test_QR A = Matrix([[1, 2], [2, 3]]) Q, S = A.QRdecomposition() R = Rational assert Q == Matrix([ [ 5**R(-1, 2), (R(2)/5)*(R(1)/5)**R(-1, 2)], [2*5**R(-1, 2), (-R(1)/5)*(R(1)/5)**R(-1, 2)]]) assert S == Matrix([ [5**R(1, 2), 8*5**R(-1, 2)], [ 0, (R(1)/5)**R(1, 2)]]) assert Q*S == A assert Q.T * Q == sparse_eye(2) R = Rational # test nullspace # first test reduced row-ech form M = SparseMatrix([[5, 7, 2, 1], [1, 6, 2, -1]]) out, tmp = M.rref() assert out == Matrix([[1, 0, -R(2)/23, R(13)/23], [0, 1, R(8)/23, R(-6)/23]]) M = SparseMatrix([[ 1, 3, 0, 2, 6, 3, 1], [-2, -6, 0, -2, -8, 3, 1], [ 3, 9, 0, 0, 6, 6, 2], [-1, -3, 0, 1, 0, 9, 3]]) out, tmp = M.rref() assert out == Matrix([[1, 3, 0, 0, 2, 0, 0], [0, 0, 0, 1, 2, 0, 0], [0, 0, 0, 0, 0, 1, R(1)/3], [0, 0, 0, 0, 0, 0, 0]]) # now check the vectors basis = M.nullspace() assert basis[0] == Matrix([-3, 1, 0, 0, 0, 0, 0]) assert basis[1] == Matrix([0, 0, 1, 0, 0, 0, 0]) assert basis[2] == Matrix([-2, 0, 0, -2, 1, 0, 0]) assert basis[3] == Matrix([0, 0, 0, 0, 0, R(-1)/3, 1]) # test eigen x = Symbol('x') y = Symbol('y') sparse_eye3 = sparse_eye(3) assert sparse_eye3.charpoly(x) == PurePoly(((x - 1)**3)) assert sparse_eye3.charpoly(y) == PurePoly(((y - 1)**3)) # test values M = Matrix([( 0, 1, -1), ( 1, 1, 0), (-1, 0, 1)]) vals = M.eigenvals() assert sorted(vals.keys()) == [-1, 1, 2] R = Rational M = Matrix([[1, 0, 0], [0, 1, 0], [0, 0, 1]]) assert M.eigenvects() == [(1, 3, [ Matrix([1, 0, 0]), Matrix([0, 1, 0]), Matrix([0, 0, 1])])] M = Matrix([[5, 0, 2], [3, 2, 0], [0, 0, 1]]) assert M.eigenvects() == [(1, 1, [Matrix([R(-1)/2, R(3)/2, 1])]), (2, 1, [Matrix([0, 1, 0])]), (5, 1, [Matrix([1, 1, 0])])] assert M.zeros(3, 5) == SparseMatrix(3, 5, {}) def test_transpose(): assert SparseMatrix(((1, 2), (3, 4))).transpose() == \ SparseMatrix(((1, 3), (2, 4))) def test_trace(): assert SparseMatrix(((1, 2), (3, 4))).trace() == 5 assert SparseMatrix(((0, 0), (0, 4))).trace() == 4 def test_CL_RL(): assert SparseMatrix(((1, 2), (3, 4))).row_list() == \ [(0, 0, 1), (0, 1, 2), (1, 0, 3), (1, 1, 4)] assert SparseMatrix(((1, 2), (3, 4))).col_list() == \ [(0, 0, 1), (1, 0, 3), (0, 1, 2), (1, 1, 4)] def test_add(): assert SparseMatrix(((1, 0), (0, 1))) + SparseMatrix(((0, 1), (1, 0))) == \ SparseMatrix(((1, 1), (1, 1))) a = SparseMatrix(100, 100, lambda i, j: int(j != 0 and i % j == 0)) b = SparseMatrix(100, 100, lambda i, j: int(i != 0 and j % i == 0)) assert (len(a._smat) + len(b._smat) - len((a + b)._smat) > 0) def test_errors(): raises(ValueError, lambda: SparseMatrix(1.4, 2, lambda i, j: 0)) raises(TypeError, lambda: SparseMatrix([1, 2, 3], [1, 2])) raises(ValueError, lambda: SparseMatrix([[1, 2], [3, 4]])[(1, 2, 3)]) raises(IndexError, lambda: SparseMatrix([[1, 2], [3, 4]])[5]) raises(ValueError, lambda: SparseMatrix([[1, 2], [3, 4]])[1, 2, 3]) raises(TypeError, lambda: SparseMatrix([[1, 2], [3, 4]]).copyin_list([0, 1], set([]))) raises( IndexError, lambda: SparseMatrix([[1, 2], [3, 4]]).submatrix((1, 2))) raises(TypeError, lambda: SparseMatrix([1, 2, 3]).cross(1)) raises(IndexError, lambda: SparseMatrix(1, 2, [1, 2])[3]) raises(ShapeError, lambda: SparseMatrix(1, 2, [1, 2]) + SparseMatrix(2, 1, [2, 1])) def test_len(): assert not SparseMatrix() assert SparseMatrix() == SparseMatrix([]) @XFAIL def test_len_different_shapes(): assert Matrix() == Matrix([[]]) assert SparseMatrix() == SparseMatrix([[]]) def test_sparse_zeros_sparse_eye(): assert SparseMatrix.eye(3) == eye(3, cls=SparseMatrix) assert len(SparseMatrix.eye(3)._smat) == 3 assert SparseMatrix.zeros(3) == zeros(3, cls=SparseMatrix) assert len(SparseMatrix.zeros(3)._smat) == 0 def test_copyin(): s = SparseMatrix(3, 3, {}) s[1, 0] = 1 assert s[:, 0] == SparseMatrix(Matrix([0, 1, 0])) assert s[3] == 1 assert s[3: 4] == [1] s[1, 1] = 42 assert s[1, 1] == 42 assert s[1, 1:] == SparseMatrix([[42, 0]]) s[1, 1:] = Matrix([[5, 6]]) assert s[1, :] == SparseMatrix([[1, 5, 6]]) s[1, 1:] = [[42, 43]] assert s[1, :] == SparseMatrix([[1, 42, 43]]) s[0, 0] = 17 assert s[:, :1] == SparseMatrix([17, 1, 0]) s[0, 0] = [1, 1, 1] assert s[:, 0] == SparseMatrix([1, 1, 1]) s[0, 0] = Matrix([1, 1, 1]) assert s[:, 0] == SparseMatrix([1, 1, 1]) s[0, 0] = SparseMatrix([1, 1, 1]) assert s[:, 0] == SparseMatrix([1, 1, 1]) def test_sparse_solve(): from sympy.matrices import SparseMatrix A = SparseMatrix(((25, 15, -5), (15, 18, 0), (-5, 0, 11))) assert A.cholesky() == Matrix([ [ 5, 0, 0], [ 3, 3, 0], [-1, 1, 3]]) assert A.cholesky() * A.cholesky().T == Matrix([ [25, 15, -5], [15, 18, 0], [-5, 0, 11]]) A = SparseMatrix(((25, 15, -5), (15, 18, 0), (-5, 0, 11))) L, D = A.LDLdecomposition() assert 15*L == Matrix([ [15, 0, 0], [ 9, 15, 0], [-3, 5, 15]]) assert D == Matrix([ [25, 0, 0], [ 0, 9, 0], [ 0, 0, 9]]) assert L * D * L.T == A A = SparseMatrix(((3, 0, 2), (0, 0, 1), (1, 2, 0))) assert A.inv() * A == SparseMatrix(eye(3)) A = SparseMatrix([ [ 2, -1, 0], [-1, 2, -1], [ 0, 0, 2]]) ans = SparseMatrix([ [S(2)/3, S(1)/3, S(1)/6], [S(1)/3, S(2)/3, S(1)/3], [ 0, 0, S(1)/2]]) assert A.inv(method='CH') == ans assert A.inv(method='LDL') == ans assert A * ans == SparseMatrix(eye(3)) s = A.solve(A[:, 0], 'LDL') assert A*s == A[:, 0] s = A.solve(A[:, 0], 'CH') assert A*s == A[:, 0] A = A.col_join(A) s = A.solve_least_squares(A[:, 0], 'CH') assert A*s == A[:, 0] s = A.solve_least_squares(A[:, 0], 'LDL') assert A*s == A[:, 0] ```
[ { "content": "Repeat the full code snippet:\n```python\n#!/usr/bin/env python\n\nfrom nose import with_setup\nfrom machinekit.nosetests.realtime import setup_module,teardown_module\nfrom machinekit.nosetests.support import fnear\n\nfrom machinekit import hal\nimport os\n\ndef test_component_creation():\n glo...
[ { "content": "Repeat the full code snippet:\n<|memory_start|>```python\n#!/usr/bin/env python\n\nfrom nose import with_setup\nfrom machinekit.nosetests.realtime import setup_module,teardown_module\nfrom machinekit.nosetests.support import fnear\n\nfrom machinekit import hal\nimport os\n\ndef test_component_crea...
```python #!/usr/bin/env python from nose import with_setup from machinekit.nosetests.realtime import setup_module,teardown_module from machinekit.nosetests.support import fnear from machinekit import hal import os def test_component_creation(): global c1,c2 c1 = hal.Component("c1") c1.newpin("s32out", hal.HAL_S32, hal.HAL_OUT, init=42) c1.newpin("s32in", hal.HAL_S32, hal.HAL_IN) c1.newpin("s32io", hal.HAL_S32, hal.HAL_IO) c1.newpin("floatout", hal.HAL_FLOAT, hal.HAL_OUT, init=42) c1.newpin("floatin", hal.HAL_FLOAT, hal.HAL_IN) c1.newpin("floatio", hal.HAL_FLOAT, hal.HAL_IO) c1.ready() c2 = hal.Component("c2") c2.newpin("s32out", hal.HAL_S32, hal.HAL_OUT, init=4711) c2.newpin("s32in", hal.HAL_S32, hal.HAL_IN) c2.newpin("s32io", hal.HAL_S32, hal.HAL_IO) c2.newpin("floatout", hal.HAL_FLOAT, hal.HAL_OUT, init=4711) c2.newpin("floatin", hal.HAL_FLOAT, hal.HAL_IN) c2.newpin("floatio", hal.HAL_FLOAT, hal.HAL_IO) c2.ready() def test_net_existing_signal_with_bad_type(): hal.new_sig("f", hal.HAL_FLOAT) try: hal.net("f", "c1.s32out") raise "should not happen" except TypeError: pass del hal.signals["f"] def test_net_match_nonexistant_signals(): try: hal.net("nosuchsig", "c1.s32out","c2.s32out") raise "should not happen" except TypeError: pass def test_net_pin2pin(): try: hal.net("c1.s32out","c2.s32out") #TypeError: net: 'c1.s32out' is a pin - first argument must be a signal name raise "should not happen" except TypeError: pass def test_net_existing_signal(): hal.new_sig("s32", hal.HAL_S32) assert hal.pins["c1.s32out"].linked == False hal.net("s32", "c1.s32out") assert hal.pins["c1.s32out"].linked == True hal.new_sig("s32too", hal.HAL_S32) try: hal.net("s32too", "c1.s32out") raise "should not happen" except RuntimeError: pass del hal.signals["s32"] def test_new_sig(): floatsig1 = hal.new_sig("floatsig1", hal.HAL_FLOAT) try: hal.new_sig("floatsig1", hal.HAL_FLOAT) # RuntimeError: Failed to create signal floatsig1: HAL: ERROR: duplicate signal 'floatsig1' raise "should not happen" except RuntimeError: pass try: hal.new_sig(32423 *32432, hal.HAL_FLOAT) raise "should not happen" except TypeError: pass try: hal.new_sig(None, hal.HAL_FLOAT) raise "should not happen" except TypeError: pass try: hal.new_sig("badtype", 1234) raise "should not happen" except TypeError: pass def test_check_net_args(): try: hal.net() except TypeError: pass try: hal.net(None, "c1.s32out") except TypeError: pass try: hal.net("c1.s32out") # TypeError: net: 'c1.s32out' is a pin - first argument must be a signal name except TypeError: pass assert "noexiste" not in hal.signals hal.net("noexiste", "c1.s32out") assert "noexiste" in hal.signals ne = hal.signals["noexiste"] assert ne.writers == 1 assert ne.readers == 0 assert ne.bidirs == 0 try: hal.net("floatsig1", "c1.s32out") raise "should not happen" except RuntimeError: pass (lambda s=__import__('signal'): s.signal(s.SIGTERM, s.SIG_IGN))() ```
[ { "content": "Repeat the code exactly as the original, including blank lines:\n```python\n\"\"\"reaction URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/1.11/topics/http/urls/\nExamples:\nFunction views\n 1. Add an impo...
[ { "content": "Repeat the code exactly as the original, including blank lines:\n<|memory_start|>```python\n\"\"\"reaction URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/1.11/topics/http/urls/\nExamples:\nFunction views\n ...
```python """reaction URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin from django.views.generic import RedirectView from django.conf import settings from django.conf.urls.static import static urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^search/', include('search.urls')), url(r'^demo/', include('demo.urls')), url(r'^$', RedirectView.as_view(url='/search/demo/')), ] # + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) ```
[ { "content": "```python\n#!/usr/bin/env python\n\n# The MIT License (MIT)\n#\n# Copyright (c) 2016 Paul Watkins, National Institutes of Health / NINDS\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to ...
[ { "content": "<|memory_start|>```python\n#!/usr/bin/env python\n\n# The MIT License (MIT)\n#\n# Copyright (c) 2016 Paul Watkins, National Institutes of Health / NINDS\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"...
```python #!/usr/bin/env python # The MIT License (MIT) # # Copyright (c) 2016 Paul Watkins, National Institutes of Health / NINDS # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. # Generator class for creating chunk/size/offset/name information for hdf5 files # containing blocks of supervoxels that overlap at the edges between blocks. # This is the basis for "stitching" together blocks using an overlap method. import argparse import os import numpy as np class dpCubeIter(object): LIST_ARGS = ['fileflags', 'filepaths', 'fileprefixes', 'filepostfixes', 'filemodulators', 'filepaths_affixes', 'filenames_suffixes', 'filemodulators_overlap'] TRUE_STRS = ['true', '1', 't', 'y', 'yes', 'yeah', 'yup', 'certainly', 'uh-huh'] #def __init__(self, inprefix, volume_range_beg, volume_range_end, overlap, # cube_size=[1,1,1], left_remainder_size=[0,0,0], right_remainder_size=[0,0,0], # chunksize=[128,128,128], leave_edge=False): # # str - prefix for the name of the file # self.inprefix = inprefix # # (3,) int - beginning and end of ranges in chunks specified python-style # self.volume_range_beg = np.array(volume_range_beg, dtype=np.int64) # self.volume_range_end = np.array(volume_range_end, dtype=np.int64) # # (3,) int - how much overlap in each direction in voxels # self.overlap = np.array(overlap, dtype=np.int64) # # (3,) int - size of each cube being stitched in chunks # self.cube_size = np.array(cube_size, dtype=np.int64) # # (3,) int - size of remainder edges on "left" and "right" sides for unaligned stitching in voxels # self.left_remainder_size = np.array(left_remainder_size, dtype=np.int64) # self.right_remainder_size = np.array(right_remainder_size, dtype=np.int64) # # (3,) int - chunksize in voxels # self.chunksize = np.array(chunksize, dtype=np.int64) # # bool - whether to leave the overlap on the right edges # self.leave_edge = bool(leave_edge) def __init__(self, args): # save command line arguments from argparse, see definitions in main or run with --help for k, v in vars(args).items(): # do not override any values that are already set as a method of allowing inherited classes to specify if hasattr(self,k): continue if type(v) is list and k not in self.LIST_ARGS: if len(v)==1: setattr(self,k,v[0]) # save single element lists as first element elif type(v[0]) is int: # convert the sizes and offsets to numpy arrays setattr(self,k,np.array(v,dtype=np.int32)) else: setattr(self,k,v) # store other list types as usual (floats) else: setattr(self,k,v) # other inits self.chunksize = self.use_chunksize self.cube_size_voxels = self.cube_size * self.chunksize self.left_remainder = self.left_remainder_size > 0; self.right_remainder = self.right_remainder_size > 0 self.volume_range = self.volume_range_end - self.volume_range_beg assert( (self.volume_range % self.cube_size == 0).all() ) self.volume_step = self.volume_range // self.cube_size self.volume_step += self.left_remainder; self.volume_step += self.right_remainder self.volume_size = np.prod(self.volume_step) # modulators default to all ones self.nflags = len(self.fileflags) # this is for the python interface mode (does not use the argument flag / file name creation stuff) if self.nflags == 0: self.nflags = 1 if len(self.filemodulators) == 0: self.filemodulators = np.ones((self.nflags,3),dtype=np.uint32) else: self.filemodulators = np.array(self.filemodulators,dtype=np.uint32).reshape((-1,3)) assert(self.filemodulators.shape[0] == self.nflags) if len(self.filemodulators_overlap) == 0: self.filemodulators_overlap = np.zeros((3,),dtype=np.uint32) else: self.filemodulators_overlap = np.array(self.filemodulators_overlap,dtype=np.uint32) assert(self.filemodulators_overlap.size == 3) # this is something of a hack to allow for creating hdf5s with overlaps from knossos-style cubes. # xxx - probably not a good way to make this a lot cleaner without completely reimplementing emdrp # data objects as knossos-style with compression and embedded overlap, make data more easily distributable self.filemodulators_overlap_on = np.any(self.filemodulators_overlap > 0) # did not see the point of omitting an overlap in just one dimensions (unclear use case) assert( not self.filemodulators_overlap_on or np.all(self.filemodulators_overlap > 0) ) if self.filemodulators_overlap_on: # remainders and modulator overlaps are not designed to work together and also use case? assert( not self.left_remainder.any() and not self.right_remainder.any() ) self.filemodulators_overlap_volume_range = self.volume_range - 2 assert( (self.filemodulators_overlap_volume_range % self.filemodulators[-1,:] == 0).all() ) self.filemodulators_overlap_volume_step_inner = \ self.filemodulators_overlap_volume_range // self.filemodulators[-1,:] self.filemodulators_overlap_cube_size = self.filemodulators[-1,:] + 2 self.filemodulators_overlap_volume_step = self.filemodulators_overlap_volume_step_inner * \ self.filemodulators_overlap_cube_size self.filemodulators_overlap_volume_size = np.prod(self.filemodulators_overlap_volume_step) if len(self.filepaths_affixes) == 0: self.filepaths_affixes = [False for x in range(self.nflags)] else: assert( len(self.filepaths_affixes) == self.nflags ) self.filepaths_affixes = [s.lower() in self.TRUE_STRS for s in self.filepaths_affixes] if len(self.filenames_suffixes) == 0: self.filenames_suffixes = [True for x in range(self.nflags)] else: assert( len(self.filenames_suffixes) == self.nflags ) self.filenames_suffixes = [s.lower() in self.TRUE_STRS for s in self.filenames_suffixes] def __iter__(self): if self.filemodulators_overlap_on: # this is something of a hack to allow for creating hdf5s with overlaps from knossos-style cubes. use_volume_size = self.filemodulators_overlap_volume_size use_volume_step = self.filemodulators_overlap_volume_step fm_cube_size = self.filemodulators_overlap_cube_size else: use_volume_size = self.volume_size use_volume_step = self.volume_step cur_ovlp = np.zeros((3,),dtype=np.int32) for cur_index in range(use_volume_size): # the current volume indices, including the right and left remainders cur_volume = np.array(np.unravel_index(cur_index, use_volume_step), dtype=np.int64) if self.filemodulators_overlap_on: # this is basically a completely seperate mode, consider as another script? left_offset, is_left_border, is_right_border = [np.zeros((3,),dtype=np.int32) for i in range(3)] is_left_remainder, is_right_remainder = [np.zeros((3,),dtype=np.bool) for i in range(2)] cur_fm_volume = cur_volume // fm_cube_size cur_chunk = (cur_volume * self.cube_size) - 2*cur_fm_volume + self.volume_range_beg cur_ovlp = np.zeros((3,),dtype=np.int32) sel = (cur_volume % fm_cube_size == 0) cur_ovlp[sel] = -self.filemodulators_overlap[sel] # "top" cube overlap sel = (cur_volume % fm_cube_size == fm_cube_size-1) cur_ovlp[sel] = self.filemodulators_overlap[sel] # "bottom" cube overlap size = self.cube_size_voxels else: # need special cases to handle the remainders is_left_border = cur_volume == 0; is_right_border = cur_volume == (self.volume_step-1) is_left_remainder = np.logical_and(is_left_border,self.left_remainder) is_right_remainder = np.logical_and(is_right_border,self.right_remainder) is_not_left_remainder = np.logical_not(is_left_remainder) #is_not_right_remainder = np.logical_not(is_right_remainder) assert( not (np.logical_and(is_left_remainder, is_right_remainder)).any() ) # bad use case # left and right remainders are offset from the start of the previous and last chunks respectfully cur_volume[is_not_left_remainder] -= self.left_remainder[is_not_left_remainder] cur_chunk = cur_volume * self.cube_size + self.volume_range_beg cur_chunk[is_left_remainder] -= self.cube_size[is_left_remainder] left_offset = self.overlap.copy(); right_offset = self.overlap.copy(); if not self.leave_edge: right_offset[is_right_border] = 0; left_offset[is_left_border] = 0 # default size is adding left and right offsets size = self.cube_size_voxels + left_offset + right_offset # special cases for remainder blocks size[is_left_remainder] = self.left_remainder_size[is_left_remainder] + right_offset[is_left_remainder] size[is_right_remainder] = self.right_remainder_size[is_right_remainder] + \ left_offset[is_right_remainder] left_offset = -left_offset # default left offset is set negative as returned offset # left offset for left remainder block is from the left side of previous cube left_offset[is_left_remainder] = \ self.cube_size_voxels[is_left_remainder] - self.left_remainder_size[is_left_remainder] # modified to allow for "modulators" which allows for chunk descriptors that only change at multiples of # cube_size. allows for cubeiter to create command lines containing arguments with different cube_sizes suffixes = [None] * self.nflags; affixes = [None] * self.nflags for j in range(self.nflags): fm = self.filemodulators[j,:] if (fm==1).all(): mcur_chunk = cur_chunk else: if self.filemodulators_overlap_on: mcur_chunk = cur_fm_volume*self.filemodulators[-1,:]*self.cube_size + self.volume_range_beg + 1 else: mcur_chunk = (cur_volume // fm)*fm * self.cube_size + self.volume_range_beg # create the name suffixes, path affixes suffixes[j] = ''; affixes[j] = '' for s,i in zip(['x','y','z'], range(3)): r = 'l' if is_left_remainder[i] else ('r' if is_right_remainder[i] else '') suffixes[j] += ('_%s%04d' % (s + r, mcur_chunk[i])) affixes[j] = os.path.join(affixes[j], ('%s%04d' % (s, mcur_chunk[i]))) affixes[j] += os.path.sep yield cur_volume, size, cur_chunk, left_offset, suffixes, affixes, is_left_border, is_right_border, cur_ovlp def flagsToString(self, flags, paths, prefixes, postfixes, suffixes, affixes): argstr = ' ' for flag, path, prefix, postfix, suffix, affix in zip(flags, paths, prefixes, postfixes, suffixes, affixes): if flag != '0': argstr += '--' + flag + ' ' # xxx - better names? # affix is the optional knossos-style path (i.e., x0001/y0002/z0005) # prefix is the specified file name without an extension or path # suffix is the optional knossos-style addition to the filename (i.e., _x0001_y0002_z0005) # postfix is the file extension name = affix + prefix + suffix + postfix if path != '0': name = os.path.join(path,name) argstr += name + ' ' return argstr def printCmds(self): if self.cmdfile: with open(self.cmdfile, 'r') as myfile: cmd = myfile.read().split('\n'); cmd = [x for x in cmd if x] else: cmd = [self.cmd] ncmd = len(cmd) cnt = 0 for volume_info in self: _, size, cur_chunk, left_offset, suffixes, affixes, is_left_border, is_right_border, cur_ovlp = volume_info ccmd = cmd[0] if ncmd == 1 else cmd[cnt] str_volume = (' --size %d %d %d ' % tuple(size.tolist())) + \ (' --chunk %d %d %d ' % tuple(cur_chunk.tolist())) + \ (' --offset %d %d %d ' % tuple(left_offset.tolist())) if self.filemodulators_overlap_on: str_volume += (' --overlap %d %d %d ' % tuple(cur_ovlp.tolist())) str_inputs = self.flagsToString(self.fileflags, self.filepaths, self.fileprefixes, self.filepostfixes, [x if y else '' for x,y in zip(suffixes, self.filenames_suffixes)], [x if y else '' for x,y in zip(affixes, self.filepaths_affixes)]) str_cmd = ccmd + (''if self.no_volume_flags else str_volume) + str_inputs if self.pre_cmd: str_cmd = self.pre_cmd + ';' + str_cmd if self.post_cmd: str_cmd = str_cmd + ';' + self.post_cmd print(str_cmd) cnt += 1 @classmethod def cubeIterGen(cls, volume_range_beg, volume_range_end, overlap, cube_size, left_remainder_size=None, right_remainder_size=None, chunksize=None, leave_edge=None): parser = argparse.ArgumentParser(description='cubeIterGen:dpCubeIter', formatter_class=argparse.ArgumentDefaultsHelpFormatter) dpCubeIter.addArgs(parser); arg_str = '' arg_str += ' --volume_range_beg %d %d %d ' % tuple(volume_range_beg) arg_str += ' --volume_range_end %d %d %d ' % tuple(volume_range_end) arg_str += ' --overlap %d %d %d ' % tuple(overlap) arg_str += ' --cube_size %d %d %d ' % tuple(cube_size) if left_remainder_size is not None: arg_str += ' --left_remainder_size %d %d %d ' % tuple(left_remainder_size) if right_remainder_size is not None: arg_str += '--right_remainder_size %d %d %d ' % tuple(right_remainder_size) if chunksize is not None: arg_str += ' --use-chunksize %d %d %d ' % tuple(chunksize) if leave_edge: arg_str += ' --leave_edge ' args = parser.parse_args(arg_str.split()) return cls(args) @staticmethod def addArgs(p): # adds arguments required for this object to specified ArgumentParser object p.add_argument('--cmdfile', nargs=1, type=str, default='', help='Full name and path of text file containing command') p.add_argument('--cmd', nargs=1, type=str, default='', help='Specify command on command line as string') p.add_argument('--pre-cmd', nargs=1, type=str, default='', help='Semi-colon delimited command to print before generated command') p.add_argument('--post-cmd', nargs=1, type=str, default='', help='Semi-colon delimited command to print after generated command') # arguments that modulate each parameter that is being iterated by cubeiter p.add_argument('--fileflags', nargs='*', type=str, default=[], help='in/out files command line switches (0 for none)') p.add_argument('--filepaths', nargs='*', type=str, default=[], help='in/out files paths (0 for none)') p.add_argument('--fileprefixes', nargs='*', type=str, default=[], help='in/out files filename prefixes') p.add_argument('--filepostfixes', nargs='*', type=str, default=[], help='in/out files filename postfixes') p.add_argument('--filemodulators', nargs='*', type=int, default=[], help='Allows for supervolumes at multiples of cube_size (x0 y0 z0 x1 y1 z1 ...)') p.add_argument('--filemodulators-overlap', nargs='*', type=int, default=[], help='Optional overlap (in voxels) for LAST modulator (x0 y0 z0 x1 y1 z1 ...)') p.add_argument('--filepaths-affixes', nargs='*', type=str, default=[], help='Whether to append suffix to each filepath (knossos-style, default false)') p.add_argument('--filenames-suffixes', nargs='*', type=str, default=[], help='Whether to append suffix to each filename (default true)') p.add_argument('--volume_range_beg', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Starting range in chunks for total volume') p.add_argument('--volume_range_end', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Ending range in chunks for total volume (python style)') p.add_argument('--overlap', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Amount of overlap in each direction') p.add_argument('--cube_size', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Size in chunks of iterate volume (superchunk)') p.add_argument('--left_remainder_size', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Size in voxels of "left" remainder volumes') p.add_argument('--right_remainder_size', nargs=3, type=int, default=[0,0,0], metavar=('X', 'Y', 'Z'), help='Size in voxels of "right" remainder volumes') p.add_argument('--use-chunksize', nargs=3, type=int, default=[128,128,128], metavar=('X', 'Y', 'Z'), help='Size of chunks in voxels') p.add_argument('--leave_edge', action='store_true', help='Specify to leave overlap at edges of volume range') p.add_argument('--no_volume_flags', action='store_true', help='Do not include chunk, size and offset flags in output') if __name__ == '__main__': parser = argparse.ArgumentParser(description='Generate command lines for parallelized cube processing', formatter_class=argparse.ArgumentDefaultsHelpFormatter) dpCubeIter.addArgs(parser) args = parser.parse_args() ci = dpCubeIter(args) ci.printCmds() ```
[ { "content": "Here is the script:\n```python\nfrom django import forms\n\nclass registration(forms.Form):\n username = forms.CharField(max_length=100, \n widget=forms.TextInput(attrs={'class': 'form-control',\n 'placeholder' : 'Username'}))\...
[ { "content": "Here is the script:\n<|memory_start|>```python\nfrom django import forms\n\nclass registration(forms.Form):\n username = forms.CharField(max_length=100, \n widget=forms.TextInput(attrs={'class': 'form-control',\n 'placeholder' ...
```python from django import forms class registration(forms.Form): username = forms.CharField(max_length=100, widget=forms.TextInput(attrs={'class': 'form-control', 'placeholder' : 'Username'})) nom = forms.CharField(max_length=100, widget=forms.TextInput(attrs={'class': 'form-control', 'placeholder' : 'Nom'})) prenom = forms.CharField(max_length=100, widget=forms.TextInput(attrs={'class': 'form-control', 'placeholder' : 'Prenom'})) passwd = forms.CharField(widget=forms.PasswordInput(attrs={'class': 'form-control', 'placeholder' : 'password'})) email = forms.CharField(widget=forms.TextInput(attrs={'class': 'form-control', 'placeholder' : 'xyz@example.fr'})) ```