text
stringlengths
1
1.04M
language
stringclasses
25 values
Priyanka Chopra will be next in The White Tiger opposite Rajkummar Rao, which is a Netflix film. It is adapted from Aravind Adiga's prize-winning novel of the same name. Our Desi Girl of Bollywood, Priyanka Chopra has celebrated 20 years in the industry. While she made her big-screen debut in 2002, PeeCee won the Miss World crown in 2000. To celebrate 20 years of illustrous career, Piggy Chops shared a video, where she thanked her fans and said that she will share 20 monumental moments of journey with them. Sharing the video she wrote, "It’s time for a celebration... 2020 marks my 20 years in the entertainment industry! What?! How did that even happen? ? You all have been by my side throughout this journey and your loyalty and support means the world to me! Join me as I take this trip down memory lane and celebrate #20in2020." While the actress made a big name in the west, she recently revealed about facing racism in USA and told PinkVilla, "Of course, my struggles were real; everyone's is. But I choose not to talk about it or be a victim of those situations. I have always learnt one thing from my dad. When you fall down, you pick yourself up and move on. So I keep thinking about the next big thing I could do, next thing I could learn, the next glass ceiling I could break." She further said, "There was one time when America took so much from the country and I was like, 'Get lost, I'm going back to India', because I felt too small as a teenager. But it was also because I was insecure. So now when I went back this time, I was a lot more secure in my own feet. The country has given me respect, my family, my husband, my home and a different career that I started in my 30s." On the work front, Priyanka Chopra will be next in The White Tiger opposite Rajkummar Rao, which is a Netflix film. It is adapted from Aravind Adiga's prize-winning novel of the same name. The film will be helmed by Ramin Bahrani. It follows the extraordinary journey of a self-made man from being a tea-shop worker in a village to a successful entrepreneur in a big city. The global icon also has an international web series under her belt titled Citadel, which will be produced by Avengers: Endgame makers Russo Brothers and also feature Richard Madden in a key role. It will be directed by Raj and DK, who are known for The Family Man, Go Goa Gone and others. Stay tuned to BollywoodLife for the latest scoops and updates from Bollywood, Hollywood, South, TV and Web-Series. Click to join us on Facebook, Twitter, Youtube and Instagram. Also follow us on Facebook Messenger for latest updates.
english
<gh_stars>10-100 --- title: Optimizar los costes mediante la administración automática del ciclo de vida de los datos titleSuffix: Azure Storage description: Use directivas de administración del ciclo de vida de Azure Storage para crear reglas automatizadas para mover datos entre los niveles de acceso frecuente, esporádico o de archivo. author: tamram ms.author: tamram ms.date: 08/18/2021 ms.service: storage ms.subservice: common ms.topic: conceptual ms.reviewer: yzheng ms.custom: devx-track-azurepowershell, references_regions ms.openlocfilehash: 245bcbfd59644946ac6f039e35fe02147054cc8c ms.sourcegitcommit: 613789059b275cfae44f2a983906cca06a8706ad ms.translationtype: HT ms.contentlocale: es-ES ms.lasthandoff: 09/29/2021 ms.locfileid: "129273238" --- # <a name="optimize-costs-by-automatically-managing-the-data-lifecycle"></a>Optimizar los costes mediante la administración automática del ciclo de vida de los datos Los conjuntos de datos tienen ciclos de vida únicos. Al principio del ciclo de vida, las personas acceden con frecuencia a algunos datos. Pero la necesidad de acceso suele descender drásticamente a medida que los datos se hacen más antiguos. Algunos datos permanecen inactivos en la nube y, una vez almacenados, no se suele acceder a ellos. Algunos conjuntos de datos expiran días o meses después de su creación, mientras que otros conjuntos de datos se leen y modifican de forma activa en el transcurso de sus ciclos de vida. La administración del ciclo de vida de Azure Storage ofrece una directiva basada en reglas que se puede usar para trasladar los datos de blob al nivel de acceso adecuado y para hacer que los datos expiren cuando finalice su ciclo de vida. La directiva de administración del ciclo de vida permite hacer lo siguiente: - Pasar blobs de nivel de acceso esporádico a nivel de acceso frecuente en el mismo instante en que se accede a ellos para optimizar el rendimiento. - Pasar blobs, versiones de blobs e instantáneas de blobs a un nivel de almacenamiento de acceso esporádico si no se accede a estos objetos o si se modifican durante un período de tiempo para optimizar el coste. En este escenario, la directiva de administración del ciclo de vida puede mover objetos del nivel de acceso frecuente al esporádico o de archivo y, asimismo, del nivel de acceso esporádico al de archivo. - Eliminar blobs, versiones de blobs e instantáneas de blobs al final de su ciclo de vida. - Definir reglas que se ejecutarán una vez al día en el nivel de cuenta de almacenamiento - Aplicar reglas a contenedores o a un subconjunto de blobs mediante prefijos de nombre o [etiquetas de índice de blobs](storage-manage-find-blobs.md) como filtros. Pensemos en un escenario donde se accede frecuentemente a los datos durante las primeras fases del ciclo de vida, pero solo ocasionalmente al cabo de dos semanas. Transcurrido el primer mes, rara vez se accede al conjunto de datos. En este escenario, es mejor el almacenamiento de acceso frecuente durante las primeras etapas. El almacenamiento de acceso esporádico es más adecuado para un acceso ocasional. El almacenamiento de archivo es la mejor opción de nivel una vez que los datos tengan un mes. Al mover datos al nivel de almacenamiento adecuado en función de su antigüedad mediante las reglas de directivas de administración del ciclo de vida, se puede diseñar la solución más asequible para sus necesidades. Se pueden usar directivas de administración del ciclo de vida con blobs en bloques y blobs en anexos en cuentas de uso general v2, en cuentas de almacenamiento Premium de blobs en bloques y en cuentas de Blob Storage. La administración del ciclo de vida no afecta a los contenedores del sistema como *$logs* o *$web*. > [!IMPORTANT] > Si un conjunto de datos debe ser legible, no establezca una directiva para mover blobs al nivel de archivo, ya que los blobs en este nivel no se pueden leer a menos que se rehidraten antes, un proceso que puede conllevar mucho tiempo y dinero. Para más información, consulte [Introducción a la rehidratación de blobs desde el nivel de archivo](archive-rehydrate-overview.md). ## <a name="lifecycle-management-policy-definition"></a>Definición de la directiva de administración del ciclo de vida Una directiva de administración del ciclo de vida es una colección de reglas en un documento JSON. En el siguiente JSON de ejemplo se muestra una definición de regla completa: ```json { "rules": [ { "name": "rule1", "enabled": true, "type": "Lifecycle", "definition": {...} }, { "name": "rule2", "type": "Lifecycle", "definition": {...} } ] } ``` Una directiva es una colección de reglas, como se describe en la siguiente tabla: | Nombre de parámetro | Tipo de parámetro | Notas | |----------------|----------------|-------| | `rules` | Una matriz de objetos de regla | Se requiere al menos una regla en una directiva. Puede definir hasta 100 reglas en una directiva.| Cada regla de la directiva tiene varios parámetros, descritos en la siguiente tabla: | Nombre de parámetro | Tipo de parámetro | Notas | Obligatorio | |--|--|--|--| | `name` | String | Un nombre de regla puede incluir hasta 256 caracteres alfanuméricos. El nombre de regla distingue mayúsculas de minúsculas. Debe ser único dentro de una directiva. | True | | `enabled` | Boolean | Un valor booleano opcional para permitir que una regla se deshabilite de forma temporal. El valor predeterminado es true si no se establece. | False | | `type` | Un valor de enumeración | El tipo actual válido es `Lifecycle`. | True | | `definition` | Un objeto que define la regla del ciclo de vida | Cada definición se compone de un conjunto de filtros y un conjunto de acciones. | True | ## <a name="lifecycle-management-rule-definition"></a>Definición de regla de administración del ciclo de vida Cada definición de regla incluye un conjunto de filtros y un conjunto de acciones. El [conjunto de filtros](#rule-filters) limita las acciones de regla a un determinado conjunto de objetos dentro de un contenedor o nombres de objetos. El [conjunto de acciones ](#rule-actions) aplica las acciones de nivel o eliminación al conjunto filtrado de objetos. ### <a name="sample-rule"></a>Ejemplo de regla La siguiente regla de ejemplo filtra la cuenta para ejecutar las acciones en objetos que existen dentro de `sample-container` y empiezan por `blob1`. - Establecer el nivel de blob en nivel esporádico 30 días después de la última modificación - Establecer el nivel de blob en nivel de almacenamiento de archivo 90 días después de la última modificación - Eliminar el blob 2555 días (siete años) después de la última modificación - Eliminar las versiones de blobs anteriores 90 días después de su creación ```json { "rules": [ { "enabled": true, "name": "sample-rule", "type": "Lifecycle", "definition": { "actions": { "version": { "delete": { "daysAfterCreationGreaterThan": 90 } }, "baseBlob": { "tierToCool": { "daysAfterModificationGreaterThan": 30 }, "tierToArchive": { "daysAfterModificationGreaterThan": 90 }, "delete": { "daysAfterModificationGreaterThan": 2555 } } }, "filters": { "blobTypes": [ "blockBlob" ], "prefixMatch": [ "sample-container/blob1" ] } } } ] } ``` ### <a name="rule-filters"></a>Filtros de reglas Los filtros limitan las acciones de regla a un subconjunto de blobs dentro de la cuenta de almacenamiento. Si se define más de un filtro, se ejecuta un valor lógico `AND` en todos los filtros. Entre los filtros están los siguientes: | Nombre de filtro | Tipo de filtro | Notas | Es obligatorio | |-------------|-------------|-------|-------------| | blobTypes | Una matriz de valores de enumeración predefinidos. | La versión actual admite `blockBlob` y `appendBlob`. En `appendBlob` solo se puede eliminar, no se puede establecer el nivel. | Sí | | prefixMatch | Una matriz de cadenas de prefijos con los que debe hacer coincidencias. Cada regla puede definir hasta 10 prefijos (con distinción entre mayúsculas y minúsculas). Una cadena de prefijos debe comenzar con el nombre de un contenedor. Por ejemplo, si quiere que todos los blobs de `https://myaccount.blob.core.windows.net/sample-container/blob1/...` coincidan en una regla, prefixMatch es `sample-container/blob1`. | Si no define prefixMatch, la regla se aplica a todos los blobs de la cuenta de almacenamiento. | No | | blobIndexMatch | Una matriz de valores de diccionario que se compone de las condiciones de clave y valor de la etiqueta de índice de blobs con las que debe haber coincidencias. Cada regla puede definir hasta 10 condiciones de etiqueta de índice de blobs. Por ejemplo, si quiere que todos los blobs coincidan con `Project = Contoso` en `https://myaccount.blob.core.windows.net/` en relación a una regla, el valor de blobIndexMatch es `{"name": "Project","op": "==","value": "Contoso"}`. | Si no define blobIndexMatch, la regla se aplica a todos los blobs de la cuenta de almacenamiento. | No | Para obtener más información sobre la característica de índice de blobs, así como sus limitaciones y problemas conocidos, consulte [Administración y búsqueda de datos en Azure Blob Storage con el índice de blobs](storage-manage-find-blobs.md). ### <a name="rule-actions"></a>Acciones de regla Las acciones se aplican a los blobs filtrados cuando se cumple la condición de ejecución. La administración del ciclo de vida admite tanto la organización en niveles como la eliminación de blobs, versiones anteriores de los blobs e instantáneas de blobs. Defina al menos una acción para cada regla en los blobs de base, las versiones anteriores de los blobs o las instantáneas de los blobs. | Acción | Blob de base | Instantánea | Versión |-----------------------------|--------------------------------------------|---------------|---------------| | tierToCool | Se admite para `blockBlob` | Compatible | Compatible | | enableAutoTierToHotFromCool | Se admite para `blockBlob` | No compatible | No compatible | | tierToArchive | Se admite para `blockBlob` | Compatible | Compatible | | delete | Compatible en `blockBlob` y `appendBlob`. | Compatible | Compatible | > [!NOTE] > Si define más de una acción en el mismo blob, la administración del ciclo de vida aplica la acción menos cara al blob. Por ejemplo, la acción `delete` es más económica que la acción `tierToArchive`. La acción `tierToArchive` es más económica que la acción `tierToCool`. Las condiciones de ejecución se basan en la antigüedad. Para realizar el seguimiento de la antigüedad, los blobs de base usan la hora de la última modificación, las versiones de los blobs usan la hora de creación de la versión y las instantáneas de los blobs usan la hora de creación de la instantánea. | Condición de ejecución de acción | Valor de la condición | Descripción | |--|--|--| | daysAfterModificationGreaterThan | Valor entero que indica la antigüedad en días | Condición de las acciones de blob de base | | daysAfterCreationGreaterThan | Valor entero que indica la antigüedad en días | La condición de la versión del blob y de las acciones de la instantánea del blob | | daysAfterLastAccessTimeGreaterThan | Valor entero que indica la antigüedad en días | Condición de las acciones del blob base cuando el seguimiento de acceso está habilitado | ## <a name="examples-of-lifecycle-policies"></a>Ejemplos de directivas de ciclo de vida Los ejemplos siguientes muestran cómo abordar escenarios comunes con las reglas de directivas del ciclo de vida. ### <a name="move-aging-data-to-a-cooler-tier"></a>Cambio de los datos antiguos a un nivel de acceso más esporádico En este ejemplo se muestra cómo realizar la transición de blobs en bloques con el prefijo `sample-container/blob1` o `container2/blob2`. La directiva realiza la transición de los blobs que no se han modificado durante más de 30 días al almacenamiento de acceso esporádico, y los blobs no modificados en 90 días al nivel de acceso de archivo: ```json { "rules": [ { "name": "agingRule", "enabled": true, "type": "Lifecycle", "definition": { "filters": { "blobTypes": [ "blockBlob" ], "prefixMatch": [ "sample-container/blob1", "container2/blob2" ] }, "actions": { "baseBlob": { "tierToCool": { "daysAfterModificationGreaterThan": 30 }, "tierToArchive": { "daysAfterModificationGreaterThan": 90 } } } } } ] } ``` ### <a name="move-data-based-on-last-accessed-time"></a>Mover datos en función de la hora de último acceso Se puede habilitar el seguimiento de la hora de último acceso para mantener un registro de cuándo se lee o escribe por última vez el blob y también como filtro para administrar los niveles y la retención de los datos de blob. Para obtener información sobre cómo habilitar el seguimiento de la hora de último acceso, consulte [Habilitar el seguimiento de hora de acceso opcionalmente](lifecycle-management-policy-configure.md#optionally-enable-access-time-tracking). Si se habilita el seguimiento de la hora del último acceso, la propiedad de blob denominada `LastAccessTime` se actualiza cuando se lee o se escribe en un blob. Una operación [Get Blob](/rest/api/storageservices/get-blob) se considera una operación de acceso. [Get Blob Properties](/rest/api/storageservices/get-blob-properties), [Get Blob Metadata](/rest/api/storageservices/get-blob-metadata) y [Get Blob Tags](/rest/api/storageservices/get-blob-tags) no son operaciones de acceso y, por lo tanto, no actualizan la hora del último acceso. Para reducir el impacto en la latencia del acceso de lectura, solo la primera lectura de las últimas 24 horas actualiza la hora del último acceso. Las lecturas posteriores en el mismo período de 24 horas no la actualizan. Si se modifica un blob entre lecturas, la hora del último acceso es la más reciente de los dos valores. En el siguiente ejemplo, los blobs se mueven al almacenamiento esporádico si no se ha accedido a ellos durante 30 días. La propiedad `enableAutoTierToHotFromCool` es un valor booleano que indica si un blob debe pasar automáticamente del nivel de acceso esporádico al nivel de acceso frecuente si se vuelve a acceder a él una vez que se haya pasado al nivel de acceso esporádico. ```json { "enabled": true, "name": "last-accessed-thirty-days-ago", "type": "Lifecycle", "definition": { "actions": { "baseBlob": { "enableAutoTierToHotFromCool": true, "tierToCool": { "daysAfterLastAccessTimeGreaterThan": 30 } } }, "filters": { "blobTypes": [ "blockBlob" ], "prefixMatch": [ "mylifecyclecontainer/log" ] } } } ``` ### <a name="archive-data-after-ingest"></a>Archivado de datos después de la ingesta Algunos datos permanecen inactivos en la nube y no se accede a ellos prácticamente nunca. La siguiente directiva del ciclo de vida está configurada para archivar los datos poco después de que se ingieran. En este ejemplo se realiza la transición de blobs en bloques de un contenedor denominado `archivecontainer` a un nivel de archivo. La transición se realiza al actuar en los blobs 0 días después de la hora de la última modificación: ```json { "rules": [ { "name": "archiveRule", "enabled": true, "type": "Lifecycle", "definition": { "filters": { "blobTypes": [ "blockBlob" ], "prefixMatch": [ "archivecontainer" ] }, "actions": { "baseBlob": { "tierToArchive": { "daysAfterModificationGreaterThan": 0 } } } } } ] } ``` > [!NOTE] > Microsoft recomienda cargar los blobs directamente en el nivel de archivo para lograr una mayor eficacia. El nivel de archivo se puede especificar en el encabezado *x-ms-access-tier* de las operaciones [Put Blob](/rest/api/storageservices/put-blob) o [Put Block List](/rest/api/storageservices/put-block-list). El encabezado *x-ms-access-tier* se puede usar con la versión de REST 2018-11-09 y versiones más recientes o con las bibliotecas cliente de Blob Storage más recientes. ### <a name="expire-data-based-on-age"></a>Expiración de datos en función de la antigüedad Se espera que algunos datos expiren días o meses después de la creación. Puede configurar una directiva de administración del ciclo de vida para que los datos expiren mediante eliminación en función de su antigüedad. En el ejemplo siguiente se muestra una directiva que elimina todos los blobs en bloques con una antigüedad superior a 365 días. ```json { "rules": [ { "name": "expirationRule", "enabled": true, "type": "Lifecycle", "definition": { "filters": { "blobTypes": [ "blockBlob" ] }, "actions": { "baseBlob": { "delete": { "daysAfterModificationGreaterThan": 365 } } } } } ] } ``` ### <a name="delete-data-with-blob-index-tags"></a>Eliminar datos con etiquetas de índice de blobs Algunos datos solo deben expirar si se marcan explícitamente para su eliminación. Puede configurar una directiva de administración del ciclo de vida para que expiren los datos etiquetados con los atributos de clave/valor del índice de blobs. En el ejemplo siguiente se muestra una directiva que elimina todos los blobs en bloques con `Project = Contoso`. Para obtener más información sobre el índice de blobs, consulte [Administración y búsqueda de datos en Azure Blob Storage con el Índice de blobs](storage-manage-find-blobs.md). ```json { "rules": [ { "enabled": true, "name": "DeleteContosoData", "type": "Lifecycle", "definition": { "actions": { "baseBlob": { "delete": { "daysAfterModificationGreaterThan": 0 } } }, "filters": { "blobIndexMatch": [ { "name": "Project", "op": "==", "value": "Contoso" } ], "blobTypes": [ "blockBlob" ] } } } ] } ``` ### <a name="manage-versions"></a>Administración de versiones En el caso de datos que se modifican y a los que se accede de forma regular a lo largo de toda su duración, puede habilitar el control de versiones de Blob Storage para mantener de forma automática las versiones anteriores de un objeto. Puede crear una directiva para organizar en niveles o eliminar las versiones anteriores. La antigüedad de la versión se determina mediante la evaluación de la hora de creación de la misma. Esta regla de directiva crea niveles de las versiones anteriores en el contenedor `activedata` que sean 90 días, o más, posteriores a la creación de la versión en el nivel de acceso esporádico, y elimina las versiones anteriores que tengan 365 días, o más. ```json { "rules": [ { "enabled": true, "name": "versionrule", "type": "Lifecycle", "definition": { "actions": { "version": { "tierToCool": { "daysAfterCreationGreaterThan": 90 }, "delete": { "daysAfterCreationGreaterThan": 365 } } }, "filters": { "blobTypes": [ "blockBlob" ], "prefixMatch": [ "activedata" ] } } } ] } ``` ## <a name="feature-support"></a>Compatibilidad de características En esta tabla se muestra cómo se admite esta característica en la cuenta y el impacto en la compatibilidad al habilitar determinadas funcionalidades. | Tipo de cuenta de almacenamiento | Blob Storage (compatibilidad predeterminada) | Data Lake Storage Gen2 <sup>1</sup> | NFS 3.0 <sup>1</sup> |-----------------------------|---------------------------------|------------------------------------|--------------------------------------------------| | De uso general estándar, v2 | ![Sí](../media/icons/yes-icon.png) |![Sí](../media/icons/yes-icon.png) | ![Sí](../media/icons/yes-icon.png) | | Blobs en bloques Premium | ![Sí](../media/icons/yes-icon.png)|![Sí](../media/icons/yes-icon.png) | ![Sí](../media/icons/yes-icon.png) | <sup>1</sup> Tanto Data Lake Storage Gen2 como el protocolo Network File System (NFS) 3.0 necesitan una cuenta de almacenamiento con un espacio de nombres jerárquico habilitado. ## <a name="regional-availability-and-pricing"></a>Disponibilidad regional y precios La característica de administración del ciclo de vida está disponible en todas las regiones de Azure. Las directivas de administración del ciclo de vida son gratuitas. A los clientes se les cobra el coste operativo estándar derivado de las llamadas API [Set Blob Tier](/rest/api/storageservices/set-blob-tier). Las operaciones de eliminación también son gratuitas. Cada actualización a la hora de último acceso de un blob se factura bajo la categoría [Todas las demás operaciones](https://azure.microsoft.com/pricing/details/storage/blobs/). Para más información sobre los precios, consulte [Precios de los blobs en bloques](https://azure.microsoft.com/pricing/details/storage/blobs/). ## <a name="faq"></a>Preguntas más frecuentes **He creado una directiva, ¿por qué las acciones no se ejecutan inmediatamente?** La plataforma ejecuta la directiva del ciclo de vida una vez al día. Una vez que configure una directiva, algunas acciones pueden tardar hasta 24 horas en ejecutarse por primera vez. **Si actualizo una directiva existente, ¿cuánto tiempo tardan en ejecutarse las acciones?** La directiva actualizada tarda hasta 24 horas en entrar en vigor. Una vez que la directiva está en vigor, las acciones pueden tardar hasta 24 horas en ejecutarse. Por lo tanto, las acciones de la directiva pueden tardar hasta 48 horas en completarse. Si la actualización va a deshabilitar o eliminar una regla y se ha usado enableAutoTierToHotFromCool, se seguirán haciendo niveles automáticos en el nivel de acceso de acceso frecuente. Por ejemplo, establezca una regla que incluya enableAutoTierToHotFromCool en función del último acceso. Si la regla está deshabilitada o eliminada, y un blob se encuentra actualmente en estado de nivel de acceso esporádico y, después, se accede a él, volverá al nivel de acceso frecuente, que es el que se aplica en el acceso fuera de la administración del ciclo de vida. Luego, el blob no pasará de nivel de acceso frecuente a nivel de acceso esporádico, ya que la regla de administración del ciclo de vida está deshabilitada o eliminada. La única manera de evitar autoTierToHotFromCool es desactivar el seguimiento de la hora del último acceso. **He rehidratado manualmente un blob archivado, ¿cómo evito que vuelva temporalmente al nivel de archivo?** Cuando se mueve un blob desde un nivel de acceso a otro, su hora de última modificación no cambia. Si rehidrata manualmente un blob archivado al nivel de acceso frecuente, el motor de administración del ciclo de vida podría devolverlo al nivel de archivo. Deshabilite la regla que afecte temporalmente a este blob para impedir que se vuelva a archivar. Vuelva a habilitar la regla cuando el blob se pueda volver a mover con seguridad al nivel de archivo. También puede copiar el blob en otra ubicación si necesita permanecer en el nivel de acceso frecuente o esporádico de forma permanente. ## <a name="next-steps"></a>Pasos siguientes - [Configurar una directiva de administración del ciclo de vida](lifecycle-management-policy-configure.md) - [Niveles de acceso frecuente, esporádico y de archivo de los datos de blob](access-tiers-overview.md) - [Administración y búsqueda de datos en Azure Blob Storage con el Índice de blobs](storage-manage-find-blobs.md)
markdown
OPEC’s oil production has slumped to the lowest level since 2011 following attacks on the heart of Saudi Arabia’s oil industry, new data shows. The 14-member organization’s monthly total in September was down 750,000 barrels per day (bpd) from August, Reuters reported. A brazen attack by Yemeni forces last month shut down 5. 7 million bpd of Saudi Arabia’s oil production, which represents more than half of the kingdom’s or five percent of global output. It sent crude prices up 20% to $72 a barrel and exposed the vulnerability of the world’s largest oil exporter to a new situation. Energy analysts have said the raid was akin to a massive heart attack for the oil market and global economy. The global oil market is already tight because of unilateral US sanctions on Iran and Venezuela. In September, Saudi Arabia supplied 9. 05 million bpd or 700,000 bpd less than in August, according to Reuters. However, the drop was larger which state oil company Aramco limited by releasing stored crude from its inventories, the news agency said.
english
{"name":"<NAME>","dna":[{"code":1.1121,"color":"#A7194B","scale":0.1121},{"code":1.3896,"color":"#FD5308","scale":0.3896},{"code":1.4215,"color":"#FB9902","scale":0.4215},{"code":1.4808,"color":"#FABC02","scale":0.4808},{"code":1.9661,"color":"#D0EA2B","scale":0.9661},{"code":1.0751,"color":"#66B032","scale":0.0751},{"code":2.969,"color":"#FEFE33","scale":0.969},{"code":1.2383,"color":"#3D01A4","scale":0.2383},{"code":2.6019,"color":"#FEFE33","scale":0.6019}],"attributes":[{"trait_type":"Speed","value":2},{"trait_type":"Stamina","value":5},{"trait_type":"Strength","value":5},{"trait_type":"Aggression","value":5},{"trait_type":"Creativity","value":10},{"trait_type":"Luck","value":2},{"trait_type":"Focus","value":10},{"trait_type":"Influence","value":3},{"trait_type":"Agility","value":6},{"trait_type":"Phobia","value":"None"},{"trait_type":"Vice","value":"None"},{"trait_type":"Role","value":"Superfan"},{"trait_type":"Personality","value":"Modest"},{"trait_type":"Class","value":"Die-hard"},{"trait_type":"Affinity","value":"Mind"}],"image":""}
json
<gh_stars>0 {% extends 'accounting/base.html' %} {% block breadcrumbs %} {% endblock %} {% block mainContainer %} {% endblock %}
html
The Congress’ MP from Thiruvananthapuram opposed the Centre’s bid to make Hindi an official language at the United Nations. Shashi Tharoor, the Congress’ MP from Thiruvananthapuram, on Wednesday asked why India should make Hindi one of the languages spoken at the United Nations, PTI reported. Tharoor made the remark after two Bharatiya Janata Party MPs asked External Affairs Minister Sushma Swaraj in Parliament about the steps taken to make Hindi an official language at the UN. “If tomorrow someone from Tamil Nadu or from West Bengal becomes the prime minister, why should we force him to speak in Hindi at the UN?” Tharoor asked during Question Hour in the Lok Sabha. He also said that India is the only country where Hindi was an official language. Responding to Tharoor and the BJP MPs, Swaraj said Hindi was an official language even in Fiji, and was spoken widely in Mauritius, Suriname, Trinidad and Tobago and many other countries. However, she said United Nations rules do not allow Hindi to be made an official language at the world body. She said that according to the UN’s rules, the motion to make a language official has to be supported by two-thirds of the body’s members (that is, by 129 out of 193 countries). All member nations have to bear the expenses of making Hindi an official UN language, she said. “It is not difficult to get the support of two-third member nations. But when the issue of bearing the expenses comes, many small nations become hesitant,” the minister added. “This is a big hurdle in making Hindi an official language at the UN”. An unidentified BJP MP said India would have to pay Rs 40 crore to the UN make Hindi one of the official languages. “We are ready to pay even Rs 400 crore if required,” Swaraj said. However, she added that the global body’s rules do not permit such payments.
english
Every year the 5th of September, the birthday of Dr Sarvepalli Radhakrishnan, our second President, gives us an opportunity to think about teachers; to reminisce our very own favourite teachers who have touched our lives in a very special way and to ponder on the role played by teachers in the moulding of humane men and women, especially in the strife - ridden, commercialized dog-eat-dog world we are living in today. What I’ll try to do here is pen down a few random thoughts that inevitably cross my mind when I think aloud on the efficacies of what has been and will continue to be my chosen vocation. A teacher is essentially a ‘Sakalakalavallabhan”, a veritable all rounder, a jack of all trades and master of all, someone who makes an indelible impact on almost all aspects of his students’ lives. The old adage aptly proclaims that the mediocre teacher tells, the good teacher explains, the superior teacher demonstrates and the great teacher inspires. Teaching is a ‘calling’ and all teachers are missionaries with the common passion to make their students ‘men and women for others. ’ As we celebrate Teachers’ Day, the evergreen Latin proverb inspires all of us teachers to more ahead in our chosen profession with renewed vigour and zest - “By learning you will teach; by teaching you will learn”. I’d like to round off my ruminations on teachers by presenting a poem entitled ‘An Ode to Teachers’, written by my daughter Rachel last year. This was published in the Golden Jubilee Souvenir of St Mary’s English Higher Primary School, Falnir. While God is our Creator, And parents are our Procreators, Teachers are undoubtedly our Co-creators. Teachers enter our lives when we are small, And are sensitive to our every beck and call. They mould us like a potter who moulds the clay, And accompany us to ensure that we never stray. Teachers are embodiments of selfless sacrifice, Beacon lights who guide our student lives. They teach us reading, writing and arithmetic, And are adept at using both the carrot and the stick. Teachers, by example, our values fashion, Encourage, empower and channel our passion. When we are hurt, our wounds they dress, When we are troubled, our hearts they bless. Teachers, I believe, are gifts from heaven, I salute my teachers in 2011. Long live beloved teachers, may God bless you, Hold you and keep you, and your families too. Dr Furtado - Archives:
english
<reponame>prophet-2019/ProphetCapstone import React, {Component} from 'react' import HomePageChart from '../HomePageChart' export default class FeaturedChart extends Component { constructor(props) { super(props) } render() { return ( <div className="dashboard-feature"> <HomePageChart /> </div> ) } }
javascript
Dr APJ Abdul Kalam Technical University (AKTU) vice chancellor professor Vineet Kansal reviewed preparations for the term end examinations to be held from December 28. Dr APJ Abdul Kalam Technical University (AKTU) Vice-Chancellor professor Vineet Kansal reviewed preparations for the term end examinations to be held from December 28. During the review meeting, instructions were given to conduct the examinations in a systematic manner. About 1,10,000 students will participate in these examinations at 122 examination centers set up across the state. “The examination center coordinators were ordered to pay special attention to ensure that the students do not face any inconvenience at the examination centre, said Asheesh Misra, media in-charge, AKTU, Lucknow. The superintendents of the examination centers and the directors of the respective institutions said that all the preparations related to the examination have been completed. “The center superintendents assured that the students will not face any inconvenience at the examination center,” Misra said. Controller of examinations professor Anurag Tripathi said that adherence to the Covid-19 protocol will be ensured at all examination centers. For this, the affiliated institutions have been directed to ensure thermal checking, sanitisation, masks etc. Exam center coordinators said that elaborate arrangements have been made at the center to ensure complete adherence to the COVID-19 protocol.
english
<reponame>achooan/simple-ecommerce const { describe, it } = require('mocha'); const mongoose = require('mongoose'); const { request } = require('../context'); describe('Carts API responses', () => { it('<200> GET /carts', (done) => { request.get('/carts') .expect(200) .end(done); }); it('<302> GET /carts/clear', (done) => { request.get('/carts/clear') .expect(302) .expect('Location', '/') // redirect back .end(done); }); it('<302> POST /carts/checkout', (done) => { request.post('/carts/checkout') .expect(302) .expect('Location', '/orders/checkout') .end(done); }); it('<302> POST /carts/1001', (done) => { const form = { option: 'default', productId: mongoose.Types.ObjectId(), quantity: 3, }; request.post('/carts/1001') .send(form) .expect(302) .expect('Location', '/') .end(done); }); });
javascript
In this Molly Aunty Rocks film, Prithviraj Sukumaran, Revathi played the primary leads. The Molly Aunty Rocks was released in theaters on 14 Sep 2012. Movies like Jeenthoal, Madhura Manohara Moham, Anveshippin Kandethum and others in a similar vein had the same genre but quite different stories. The soundtracks and background music were composed by Anand Madhusoodhanan for the movie Molly Aunty Rocks. The movie Molly Aunty Rocks belonged to the Comedy, genre.
english
<filename>unidbg-android/src/test/java/u/aly/cv.java package u.aly; /* compiled from: BL */ public final class cv { public static final byte a = 0; public static final byte b = 1; /* renamed from: c reason: collision with root package name */ public static final byte f23680c = 2; public static final byte d = 3; public static final byte e = 4; public static final byte f = 6; public static final byte g = 8; public static final byte h = 10; /* renamed from: i reason: collision with root package name */ public static final byte f23681i = 11; public static final byte j = 12; public static final byte k = 13; /* renamed from: l reason: collision with root package name */ public static final byte f23682l = 14; public static final byte m = 15; public static final byte n = 16; }
java
<gh_stars>1-10 {"web":[{"value":["天主教","天主教","天主教教义"],"key":"Catholicism"},{"value":["民间天主教"],"key":"Folk Catholicism"},{"value":["父权制天主教"],"key":"patriarchal Catholicism"}],"query":"Catholicism","translation":["天主教"],"errorCode":"0","dict":{"url":"yddict://m.youdao.com/dict?le=eng&q=Catholicism"},"webdict":{"url":"http://m.youdao.com/dict?le=eng&q=Catholicism"},"basic":{"us-phonetic":"kə'θɔlisizəm","phonetic":"kə'θɔlisizəm","uk-phonetic":"kə'θɔlisizəm","explains":["n. 天主教;天主教义"]},"l":"EN2zh-CHS"}
json
The eighth edition of the Karnataka Premier League is up and running and Hubli Tigers will be taking on Belagavi Panthers at the M Chinnaswamy Stadium on 21st August, Wednesday. Both teams are outside the top four after playing two matches each and they would be dearly hoping that rain doesn't play spoilsport yet again. After losing their first match by a mere five-run margin, Belagavi Panthers' second fixture was abandoned due to heavy rain. They haven't played much and as a result, there isn't much to talk about their performances. If anything, medium-pacer Zahoor Farooqui has been their best bowler and has taken three wickets so far. For Hubli, the situation is even worse as they have lost both of their matches so far and lie at the very bottom of the points table. Even though their most recent match against Ballari was close, a complete performance has been missing. Except for KB Pawan's half-century in the first game, no other Hubli batsman has registered a 50+ score and that reflects a failure to convert starts. The bowling has been largely disciplined but has lacked penetration. Belagavi Panthers: Mir Kaunian Abbas (C), Shubhang Hegde, Avinash D, Manish Pandey, D Negi, Abhinav Manohar, Arshdeep Singh Brar, Ritesh Bhatkal, Stalin Hoover, Darshan MB, Ravikumar Samarth, Zahoor Farooqui, Kiran AM, Rakshith S, Lochan Appanna, Darshan Machaiah, Abdul Majid. Hubli Tigers: R Vinay Kumar (C), Praveen Dubey, Mohammed Taha, Pawan KB, Aditya Somanna, Shishir Bhavane, Vidyadhar Patil, Mahesh Patel, Abhilash Shetty, David Mathias, Shivil Kaushik, Suraj Seshadri, Mithrakanth Yadav, Vishwanath M, KL Srijith, Luvnith Sisodia, Parikshith Shetty, Dheeraj Shashidhar. Belagavi are expected to go in with the same line-up whereas Hubli might make one change by replacing young Vidhyadhar Patil for Shivil Kaushik or Mahesh Patel. Belagavi Panthers: Ravikumar Samarth, Stallin Hoover, Manish Pandey, Mir Kaunain Abbas (C & WK), Abhinav Manohar, Dikshanshu Negi, Arshdeep Singh Brar, Ritesh Bhatkal, Shubhang Hegde, Avinash D, Zahoor Farooqui. Hubli Tigers: Mohammad Taha, M Vishwanathan, Luvnith Sisodia (WK), KL Shrijith, KB Pawan, Vinay Kumar (C), Praveen Dubey, David Mathias, Aditya Somanna, Mithrakanth Yadav, Vidyadhar Patil/ Shivil Kaushik/ Mahesh Patel. Unfortunately, rain is still expected to play a part on Wednesday and this means that any team that might win the toss would look to field first and take advantage of the VJD method. Fast-bowlers can also make use of the moisture on the pitch. Wicket-keeper: We can pick two wicket-keepers for this game. KB Pawan scored a half-century in the first game and he can be a solid pick. Partnering him will be M Vishwanathan, Pawan's team-mate and opening batsman. Vishwanathan scored 30 runs against Ballari. Batsmen: No surprises here. Without an iota of doubt, Manish Pandey will be the first pick. He didn't get to bat in the last game and will be raring to make his presence felt immediately. Belagavi skipper Kaunain Abbas is also a good choice, more so because he is also keeping for the team. The lone specialist batsman from Hubli will be KL Shrijith. Shrijith has scored 33 and 22 in the last two matches. He will want to convert his scores into big ones and the match against Belagavi will be the perfect opportunity to do so. All-rounders: Karnataka's favourite son Vinay Kumar will be a key all-rounder for both Belagavi and also Dream11 players. Apart from scoring 48 runs with the bat, Vinay has also been effective with the ball and has taken a wicket while maintaining a good economy rate in the two matches so far. Aditya Somanna is another good all-rounder who can pick up crucial points. After finishing as the 4th highest wicket-taker last year, the 23-year-old has started this season very well and went on to take a couple of wickets against Ballari. The final all-rounder will be Dikshanshu Negi. Negi is a wonderful leg-spinner and is also more than serviceable with the bat. He has taken a wicket and scored 17 runs so far. Bowlers: The experienced Zahoor Farooqui has been really good in the two matches so far by picking up three wickets. He will the lone Belagavi bowler with David Mathias and Mithrakanth Yadav completing the Dream11 team. Mithrakanth Yadav has already taken four wickets and given his consistency, he can be trusted to deliver again. David Mathias came into the side for the last match and picked up a couple of wickets. Captain: Manish Pandey will be the captain of the Dream11 team. Vinay Kumar will be the vice-captain due to dual skills. Alternate options for captaincy would be Kaunain Abbas and KB Pawan. Fantasy Suggestion #1: KB Pawan, M Vishwanathan, Manish Pandey (C), Kaunain Abbas, KL Shrijith, Vinay Kumar (VC), Aditya Somanna, Dikshanshu Negi, Zahoor Farooqui, David Mathias, Mithrakanth Yadav. Fantasy Suggestion #2: KB Pawan (VC), Luvnith Sisodia, Manish Pandey, Kaunain Abbas (C), R Samarth, KL Shrijith, Vinay Kumar, Praveen Dubey, Zahoor Farooqui, D Avinash, Mithrakanth Yadav. Follow Sportskeeda for all the updates on KPL teams & squads, points table, news, results, KPL schedule, most runs, most wickets and fantasy tips.
english
const fs = require('fs'); const path = require('path'); const pkgJson = path.join(process.cwd(), 'package.json'); module.exports = fs.existsSync(pkgJson) && require(pkgJson).name === '@discoveryjs/discovery' ? process.cwd() : path.dirname(require.resolve('@discoveryjs/discovery/package.json', { paths: [process.cwd()] }));
javascript
<gh_stars>0 // This file implements a ppm to (dc-ppm, ac-ppm) mapping that allows // us to experiment in different ways to compose the image into // (4x4) pseudo-dc and respective ac components. #include "upscaler.h" #include <stdio.h> #include <stdlib.h> #include <string.h> #include <cmath> #include <vector> #include "butteraugli_distance.h" #include "gamma_correct.h" #include "image.h" #include "image_io.h" #include "resample.h" #define BUTTERAUGLI_RESTRICT __restrict__ namespace pik { namespace { std::vector<float> ComputeKernel(float sigma) { // Filtering becomes slower, but more Gaussian when m is increased. // More Gaussian doesn't mean necessarily better results altogether. const float m = 2.5; const float scaler = -1.0 / (2 * sigma * sigma); const int diff = std::max<int>(1, m * fabs(sigma)); std::vector<float> kernel(2 * diff + 1); for (int i = -diff; i <= diff; ++i) { kernel[i + diff] = exp(scaler * i * i); } return kernel; } void ConvolveBorderColumn(const ImageF& in, const std::vector<float>& kernel, const float weight_no_border, const float border_ratio, const size_t x, float* const BUTTERAUGLI_RESTRICT row_out) { const int offset = kernel.size() / 2; int minx = x < offset ? 0 : x - offset; int maxx = std::min<int>(in.xsize() - 1, x + offset); float weight = 0.0f; for (int j = minx; j <= maxx; ++j) { weight += kernel[j - x + offset]; } // Interpolate linearly between the no-border scaling and border scaling. weight = (1.0f - border_ratio) * weight + border_ratio * weight_no_border; float scale = 1.0f / weight; for (size_t y = 0; y < in.ysize(); ++y) { const float* const BUTTERAUGLI_RESTRICT row_in = in.Row(y); float sum = 0.0f; for (int j = minx; j <= maxx; ++j) { sum += row_in[j] * kernel[j - x + offset]; } row_out[y] = sum * scale; } } // Computes a horizontal convolution and transposes the result. ImageF Convolution(const ImageF& in, const std::vector<float>& kernel, const float border_ratio) { ImageF out(in.ysize(), in.xsize()); const int len = kernel.size(); const int offset = kernel.size() / 2; float weight_no_border = 0.0f; for (int j = 0; j < len; ++j) { weight_no_border += kernel[j]; } float scale_no_border = 1.0f / weight_no_border; const int border1 = in.xsize() <= offset ? in.xsize() : offset; const int border2 = in.xsize() - offset; int x = 0; // left border for (; x < border1; ++x) { ConvolveBorderColumn(in, kernel, weight_no_border, border_ratio, x, out.Row(x)); } // middle for (; x < border2; ++x) { float* const BUTTERAUGLI_RESTRICT row_out = out.Row(x); for (size_t y = 0; y < in.ysize(); ++y) { const float* const BUTTERAUGLI_RESTRICT row_in = &in.Row(y)[x - offset]; float sum = 0.0f; for (int j = 0; j < len; ++j) { sum += row_in[j] * kernel[j]; } row_out[y] = sum * scale_no_border; } } // right border for (; x < in.xsize(); ++x) { ConvolveBorderColumn(in, kernel, weight_no_border, border_ratio, x, out.Row(x)); } return out; } // A blur somewhat similar to a 2D Gaussian blur. // See: https://en.wikipedia.org/wiki/Gaussian_blur ImageF Blur(const ImageF& in, float sigma, float border_ratio) { std::vector<float> kernel = ComputeKernel(sigma); return Convolution(Convolution(in, kernel, border_ratio), kernel, border_ratio); } Image3F Blur(const Image3F& image, float sigma) { float border = 0.0; return Image3F(Blur(image.plane(0), sigma, border), Blur(image.plane(1), sigma, border), Blur(image.plane(2), sigma, border)); } // DoGBlur is an approximate of difference of Gaussians. We use it to // approximate LoG (Laplacian of Gaussians). // See: https://en.wikipedia.org/wiki/Difference_of_Gaussians // For motivation see: // https://en.wikipedia.org/wiki/Pyramid_(image_processing)#Laplacian_pyramid ImageF DoGBlur(const ImageF& in, float sigma, float border_ratio) { ImageF blur1 = Blur(in, sigma, border_ratio); ImageF blur2 = Blur(in, sigma * 2.0f, border_ratio); static const float mix = 0.25; ImageF out(in.xsize(), in.ysize()); for (size_t y = 0; y < in.ysize(); ++y) { const float* const BUTTERAUGLI_RESTRICT row1 = blur1.Row(y); const float* const BUTTERAUGLI_RESTRICT row2 = blur2.Row(y); float* const BUTTERAUGLI_RESTRICT row_out = out.Row(y); for (size_t x = 0; x < in.xsize(); ++x) { row_out[x] = (1.0f + mix) * row1[x] - mix * row2[x]; } } return out; } Image3F DoGBlur(const Image3F& image, float sigma) { float border = 0.0; return Image3F(DoGBlur(image.plane(0), sigma, border), DoGBlur(image.plane(1), sigma, border), DoGBlur(image.plane(2), sigma, border)); } void SelectiveBlur(Image3F& image, float sigma, float select) { Image3F copy = Blur(image, sigma); float select2 = select * 2; float ramp = 0.8f; float onePerSelect = ramp / select; float onePerSelect2 = ramp / select2; for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < image.ysize(); ++y) { const float* PIK_RESTRICT row_copy = copy.ConstPlaneRow(c, y); float* PIK_RESTRICT row = image.PlaneRow(c, y); for (size_t x = 0; x < image.xsize(); ++x) { float dist = fabs(row_copy[x] - row[x]); float w = 0.0f; if ((x & 7) == 0 || (x & 7) == 7 || (y & 7) == 0 || (y & 7) == 7) { if (dist < select2) { w = ramp - dist * onePerSelect2; if (w > 1.0f) w = 1.0f; } } else if (dist < select) { w = ramp - dist * onePerSelect; if (w > 1.0f) w = 1.0f; } row[x] = w * row_copy[x] + (1.0 - w) * row[x]; } } } } void SelectiveBlur8x8(Image3F& image, Image3F& ac, float sigma, float select_mod) { Image3F copy = Blur(image, sigma); float ramp = 1.0f; for (int c = 0; c < 3; ++c) { for (size_t wy = 0; wy < image.ysize(); wy += 8) { for (size_t wx = 0; wx < image.xsize(); wx += 8) { // Find maxdiff double max = 0; for (int dy = 0; dy < 6; ++dy) { for (int dx = 0; dx < 6; ++dx) { int y = wy + dy; int x = wx + dx; if (y >= image.ysize() || x >= image.xsize()) { break; } // Look at the criss-cross of diffs between two pixels. // Scale the smoothing within the block of the amplitude // of such local change. const float* PIK_RESTRICT row_ac0 = ac.PlaneRow(c, y); const float* PIK_RESTRICT row_ac2 = ac.PlaneRow(c, y + 2); float dist = fabs(row_ac0[x] - row_ac0[x + 2]); if (max < dist) max = dist; dist = fabs(row_ac0[x] - row_ac2[x]); if (max < dist) max = dist; dist = fabs(row_ac0[x] - row_ac2[x + 2]); if (max < dist) max = dist; dist = fabs(row_ac0[x + 2] - row_ac2[x]); if (max < dist) max = dist; } } float select = select_mod * max; float select2 = 2.0 * select; float onePerSelect = ramp / select; float onePerSelect2 = ramp / select2; for (int dy = 0; dy < 8; ++dy) { for (int dx = 0; dx < 8; ++dx) { int y = wy + dy; int x = wx + dx; if (y >= image.ysize() || x >= image.xsize()) { break; } const float* PIK_RESTRICT row_copy = copy.PlaneRow(c, y); float* PIK_RESTRICT row = image.PlaneRow(c, y); float dist = fabs(row_copy[x] - row[x]); float w = 0.0f; if ((x & 7) == 0 || (x & 7) == 7 || (y & 7) == 0 || (y & 7) == 7) { if (dist < select2) { w = ramp - dist * onePerSelect2; if (w > 1.0f) w = 1.0f; } } else if (dist < select) { w = ramp - dist * onePerSelect; if (w > 1.0f) w = 1.0f; } row[x] = w * row_copy[x] + (1.0 - w) * row[x]; } } } } } } Image3F SubSampleSimple8x8(const Image3F& image) { const size_t nxs = (image.xsize() + 7) >> 3; const size_t nys = (image.ysize() + 7) >> 3; Image3F retval(nxs, nys, 0.0f); float mul = 1 / 64.0; for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < image.ysize(); ++y) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, y); float* PIK_RESTRICT row_out = retval.PlaneRow(c, y >> 3); for (size_t x = 0; x < image.xsize(); ++x) { row_out[x >> 3] += mul * row_in[x]; } } } if ((image.xsize() & 7) != 0) { const float last_column_mul = 8.0 / (image.xsize() & 7); for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < nys; ++y) { retval.PlaneRow(c, y)[nxs - 1] *= last_column_mul; } } } if ((image.ysize() & 7) != 0) { const float last_row_mul = 8.0 / (image.ysize() & 7); for (int c = 0; c < 3; ++c) { for (size_t x = 0; x < nxs; ++x) { retval.PlaneRow(c, nys - 1)[x] *= last_row_mul; } } } return retval; } Image3F SubSampleSimple4x4(const Image3F& image) { const size_t nxs = (image.xsize() + 3) >> 2; const size_t nys = (image.ysize() + 3) >> 2; Image3F retval(nxs, nys, 0.0f); float mul = 1 / 16.0; for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < image.ysize(); ++y) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, y); float* PIK_RESTRICT row_out = retval.PlaneRow(c, y >> 2); for (size_t x = 0; x < image.xsize(); ++x) { row_out[x >> 2] += mul * row_in[x]; } } } if ((image.xsize() & 3) != 0) { const float last_column_mul = 4.0 / (image.xsize() & 3); for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < nys; ++y) { retval.PlaneRow(c, y)[nxs - 1] *= last_column_mul; } } } if ((image.ysize() & 3) != 0) { const float last_row_mul = 4.0 / (image.ysize() & 3); for (int c = 0; c < 3; ++c) { for (size_t x = 0; x < nxs; ++x) { retval.PlaneRow(c, nys - 1)[x] *= last_row_mul; } } } return retval; } Image3F SuperSample2x2(const Image3F& image) { size_t nxs = image.xsize() << 1; size_t nys = image.ysize() << 1; Image3F retval(nxs, nys); for (int c = 0; c < 3; ++c) { for (size_t ny = 0; ny < nys; ++ny) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, ny >> 1); float* PIK_RESTRICT row_out = retval.PlaneRow(c, ny); for (size_t nx = 0; nx < nxs; ++nx) { row_out[nx] = row_in[nx >> 1]; } } } return retval; } Image3F SuperSample4x4(const Image3F& image) { size_t nxs = image.xsize() << 2; size_t nys = image.ysize() << 2; Image3F retval(nxs, nys); for (int c = 0; c < 3; ++c) { for (size_t ny = 0; ny < nys; ++ny) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, ny >> 2); float* PIK_RESTRICT row_out = retval.PlaneRow(c, ny); for (size_t nx = 0; nx < nxs; ++nx) { row_out[nx] = row_in[nx >> 2]; } } } return retval; } Image3F SuperSample8x8(const Image3F& image) { size_t nxs = image.xsize() << 3; size_t nys = image.ysize() << 3; Image3F retval(nxs, nys); for (int c = 0; c < 3; ++c) { for (size_t ny = 0; ny < nys; ++ny) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, ny >> 3); float* PIK_RESTRICT row_out = retval.PlaneRow(c, ny); for (size_t nx = 0; nx < nxs; ++nx) { row_out[nx] = row_in[nx >> 3]; } } } return retval; } void Smooth4x4Corners(Image3F& ima) { static const float overshoot = 3.5; static const float m = 1.0 / (4.0 - overshoot); for (int y = 3; y + 3 < ima.ysize(); y += 4) { for (int x = 3; x + 3 < ima.xsize(); x += 4) { float ave[3] = {0}; for (int c = 0; c < 3; ++c) { ave[c] += ima.PlaneRow(c, y)[x]; ave[c] += ima.PlaneRow(c, y)[x + 1]; ave[c] += ima.PlaneRow(c, y + 1)[x]; ave[c] += ima.PlaneRow(c, y + 1)[x + 1]; } const int off = 2; for (int c = 0; c < 3; ++c) { float others = (ave[c] - overshoot * ima.PlaneRow(c, y)[x]) * m; ima.PlaneRow(c, y - off)[x - off] -= (others - ima.PlaneRow(c, y)[x]); ima.PlaneRow(c, y)[x] = others; } for (int c = 0; c < 3; ++c) { float others = (ave[c] - overshoot * ima.PlaneRow(c, y)[x + 1]) * m; ima.PlaneRow(c, y - off)[x + off + 1] -= (others - ima.PlaneRow(c, y)[x + 1]); ima.PlaneRow(c, y)[x + 1] = others; } for (int c = 0; c < 3; ++c) { float others = (ave[c] - overshoot * ima.PlaneRow(c, y + 1)[x]) * m; ima.PlaneRow(c, y + off + 1)[x - off] -= (others - ima.PlaneRow(c, y + 1)[x]); ima.PlaneRow(c, y + 1)[x] = others; } for (int c = 0; c < 3; ++c) { float others = (ave[c] - overshoot * ima.PlaneRow(c, y + 1)[x + 1]) * m; ima.PlaneRow(c, y + off + 1)[x + off + 1] -= (others - ima.PlaneRow(c, y + 1)[x + 1]); ima.PlaneRow(c, y + 1)[x + 1] = others; } } } } void Subtract(Image3F& a, const Image3F& b) { for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < a.ysize(); ++y) { const float* PIK_RESTRICT row_b = b.PlaneRow(c, y); float* PIK_RESTRICT row_a = a.PlaneRow(c, y); for (size_t x = 0; x < a.xsize(); ++x) { row_a[x] -= row_b[x]; } } } } void Add(Image3F& a, const Image3F& b) { for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < a.ysize(); ++y) { const float* PIK_RESTRICT row_b = b.PlaneRow(c, y); float* PIK_RESTRICT row_a = a.PlaneRow(c, y); for (size_t x = 0; x < a.xsize(); ++x) { row_a[x] += row_b[x]; } } } } // Clamps pixel values to 0, 255. Image3F Crop(const Image3F& image, int newxsize, int newysize) { Image3F retval(newxsize, newysize); for (int c = 0; c < 3; ++c) { for (int y = 0; y < newysize; ++y) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, y); float* PIK_RESTRICT row_out = retval.PlaneRow(c, y); for (int x = 0; x < newxsize; ++x) { float v = row_in[x]; if (v < 0) { v = 0; } if (v > 255) { v = 255; } row_out[x] = v; } } } return retval; } Image3F ToLinear(const Image3F& image) { Image3F out(image.xsize(), image.ysize()); for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < image.ysize(); ++y) { const float* PIK_RESTRICT row_in = image.PlaneRow(c, y); float* PIK_RESTRICT row_out = out.PlaneRow(c, y); for (size_t x = 0; x < image.xsize(); ++x) { row_out[x] = Srgb8ToLinearDirect(row_in[x]); } } } return out; } Image3F EncodePseudoDC(const Image3F& in) { Image3F goal = CopyImage(in); Image3F image8x8sub; static const int kIters = 2; for (int ii = 0; ii < kIters; ++ii) { if (ii != 0) { Image3F normal = UpscalerReconstruct(image8x8sub); // adjust the image by diff of normal and image. for (int c = 0; c < 3; ++c) { for (size_t y = 0; y < in.ysize(); ++y) { const float* PIK_RESTRICT row_normal = normal.PlaneRow(c, y); const float* PIK_RESTRICT row_in = in.PlaneRow(c, y); float* PIK_RESTRICT row_goal = goal.PlaneRow(c, y); for (size_t x = 0; x < in.xsize(); ++x) { row_goal[x] -= 0.1 * (row_normal[x] - row_in[x]); } } } } image8x8sub = SubSampleSimple8x8(goal); } // Encode pseudo dc. return image8x8sub; } } // namespace Image3F UpscalerReconstruct(const Image3F& in) { Image3F out1 = SuperSample4x4(in); Smooth4x4Corners(out1); out1 = Blur(out1, 2.5); Image3F out(out1.xsize() * 2, out1.ysize() * 2); Upsample<slow::Upsampler>(ExecutorLoop(), out1, kernel::CatmullRom(), &out); return out; } } // namespace pik
cpp
<reponame>Yadon079/yadon079.github.io --- layout: post date: 2020-11-04 18:16:00 title: "리스트" description: "자료구조" subject: 자료구조 category: [ data structure ] tags: [ data structure, abstract data type, array list ] use_math: true comments: true --- # 리스트(List) ## 추상 자료형 : Abstract Data Type ADT라고도 불리는 추상 자료형은 '구체적인 기능의 완성과정을 언급하지 않고, 순수하게 기능이 무엇인지를 나열한 것'이다. 자료구조는 추상 자료형이 정의한 연산들을 구현한 구현체를 가리킨다. 쉽게 구분하자면 클래스인지 인터페이스인지를 확인하면 되는 것이다. 스택이나 큐와 같이 구현 방법이 정의되어 있지 않으면 ADT이고, 배열이나 연결 리스트처럼 저장 방식이 정해져 있다면 자료구조이다. ## 배열을 이용한 리스트의 구현 ### 리스트의 이해 리스트라는 자료구조는 구현방법에 따라서 크게 두 가지로 나눌 수 있다. + 순차 리스트 배열을 기반으로 구현된 리스트 + 연결 리스트 메모리의 동적 할당을 기반으로 구현된 리스트 리스트 자료구조는 데이터를 나란히 저장한다. 그리고 중복된 데이터의 저장을 허용한다.
markdown
from PyQt4.QtCore import QMimeData, QByteArray, Qt, QSize, QPoint, QVariant,SIGNAL from PyQt4.QtGui import QTreeWidget, QImage, QDrag, QPixmap, QIcon, QPalette, QColor,QTreeWidgetItem import logging class SimpleDraggableTreeWidget(QTreeWidget): """ TreeWidget suitable for holding a list of strings. """ MIME_TYPE = "text/plain" def __init__(self, headerLabel, dragEnabled=False, mimeType=None, parent=None): """ Constructor. """ QTreeWidget.__init__(self,parent) self.setMimeType(mimeType) self.setAutoFillBackground(True) #print "color roles", QPalette.Base, QPalette.Window, self.backgroundRole() lightBlueBackgroundColor = QColor(Qt.blue).lighter(195) #lightBlueBackgroundColor = QColor(Qt.red) self.palette().setColor(QPalette.Base, lightBlueBackgroundColor) # OS X self.palette().setColor(QPalette.Window, lightBlueBackgroundColor) self.setColumnCount(1) self.setHeaderLabels([headerLabel]) if dragEnabled: self.setDragEnabled(True) def populate(self, items): """ Fills items into list. """ self.insertTopLevelItems(0, items) def setDragEnable(self, dragEnabled, mimeType=None): """ Usual behavior of QWidget's setDragEnabled() function plus optional setting of mimeType. """ QTreeWidget.setDragEnabled(dragEnabled) self.setMimeType(mimeType) def mimeType(self): """ Returns mime type which will be used to encode list entries while dragging. """ return self._mimeType def setMimeType(self, mimeType): """ Sets mime type of this widget to type if type is not None. If type is None the default mime type MIME_TYPE will be used. """ if mimeType: self._mimeType = mimeType else: self._mimeType = self.MIME_TYPE def mimeTypes(self): """ Returns self.mimeType() as single element of QStringList. """ list = QStringList() list << self.mimeType() return list def mimeData(self, items): """ Returns QMimeData for drag and drop. """ logging.debug(self.__class__.__name__ + ": mimeData()") mime = QMimeData() encodedData = QByteArray() for item in items: encodedData.append(item.text(0)) mime.setData(self.mimeType(), encodedData) return mime def startDrag(self, supportedActions): """ Overwritten function of QTreeWidget. This function creates a QDrag object representing the selected element of this TreeWidget. """ logging.debug(self.__class__.__name__ +": startDrag()") indexes = self.selectedIndexes() if len(indexes) > 0: data = self.model().mimeData(indexes) if not data: return drag = QDrag(self) drag.setMimeData(data) if self.model().data(indexes[0], Qt.DecorationRole).type() == QVariant.Icon: icon = QIcon(self.model().data(indexes[0], Qt.DecorationRole)) drag.setPixmap(icon.pixmap(QSize(50, 50))) drag.setHotSpot(QPoint(drag.pixmap().width()/2, drag.pixmap().height()/2)) # center icon in respect to cursor defaultDropAction = Qt.IgnoreAction drag.exec_(supportedActions, defaultDropAction) def mousePressEvent(self,event): QTreeWidget.mousePressEvent(self,event) if event.button()==Qt.RightButton: self.emit(SIGNAL("mouseRightPressed"), event.globalPos())
python
<reponame>Lovemma/pyshark import binascii from pyshark.packet.common import Pickleable, SlotsPickleable class LayerField(SlotsPickleable): """ Holds all data about a field of a layer, both its actual value and its name and nice representation. """ # Note: We use this object with slots and not just a dict because # it's much more memory-efficient (cuts about a third of the memory). __slots__ = ['name', 'showname', 'raw_value', 'show', 'hide', 'pos', 'size', 'unmaskedvalue'] def __init__(self, name=None, showname=None, value=None, show=None, hide=None, pos=None, size=None, unmaskedvalue=None): self.name = name self.showname = showname self.raw_value = value self.show = show self.pos = pos self.size = size self.unmaskedvalue = unmaskedvalue if hide and hide == 'yes': self.hide = True else: self.hide = False def __repr__(self): return '<LayerField %s: %s>' % (self.name, self.get_default_value()) def get_default_value(self): """ Gets the best 'value' string this field has. """ val = self.show if not val: val = self.raw_value if not val: val = self.showname return val @property def showname_value(self): """ For fields which do not contain a normal value, we attempt to take their value from the showname. """ if self.showname and ': ' in self.showname: return self.showname.split(': ', 1)[1] @property def showname_key(self): if self.showname and ': ' in self.showname: return self.showname.split(': ', 1)[0] @property def binary_value(self): """ Converts this field to binary (assuming it's a binary string) """ return binascii.unhexlify(self.raw_value) @property def int_value(self): """ Returns the int value of this field (assuming it's an integer integer). """ return int(self.raw_value) @property def hex_value(self): """ Returns the int value of this field if it's in base 16 (either as a normal number or in a "0xFFFF"-style hex value) """ return int(self.raw_value, 16) base16_value = hex_value class LayerFieldsContainer(str, Pickleable): """ An object which contains one or more fields (of the same name). When accessing member, such as showname, raw_value, etc. the appropriate member of the main (first) field saved in this container will be shown. """ def __new__(cls, main_field, *args, **kwargs): value = main_field.get_default_value() if value is None: value = '' obj = str.__new__(cls, value, *args, **kwargs) obj.fields = [main_field] return obj def __dir__(self): return dir(type(self)) + list(self.__dict__.keys()) + dir(self.main_field) def add_field(self, field): self.fields.append(field) @property def main_field(self): return self.fields[0] @property def alternate_fields(self): """ Return the alternate values of this field containers (non-main ones). """ return self.fields[1:] @property def all_fields(self): """ Returns all fields in a list, the main field followed by the alternate fields. """ return self.fields def __getattr__(self, item): return getattr(self.main_field, item)
python
package com.pandong.tool.nnts.server.handler; import com.pandong.tool.nnts.Global; import com.pandong.tool.nnts.model.Service; import com.pandong.tool.nnts.model.utils.TransferGenerate; import com.pandong.tool.nnts.server.utils.ServerUtils; import io.netty.buffer.ByteBuf; import io.netty.channel.Channel; import io.netty.channel.ChannelHandlerContext; import io.netty.channel.ChannelOption; import io.netty.channel.SimpleChannelInboundHandler; import lombok.extern.slf4j.Slf4j; import java.net.InetSocketAddress; @Slf4j public class ProxyChannelHandler extends SimpleChannelInboundHandler<ByteBuf> { @Override protected void channelRead0(ChannelHandlerContext ctx, ByteBuf msg) throws Exception { Channel requestChannel = ctx.channel(); Channel proxyChannel = requestChannel.attr(Global.ChannelAttribute.PROXY_CHANNEL).get(); long requestId = requestChannel.attr(Global.ChannelAttribute.REQUEST_ID).get(); if(proxyChannel==null){ log.debug("request id --> " + requestId + ", no proxy channel."); requestChannel.close(); }else { byte[] bytes = new byte[msg.readableBytes()]; msg.readBytes(bytes); log.debug("request id --> " + requestId + ", read byte size --> "+ bytes.length); proxyChannel.writeAndFlush(TransferGenerate.generateTransferRequest(requestId, ServerUtils.server(), bytes)); } } @Override public void channelActive(ChannelHandlerContext ctx) throws Exception { InetSocketAddress inetSocketAddress = (InetSocketAddress) ctx.channel().localAddress(); int port = inetSocketAddress.getPort(); Channel proxyChannel = ServerUtils.getProxyChannelWithPort(port); Service service = ServerUtils.getServiceWithPort(port); if (proxyChannel == null) { //没有client端的连接 ctx.channel().close(); } else { //向client端发起连接请求 long requestId = Global.getId(Global.SequenceName.SEQUENCE_NAME_REQUEST); log.debug("create new request, requestId --> "+requestId); //根据requestID,缓存channel,方便在传输数据时使用 ServerUtils.addConnectProxy(requestId, ctx.channel(), proxyChannel, service); //在client端连接道真实服务前,暂停读取数据 ctx.channel().config().setOption(ChannelOption.AUTO_READ, false); log.debug("send connect commend..."); proxyChannel.writeAndFlush(TransferGenerate .generateConnectRequest(requestId, ServerUtils.server(), ServerUtils.getServiceWithPort(port))); } super.channelActive(ctx); } @Override public void channelInactive(ChannelHandlerContext ctx) throws Exception { Channel requestChannel = ctx.channel(); Channel proxyChannel = requestChannel.attr(Global.ChannelAttribute.PROXY_CHANNEL).get(); long requestId = requestChannel.attr(Global.ChannelAttribute.REQUEST_ID).get(); Service service = requestChannel.attr(Global.ChannelAttribute.SERVICE).get(); if(proxyChannel==null){ requestChannel.close(); }else{ log.debug("close request, reuqest id --> "+requestId); proxyChannel.attr(Global.ChannelAttribute.REQUEST_CHANNEL).set(null); proxyChannel.writeAndFlush(TransferGenerate.generateDisconnectRequest(requestId, ServerUtils.server(), service)); } super.channelInactive(ctx); } @Override public void channelWritabilityChanged(ChannelHandlerContext ctx) throws Exception { Channel requestChannel = ctx.channel(); Channel proxyChannel = requestChannel.attr(Global.ChannelAttribute.PROXY_CHANNEL).get(); if(proxyChannel==null){ requestChannel.close(); }else { proxyChannel.config().setOption(ChannelOption.AUTO_READ, requestChannel.isWritable()); } super.channelWritabilityChanged(ctx); } @Override public void exceptionCaught(ChannelHandlerContext ctx, Throwable cause) throws Exception { super.exceptionCaught(ctx, cause); ctx.close(); } }
java
import os import sys from loguru import logger from rich.console import Console console_args = {} if "pytest" in sys.modules: console_args["width"] = 120 console = Console(**console_args) cpu_count = None def escape_logging(s): return str(s).replace("<", "\\<").replace("{", "{{").replace("}", "}}") def CPUs(): """ Detects the number of CPUs on a system. Cribbed from pp. """ global cpu_count if cpu_count is None: cpu_count = 1 # default # Linux, Unix and MacOS: if hasattr(os, "sysconf"): if "SC_NPROCESSORS_ONLN" in os.sysconf_names: # Linux & Unix: ncpus = os.sysconf("SC_NPROCESSORS_ONLN") if isinstance(ncpus, int) and ncpus > 0: cpu_count = ncpus else: # OSX: pragma: no cover cpu_count = int( os.popen2("sysctl -n hw.ncpu")[1].read() ) # pragma: no cover # Windows: if "NUMBER_OF_PROCESSORS" in os.environ: # pragma: no cover ncpus = int(os.environ["NUMBER_OF_PROCESSORS"]) if ncpus > 0: cpu_count = ncpus return cpu_count def job_or_filename(job_or_filename, invariant_class=None): """Take a filename, or a job. Return Path(filename), dependency-for-that-file ie. either the job, or a invariant_class (default: FileInvariant)""" from .jobs import Job, FileInvariant from pathlib import Path if invariant_class is None: # pragma: no cover invariant_class = FileInvariant if isinstance(job_or_filename, Job): filename = job_or_filename.files[0] deps = [job_or_filename] elif job_or_filename is not None: filename = Path(job_or_filename) deps = [invariant_class(filename)] else: filename = None deps = [] return filename, deps def assert_uniqueness_of_object( object_with_name_attribute, pipegraph=None, also_check=None ): """Makes certain there is only one object with this class & .name. This is necessary so the pipegraph jobs assign their data only to the objects you're actually working with.""" if pipegraph is None: # pragma: no branch from pypipegraph2 import global_pipegraph pipegraph = global_pipegraph if object_with_name_attribute.name.find("/") != -1: raise ValueError( "Names must not contain /, it confuses the directory calculations" ) if pipegraph is None: # pragma: no cover return if not hasattr(pipegraph, "object_uniquifier"): pipegraph.object_uniquifier = {} typ = object_with_name_attribute.__class__ if typ not in pipegraph.object_uniquifier: pipegraph.object_uniquifier[typ] = {} if object_with_name_attribute.name in pipegraph.object_uniquifier[typ]: raise ValueError( "Doublicate object: %s, %s" % (typ, object_with_name_attribute.name) ) if also_check: if not isinstance(also_check, list): also_check = [also_check] for other_typ in also_check: if ( other_typ in pipegraph.object_uniquifier and object_with_name_attribute.name in pipegraph.object_uniquifier[other_typ] ): raise ValueError( "Doublicate object: %s, %s" % (other_typ, object_with_name_attribute.name) ) object_with_name_attribute.unique_id = len(pipegraph.object_uniquifier[typ]) pipegraph.object_uniquifier[typ][object_with_name_attribute.name] = True def flatten_jobs(j): """Take an arbitrary deeply nested list of lists of jobs and return just the jobs""" from .jobs import Job if isinstance(j, Job): yield j else: for sj in j: yield from flatten_jobs(sj) do_jobtrace_log = False def log_warning(msg): logger.opt(depth=1).warning(msg) def log_error(msg): logger.opt(depth=1).error(msg) def log_info(msg): logger.opt(depth=1).info(msg) def log_debug(msg): logger.opt(depth=1).debug(msg) def log_job_trace(msg): if do_jobtrace_log: logger.opt(depth=1).log("JT", msg) def log_trace(msg): if do_jobtrace_log: # pragma: no cover logger.opt(depth=1).trace(msg) def shorten_job_id(job_id): dotdotcount = job_id.count(":::") if dotdotcount: return job_id[: job_id.find(":::") + 3] + "+" + str(dotdotcount) else: return job_id def pretty_log_errors(func): """capture exceptions (on a function outside of ppg) and format it with our fancy local logging exception logger """ def inner(*args, **kwargs): try: func(*args, **kwargs) except Exception as e: exception_type, exception_value, tb = sys.exc_info() captured_tb = ppg2.ppg_traceback.Trace(exception_type, exception_value, tb) logger.error( captured_tb._format_rich_traceback_fallback( include_locals=True, include_formating=True ) ) raise return inner
python
A group of horses were caught on camera leaving their indoor shelter area in Canada, moving back to their paddocks. A footage taken in the Alberta province in Western Canada shows the stallions jumping and pushing through deep snow amid the storms. Before they made their way out of the enclosed shades, they were shielding themselves from the wintry climate. Patricia Kielstra gave a glimpse of the horses leaving the protective property area amid the harsh weather. Kielstra spoke to the Storyful and explained that the farm witnessed a “major snowstorm”. It was the reason why horses and sheep were shifted to an “indoor arena to keep them safe during this blizzard. ” However, the following day, she unlocked the doors letting the horses out who rushed to their pens. What surprised Kielstra was that the horses furrowed. She did not expect them to hurtle through the drifts in the way they did. Signing off, she asserted it was amazing to watch though. Here is the video: The 44-second video recording has been viewed more than 25,000 times on the micro-blogging site since shared. The post has been reshared on the platform nearly 100 times. Many users of the social media site have tapped the heart button almost 670 times. According to Environment Canada, warnings of heavy storms and strong winds were indicated on Sunday. The Albertans were urged to stay off roads as poor conditions and collisions were reported throughout southern Alberta. Snowfall and high winds caused burying vehicles, prompted highway closures and led to power outages in many areas over the weekend. Reports citing Environment Canada specified that the snowfall measurements through the weekend were complicated due to blowing snow. As per veterinarian experts’ publishing in American Association of Equine Practitioners, horses are well equipped to manage everything that winters can offer. They have a knack of getting out of the wind by using their long winter hair coat to their benefit. It helps to trap air next to the skin and also provide insulation against cold weather. A mount in a healthy body condition can resist temperatures down to -40 degrees Fahrenheit with ease. The only concern is any disturbance whatsoever in the horse’s hair. If strong winds ruffle the hair, it can upset the insulating layer of air trapped underneath, which supplies warmth to them. Wet weather can flatten the hair coat, hence a simple shelter that can serve as a windbreak can prevent chilling the horse.
english
Combining a sporty style with a chic of high fashion - these are the main features of Versace's fashion house, known all over the world. This Italian brand manages to decorate unpretentious jeans with fancy rhinestones and intricate embroidery. But this happens not only in clothes. He managed to unite the two contradictory elements in his fragrances. Here we will talk about the perfume Bright Crystal Versace (Shining Crystal) - a bright and classic specimen of this authoritative brand. What is hidden under a transparent, faceted diamond lid of an elegant and luxuriously decorated bottle? Versace Crystal Bright surprises with contrasts: crystal, transparent as a mountain stream, coldness and hot, all-consuming passion. The concept of composition is reflected even on the package: a soft pink box with a luxurious silvery ornament "under the skin". This fragrance, born in 2006, is produced in the concentration of "toilet water" in bottles of 30, 50 and 90 ml. It was created for charming women, endowed with an extraordinary charm. Unlike his older brother, released two years earlier by the "Crystal Noir", "Shining" is more fresh, clear, clean. It clearly shows an icy shade of fresh January frost. Creating Versace Bright Crystal, Donatella Versace conceived a real love potion. There should have been a fragrance in which only a hint of passion could be guessed. Knowing that the game is more desirable, the faster it escapes from the hunter, the perfumer Alberto Morillas, the author of the composition, wove threads of coldness and detachment into the languid and sensual base of musk and amber. It turned out to be a real masterpiece: a woman wrapped in the veil of Shining Crystal seems to be a fragile, refined and elegant, promising "hunter" inexplicable joys of "chasing" and enjoying the coveted "trophy". We'll try to arrange the composition Versace Bright Crystal by notes. This is difficult to do, because it plays like a young wine, and shimmers with all colors, like a diamond. If you say that this is a fruit and flower melody, it's like saying nothing. As in the "noir" version, there are tart notes of pomegranate, sweet pianon chords, warmth of mahogany, mental languor of amber and musk. However, the highest note is the ice yuzu, which has a separate score in this ensemble. Frosty hues softened by the tenderness of the magnolia petals and the lotus extract. What kind of woman needs to "wear" the fragrance Versace Bright Crystal? It is tender and graceful, light, like a drop of dew trembling on the petal of a magnolia, sweet, like temptation itself. However, one should not expect an easy victory: behind the image of a fragile blonde lies a strong-willed and passionate nature, able to achieve the set goals. A gentle seductress at any moment can turn into a femme fatale for the addicted man. Thanks to this duality, Versace Bright Crystal perfume can be used in winter and summer, camping in a cafe or cinema and on a party. This combination of flavors of wood mahogany and flowers, frost and oriental smells, repeated many times in the train, make the owner of the perfume simply irresistible. "Shining Crystal" is the embodiment of feminine seductiveness, sophisticated luxury and elegance.
english
{ "title": "openage", "url": "https://openage.sft.mx/", "description": "a free (as in freedom) game engine to implement Age of Empires", "tags": [ "games" ], "dateAdded": 0 }
json
On Thursday morning of the 29th of May,in a glossy shop in Marais – Paris, Pop star Rihanna quietly unveiled the debut Collection from her new Fenty clothing line. The collection which was expected to be released with a loud bang, runaway shows and lots of editors and media was an intimidate affair. “It’s definitely different from doing a show, with a show there are so many moving parts. That’s not to say it hasn’t been hectic, but there is something about the intimacy of doing things this way that I love,” Rihanna was quoted as saying. With this unveiling, Rihanna made history twice over. This the first time the luxury group, LVMH has launched a fashion brand from scratch since Christian Lacroix in 1987. Rihanna also became the first woman of colour to preside over an LVMH brand. She got a team of 12 designers for this brand. She also got on board Jahleel Weaver, a former stylist to serve as the brand’s style director. She has decided to disregard the seasonal collections unlike traditional collections and brands and her collections with drop at surprising and frequent intervals. The debut collection mostly consists of a series of sharp-shouldered blazers which structured feminine curves which also happen to the back bone of the collection, poplin blouses with exaggerated shoulders and collars, wide-leg trousers and denim, series of Japanese denim jackets (a fabric Rihanna is partial towards) with cargo pockets and split flare pants, a signature corset blazer dress in exaggerated shoulders and fanny packs that doubled as waist-cinching sashes. Aptly titled “Release 5-19” for the month and year it will become available the collection is for bosses wanting to have some fun. With sizes going up to size 14, she is trying to break another stereotype with offering designer wear for plus size women. Since Rihanna body structure varies from time to time she has been testing the collection by trying it on herself first. “I’m a curvy girl, and if I can’t wear it myself, it’s not going to work,” Rihanna told Vogue. She has a huge fan following which helps her push this collection in the limelight. Her Instagram alone has a following on 71.2 million people. Rihanna made a guest appearance in the teaser video for Fenty’s collection. She posted it on Instagram on Wednesday, a day before the unveiling. 2 million people had viewed it in a single day. If you happen to be in Paris, you can pop over to Marais and shop the collection. For others the collection goes live from 29th May. It ranges from $200 for a T-shirt and goes to around $1,500 for an outerwear. The post Rihanna Unveils New Collection of Fenty’s Clothing Line appeared first on iFashion Network - Global Fashion business, fashion news trends.
english
Hi guys, I didn't activate 3G but it automatically showing 3.5G in place of E sign in my Nokia 5230, so i recharged my cell with Rs.99 for 2.5 GB data for GPRS. But it giving speed like 3G, today i downloaded lots of mp3 and videos in my cell with speed varying in between 110 kb/s to 180 kb/s (Downloading speed). And nothing is deducting from my account for 3G speed. This is my Delhi no. and also working in roaming in Patna. Voice call on Skype working like a phone, YouTube is now faster than GPRS. Guys, is it default in reliance or it will deduct any money later for 3G speed??? I'm just attaching a screenshots of my mobile. on my pc:: THANK YOU.mp3 in UC browser, it tooks only 2-3 min. I'm just attaching a screenshots of my mobile. on my pc:: THANK YOU.mp3 in UC browser, it tooks only 2-3 min. Last edited:
english
<reponame>zhong-tsong/express-server-dmall<filename>app.js // 使用express构建web服务器 const express = require('express'); const bodyParser = require('body-parser'); const cors = require('cors'); const cookieParser = require('cookie-parser'); const pool = require('./pool'); // 引入路由模块 const index = require('./routes/index.js'); const products = require('./routes/products.js'); const details = require('./routes/details.js'); const users = require('./routes/users.js'); const cart = require('./routes/cart.js'); const comment = require('./routes/comment.js'); const brand = require('./routes/brand.js'); const dictionaries = require('./routes/dictionaries.js'); const order = require('./routes/order.js'); const collection = require('./routes/collection.js'); const address = require('./routes/address.js'); const message = require('./routes/message.js'); const admin = require('./routes/admin.js'); let app = express(); // 接口白名单 const requestWhiteList = [ '/api/index/onepush', '/api/index/banner', '/api/index/hot', '/api/products/select', '/api/index/kw', '/api/users/log', '/api/users/vali/forgetPwd', '/api/users/update/upwd', '/api/users/reg', '/api/details/select', '/api/cart/select/num', '/api/comment/select/pid', '/api/dic/selectDic', '/api/products/select/filter', '/api/users/logout', '/api/message/select' ]; // 配置跨域访问 app.use(cors({ // 指定接收的地址 origin: [ 'http://localhost:3000', 'http://127.0.0.1:3000', 'http://localhost:8080','http://127.0.0.1:8080', 'http://172.16.66.163:3000', 'http://192.168.2.102:3000', 'http://192.168.2.101:3000', 'http://192.168.2.100:3000', 'http://localhost:9000','http://127.0.0.1:9000' ], // 指定接收的请求类型 methods: ['GET', 'POST'], // 指定header alloweHeaders: ['Content-Type', 'Authorization'], credentials: true })) // 使用body-parser中间件 app.use( bodyParser.json({ limit: '50mb' }) ); app.use( bodyParser.urlencoded({ extended: false, limit: '50mb' }) ); // cookie app.use( cookieParser() ); // 托管静态资源到public目录下 app.use('/api', express.static('public')); app.all('/*', (req, res, next) => { const { token, type, uname } = req.headers || {}; const { path, body, query } = req; if ((type == 'wx' || type == 'vue') && !requestWhiteList.includes(path)) { if ((query.hasOwnProperty('uname') && !query.uname) || (body.hasOwnProperty('uname') && !body.uname) || !uname) { res.status(401).send({ code: 401, msg: '认证失败,重新登录!' }) return; } if (!token) { res.status(401).send({ code: 401, msg: '认证token不存在,重新登录!' }) return; } let sql = "SELECT * FROM dm_user WHERE upwd=? AND uname=?"; pool.query(sql, [token, uname], (err, data) => { if (err) { res.status(503).send({ code: 2, msg: err }) } else { if (!data.length) { res.status(401).send({ code: 401, msg: '认证token已失效,重新登录!' }) return; } next(); } }); } else { next(); } }) // 使用路由器来管理路由 app.use('/api/index', index); app.use('/api/products', products); app.use('/api/details', details); app.use('/api/users', users); app.use('/api/cart', cart); app.use('/api/comment', comment); app.use('/api/brand', brand); app.use('/api/dic', dictionaries); app.use('/api/order', order); app.use('/api/collection', collection); app.use('/api/address', address); app.use('/api/message', message); app.use('/api/admin', admin); app.listen(8000, () =>{ console.log('服务器创建成功8000!!!'); });
javascript
{ "type": "none", "comment": "Added missing comments for authentication", "packageName": "@microsoft/teams-js", "email": "<EMAIL>", "dependentChangeType": "none" }
json
{"id":33199,"group":"72213117","joblink":true,"name":"Finanzbuchhalter/in","potential":88,"skills":[{"id":62220,"skill":"Buchführung, Buchhaltung","replaceable":true},{"id":62225,"skill":"Finanzbuchhaltung","replaceable":true},{"id":62250,"skill":"Zahlungsverkehr","replaceable":true},{"id":62288,"skill":"Controlling","replaceable":false},{"id":62310,"skill":"Berichtswesen, Information","replaceable":true},{"id":62484,"skill":"Steuerrecht","replaceable":true},{"id":62535,"skill":"Jahresabschluss","replaceable":true},{"id":126816,"skill":"Rechnungslegung nach HGB","replaceable":true}]}
json
{"h":[{"d":[{"e":["`Miliyasay~ `ci~`nira~ `to~ `Taypak~ `anini~.他現在離開臺北。"],"f":"正在離開、脫離了。"},{"e":["`Ci~ `kaka~ `ako~ `ko~ `miliyasay~ `to~ `loma'~ `niyam~.是我哥哥離開我們家。"],"f":"指離開者,脫離者。"}]}],"stem":"liyas","t":"miliyasay"}
json
The meeting of State Cabinet held here today under the chairmanship of Chief Minister Jai Ram Thakur gave its nod to amendment and insertion in Himachal Pradesh Buildings and Other Construction Workers (Regulation of Employment and Condition of Service) Rules, 2008 to increase the ambit of the scheme and provide more benefits to the eligible beneficiaries. Now the children of workers from 1st to 8th standard would be provided Rs. 8400 per annum instead of Rs. 8000 to girls and 5000 to boys per year, Rs. 12,000 per annum for students from 9th to 10 plus 2 standards instead of Rs. 11,000 to girls and Rs. 8000 to boys, Rs. 36,000 per annum to graduation classes instead of Rs. 16,000 to girls and 12,000 to boys, Rs. 60,000 per annum to post graduation instead of Rs. 21,000 to girls and Rs. 17,000 to boys, Rs. 48000 per annum for diploma courses instead of Rs. 21,000 to girls and Rs. 17000 to boys, Rs. 60,000 for professional courses/degrees per annum instead of Rs. 36,000 to girls and Rs. 27,000 to boys and Rs. 1. 20 lakh per annum for PHD research courses instead of Rs. 36,000 being provided to girl students and Rs. 27,000 being given to boys for their studies. It also decided to start new scheme Female Birth Gift Scheme under which an FDR of Rs. 51,000 may be given on the birth of a female child, maximum upto two girls, differently abled and Mentally Retarded Children Benefit Scheme under which Rs. 20,000 per annum would be provided to the child with disability of 50 percent or above. It also decided to start Widow Pension of Rs. 1500 per month to the widows of registered beneficiaries. It also decided to start Hostel Facility Scheme under which a registered Building and Other Construction workers would receive an amount of Rs. 20,000 maximum for bearing the expenses incurred on account of lodging, boarding and food bill of his children living in any hostel and starting Mukhyamantri Awas Yojna under which registered beneficiary who is already enrolled under Pradhan Mantri Awas Yojna or Mukhya Mantri Awas Yojna would receive a financial assistance of Rs. 1,50,000 to build his/her own house. The Cabinet also reviewed covid-19 situation in the State and decided to continue with the present restrictions. The Cabinet gave its nod to constitute Cabinet Sub Committee under chairmanship of Jal Shakti Minister Mahender Singh Thakur and Education Minister Govind Thakur and Forest Minister Rakesh Pathania as its members to resolve the issue of compensation for land acquisition linked with four lane construction projects. The Sub Committee would examine the policy related to it in neighbouring States. The Cabinet also gave its nod to create Sub Division (Civil) at Kotli in Mandi district to facilitate the people of the area. It decided to open new veterinary dispensaries at Village Adhar in Gram Panchayat Bat of Chamba district alongwith creation and filling up of posts to manage these dispensaries. It also decided to create new Sub Division of Jal Shakti Department at Ranital under Jal Shakti Division Shahpur and a new Section under this Sub Division at Thakurdwara. It also gave its approval for creating new Section of Jal Shakti Department for Pandit Jawahar Lal Nehru Government Medical College Chamba (Sarol) under Udaipur Jal Shakti Sub Division. It also gave its nod for opening Atal Adarsh Vidyalaya at Chanol in Sundernagar Vidhan Sabha area of Mandi district. It decided to open new Health Block Office at Rajgarh in Sirmaur district alongwith creation of requisite posts. The Cabinet decided to open new Government Primary Schools at village Galang in Gram Panchayat Pichhali Dhar in Education Block Naggar and at village Sarali in Gram Panchayat Bastori in Education Block Kullu-II in Kullu district. It also decided to upgrade Government High Schools Tatwali in Fatehpur area and Nadholi in Jawali area of Kangra district to Government Senior Secondary Schools alongwith creation and filling up of 12 posts of different categories to man these schools. The Cabinet decided to upgrade Government Middle School Kareri Khas in Shahpur area and Government High School Jalot in Nagrota Bagwan in Kangra district to Government High School and Government Senior Secondary School respectively alongwith creation of requisite posts for smooth functioning. It gave its nod to upgrade Government Primary School, Ree Khas in Swarghat area of Bilaspur district to Government Middle School alongwith creation and filling up of three posts of different categories. The Cabinet gave its approval to upgrade Government Primary Schools, Jhangi, Dhamgran, Oyal and Kakla in Chamba constituency of Chamba district to Government Middle Schools alongwith creation and filling up of 12 posts of different categories. It decided to start commerce classes at Government Senior Secondary School Keolidhar in Mandi district. It also decided to start science classes at Government Senior Secondary Schools, Dadoh, Bassi, Bhakhli and Devdhar in Mandi district to facilitate the science students of the area. The Cabinet gave its consent to start science classes (Medical Biology) in Government Senior Secondary School, Gandhir in Jhanduta area of Bilaspur district alongwith creation of requisite posts to man this institution. It also decided to start science classes in Atal Bihari Vajpayee Government Degree College Takipur in Kangra district. It gave its nod to upgrade Government High School Jharmajri in Doon area of Solan district to Senior Secondary School alongwith creation and filling up of requisite posts. The Secretaries incharge of Home, PWD, Jal Shakti, Education, Health, Youth Services and Sports and Animal Husbandry briefed the Cabinet on Chief Minister’s announcements pertaining to their departments. The Cabinet accorded its approval for renewal of State Government's ‘No Default Guarantee’ in favour of Himachal Road Transport Corporation for availing case credit limit of Rs. 60 crore for the financial year 2021-22. It also gave its approval for creation and filling up of 15 posts of different categories for the each newly constituted Municipal Corporations of Solan, Mandi and Palampur. It also decided that two posts of JE, two posts of Sanitary Supervisors, one post of Junior Draftsman in place of Draftsman, three posts of Data entry operators in place of PA and four posts of JOA (IT) for each newly constituted MCs of Solan, Mandi and Palampur would be filled on outsource basis till these posts are filled up on regular basis. It gave its approval to fill up two posts of Himachal Pradesh Administrative Services through direct recruitment by holding Himachal Pradesh Combined Competitive Examination 2021 on regular basis. It also gave its consent to fill up 12 posts of Junior Technician (Electrical) in Public Works Department through direct recruitment on contract basis. It decided to fill up four posts of drivers for development blocks Pragpur, Kaza, Baijnath and Ghumarwin under Rural Development Department. The Cabinet decided to fill up six posts of Associate Director in Chaudhary Sarwan Kumar Himachal Pradesh Agriculture University Palampur. The Cabinet also gave its approval for construction of additional three suites in Rest House building Sarahan in Pachhad area of Sirmaur district. It also gave its nod for construction of new PWD Rest House at Baldwara in Mandi district. It also decided to construct additional accommodation at PWD Rest House at Shillai in Sirmaur district. The Cabinet decided to create new Division of Public Works Department at Thalaut in Mandi district alongwith creation of 11 posts of different categories. It approved creation of new Sub Division of PWD at Timbi in Sirmaur district alongwith creation of requisite posts for its smooth functioning. It gave its consent to curve out new Development Block at Baroh in Kangra district from Development Blocks Nagrota Bagwan and Kangra. It also gave its consent to include six gram panchayats of Development Block Bamsan in Development Block Hamirpur to facilitate the people of these panchayats.
english
How to choose an orthopedic mattress? Such a question now increasingly arises in people who want to make sleep sound and improve well-being. These designs allow the person's spine to stay straight through the night. But if you are interested in the question of how to choose an orthopedic mattress, you should adequately assess yourself and your complexion. Full people should prefer a rigid design. For example, a mattress equipped with reinforced springs, or a springless, made of dense materials. Thin people can afford a soft spring model or a springless, made of natural latex. But those who have a medium build, can choose from more options. Springless mattresses are recommended for children and teenagers. And for babies this is a well-ventilated model, and for older children - surviving jumps and loads. Teenagers and the elderly are recommended rigid mattresses. However, it is always necessary to consider the state of health. How to choose an orthopedic mattress in size? Let's say you liked the model 160x200. But is this the outer size of the product itself or the bed? Take the roulette safely and measure it. It is best to measure the internal size of the bed, otherwise the mattress will not fit into it. How to choose an orthopedic mattress? Given the filler! This is the most important parameter. A good mattress is a guarantee that your sleep will be strong and healthy. So, all models are divided into spring and springless. Each subspecies has its advantages, but it does not suit everyone. Here everything is individual. If you do not like the "pushing" effect, it is better to prefer a springless design. It is in this category that the largest selection of hard mattresses. They can be made of natural latex. Such options perfectly work out the contours of the body. Good orthopedic properties are also available in waterlatex, foam rubber, polyurethane foam. Strong are considered multi-layer mattresses, in which coconut with latex is combined. But those that are made only from coconut, the most stringent. Sleeping on them is tantamount to lying on the boards. Usually such models are prescribed to use a doctor. If you are interested in how to choose an orthopedic mattress, then you need to do this, taking into account the cover. Springless models can be covered with double-sided covers (winter-summer) made of natural or artificial materials. This significantly affects the price. If you do not know how to choose the right orthopedic mattress, then you should find out from consultants. They will explain to you, what are the advantages of the second kind of design - spring. The spring unit in them can be dependent and independent. The first supports the body in anatomically normal position thanks to the Bonnel springs. But this design is eventually squeezed. The independent block, unlike the previous one, consists of springs, each of which accurately distributes the load without involving "neighbors". Foams for all types of spring mattresses are wool, spunbond, felt, natural latex, coconut coir. If you are interested in what firm to choose an orthopedic mattress, then now the models DORMEO, MARIAMIS, "Marquis" and others are popular.
english
<reponame>Oscaner/exam-questions { "Update": "2020-03-23 15:28:16", "answer": [ "B", "C" ], "category": "saa", "choices": [ "A.Amazon DynamoDB ", "B.Amazon Elastic 计算 Cloud ( EC2 )", "C.Amazon Elastic Load Balancing ", "D.Amazon Simple Notification Service (SNS)", "E.Amazon Simple 存储 Service ( S3 )" ], "detail": "使用以下 AWS 服务时, 高可用性解决方案应在多个 Availability Zone 中实现哪些服务?选择两个答案", "explanation": [ "正确答案是B和C, 因为 ELB 可以通过跨多个AZ跨多个 EC2 实例路由流量来提供高可用性", " Elastic Load Balancing 在多AZ中的多个 Amazon EC2 实例之间自动分配传入的应用程序流量。它使您能够在应用程序中实现容错, 无缝地提供路由应用程序流量所需的负载平衡容量。", "选项A、D&E都是错误的, 因为它们都是 AWS 托管服务, 并且都是可扩展的, 并且都是HA。" ], "id": 73, "saved_answer": "01100", "type": "Multiple Answer" }
json
Waheeda Rehman and Guru Dutt Wonderful. That is the word. As a director, Guru Dutt was a master. As an actor, he was no less accomplished. Waheeda Rehman was close to perfect in her ability to match up to him. The two of them were in love in real life. They bonded beautifully on the big screen. The end results in films like Pyaasa and Kaagaz Ke Phool were outstanding.
english
<filename>mobile/src/assets/i18n/en.json { "LOGIN_PAGE": { "TITLE": "Welcome to<br> I FOR YOU", "SIGN_IN_GOOGLE": "Login with Google", "SIGN_IN_FACEBOOK": "Login with Facebook", "DESCR": "If you are between 18 and 64 years old, you can provide your local community with any kind of support - for instance, grocery home delivery. As a volunteer, you can share your spare time and skills to respond to any needs in your community and promote targeted social support when and where it is needed." }, "WELCOME": { "TITLE": "Welcome to I FOR YOU", "MESSAGE": "Wou can provide your local community with any kind of support. You can share your spare time and skills to respond to any needs in your community.", "CTA": "Sig in!" }, "REGISTRATION_PAGE": { "TITLE": "Register User", "EMAIL": "Email", "PASSWORD": "Password", "REGISTER": "Register" }, "POSITION_PICKER": { "TITLE": "Share your location to help your neighbors.", "SAVE": "Save my position", "DISCLAIMER": "your exact location data will not be recorded or shared" }, "CARDS_PAGE": { "TITLE": "Helpers near you!", "NEAR_YOU": "near you", "ABOUT_NEAR_YOU": "about {{distance}} near you", "NO_DISTANCE_LIMIT": "no distance limit" }, "FILTER_PAGE": { "FILTER_BY": "Filtra per", "DISTANCE": "Distanza", "AVAILABILITY": "Disponibilità" }, "SETTINGS_PAGE": { "TITLE": "Settings", "PRIVACY": "Privacy Policy", "TERMS": "Terms of Services", "DELETE": "Delete your data", "LOGOUT": "Logout", "LANGUAGE_SELECT": { "LABEL": "Select the app language", "LIST": { "IT": "Italian", "EN": "English" } } }, "PROFILE_PAGE": { "TITLE": "Profile", "FIRST_MESSAGE": "Hey, I can take care of you.", "CONFIG_MESSAGE": "Can You go to the market? Pharmacy?", "NAME": "Name", "SURNAME": "Surname", "COUNTRY": "Country", "CITY": "City", "ADDRESS": "Place", "PHONE": "Phone", "SKYPE": "Skype", "MESSAGE_HELP": "I can help you with:", "AVAILABLE": "You ara available to help people, help us?", "ROLE_TYPE": { "FOOD": "food", "FARMACY": "farmacy", "MAIL": "mail", "COMPANY": "company" }, "YESIAM": "Yes, I am!", "NEXT": "How do you want to help?", "SAVE": "Save", "SIGN_UP": "Sign up", "CLOSE": "Close", "POSITION": "Share your location", "LOADER_MESSAGE": "Saving your profile", "SET_POSITION_MESSAGE": "To help us you need to share your location" }, "COMMON": { "UNIT_DISTANCE": "m", "UNIT_DISTANCE_K": "km", "ACTIVITY_TYPES": { "FOOD": "Food", "PHARMACY": "Pharmacy", "MAIL": "Mail", "COMPANY": "Company" }, "OK_BUTTON": "Ok", "CANCEL_BUTTON": "Cancel", "AVAILABILITY_TYPES": { "ALL_TIME": "All time", "MORNING": "Morning", "AFTER_NOON": "After noon", "EVENING": "Evening" } }, "EXIT": { "BACK_BUTTON": "Press back again to exit App" } }
json
Returnal, a critically acclaimed PlayStation 5 exclusive, has finally made its way to PC after almost two years of its initial release. The game boasts top-of-the-line graphics that fully take advantage of the next-gen hardware, providing players with a truly immersive experience. Developed by Housemarque, it showcases stunning visuals that bring the game's world to life and create an atmosphere that truly draws players in. RTX 2060 and RTX 2060 Super are mid-range GPU offerings from Nvidia. They are the first generation of RTX cards that made real-time ray tracing in games a reoccurring theme. Apart from RT, these were also the first GPUs that came with DLSS support, enabling players to achieve higher framerates with the same graphics settings. Despite their age, these two cards perform surprisingly well in early 2023. While some minor compromises may be needed to run the latest AAA games, these cards still manage to deliver a playable experience that gets the job done. Returnal's PC port is considered to be one of the best in recent times. The game is brilliantly optimized for PCs with only minor nuances, such as stuttering, that have been reported by a few players. Other than that, the experience is almost flawless. However, as mentioned earlier, the RTX 2060 and the RTX 2060 Super users will have to make a few compromises in this latest launch to arrive at stable results. The settings this guide suggests will provide a balanced experience that brings the best of both visuals and framerates. Keeping this in mind, here are the best graphics settings in Returnal for the RTX 2060 and RTX 2060 Super: - Display Monitor: Your primary monitor. - UI Contrast: As per the user's preference. - Field of View: As per the user's preference. - Depth of Field: As per the user's preference. - Film Grain: As per the user's preference. - Emissive Intensity: As per the user's preference. - Brightness: As per the user's preference. - Contrast: As per the user's preference. - Display Monitor: Your primary monitor. - UI Contrast: As per the user's preference. - Field of View: As per the user's preference. - Depth of Field: As per the user's preference. - Film Grain: As per the user's preference. - Emissive Intensity: As per the user's preference. - Brightness: As per the user's preference. - Contrast: As per the user's preference. The recommended settings provide a balanced experience with the best image quality and frame rates in Returnal with the RTX 2060 and RTX 2060 Super. However, if players are not satisfied with the results, they should start off with these settings and then tweak them as per their preference.
english
Don’t Miss Out on the Latest Updates. Subscribe to Our Newsletter Today! A driver and his relative were arrested from suburban Andheri for allegedly staging kidnapping of two children of a builder, for whom he worked, in a bid to extort Rs 1 crore, police said on Wednesday. During the interrogation, the driver admitted that he had hatched the conspiracy of kidnapping the twin children to get money for his daughter’s marriage, an official said. “The incident came to light when the father of the children, a real estate developer, approached the police on Monday evening with the complaint that his two kids had been kidnapped,” he said. “He said the twins were abducted after his driver was beaten by the kidnappers when they were on their way back to home in Juhu,” the official said. A kidnapper forcibly opened the door of the car and allegedly threatened the children and the driver and took their car near Juhu PVR area, the complaint said. “During the journey, the kidnapper gave sedatives to the children and the driver. It said the kidnapper got down with one child and the driver and put them in a school bus, while another child was kept in the car near Juhu PVR,” he added. After that, six kidnappers on three bikes took the car driver out of the bus and took him on Juhu-Versova Road and allegedly thrashed him, the complaint said. A police team reached the Juhu PVR area and rescued one child, while the second child escaped with the help of people and alerted his family members about the incident. “Meanwhile, the mother of the twins had received a call with the callers seeking Rs 1 crore,” the police official said. “But as there were loose ends in the complaint, the police started interrogating the car driver. After questioning him for 18 hours, he told the police that he had staged the kidnapping drama to get Rs 1 crore from his master for his daughter’s marriage,” he said. The accused driver also had called one of his relative from Delhi to be part of the conspiracy and had promised him half of the extortion money, he said. Police have arrested both the accused and claimed to have cracked the case. An FIR was registered at D N Nagar Police Station under various sections of the IPC. For breaking news and live news updates, like us on Facebook or follow us on Twitter and Instagram. Read more on Latest Viral News on India. com.
english
def xor_reverse(iterable): lenght = len(iterable) i = 0 while i < lenght // 2: iterable[i] ^= iterable[lenght - i - 1] iterable[lenght - i - 1] ^= iterable[i] iterable[i] ^= iterable[lenght - i - 1] i += 1 return iterable
python
<filename>src/App.js import React, { useState } from "react"; import Container from "react-bootstrap/Container"; import "./App.css"; import { initialApplicationContext } from "./context/ApplicationContext"; import ApplicationProvider from "./context/ApplicationProvider"; import Routes from "./Routes.js"; function App() { const [appContext, setAppContext] = useState(initialApplicationContext); return ( <Container fluid={true}> <ApplicationProvider> <Routes /> </ApplicationProvider> </Container> ); } export default App;
javascript
Gay porn star Al Parker was born Drew Okum in Natick, Massachusetts. When he was a teenager his loaned him their brand-new Mustang to attend Woodstock, which they thought was a classical music festival. While at Woodstock, Drew had several homosexual experiences, the most notable in the back of a van--a theme that would be repeated throughout his adult-film career. Nickname: Pony Boy After graduating from high school in Natick, he moved to Los Angeles, where he landed a job operating the film projector booth at 'Hugh Hefner"s Playboy mansion. He was "discovered" by gay-porn producer Rip Colt, who changed his name to Al Parker because it sounded like a man's name. Parker starred in several of Colt's 8mm shorts. Parker began a long-term, and open, lover relationship with Steve Taylor; together the two established Surge Studios, a gay-adult film production studio. Following Taylor's sudden death from AIDS, Parker's interest in the studio faded. Parker himself serio-converted and tested positive, following which he made several "safer" porn movies with other HIV-positive actors. He died of complications from AIDS in San Francisco, California, on August 17, 1992.
english
Manny Machado has signed a blockbuster 11-year deal with the San Diego Padres worth $350 million, thereby ending weeks of uncertainty regarding his future with the Friars. It was only a little over a week ago that Machado made public his intentions to opt out of his contract at the end of the season. However, in a spectacular turnaround, it now seems very likely Machado will remain a part of the stellar Padres lineup for the years to come. The staggering deal sees Machado earn an average of almost $32 million every year. But that still does not make him the highest-paid player in the MLB. BREAKING: Star third baseman Manny Machado and the San Diego Padres are finalizing an 11-year, $350 million contract extension, sources familiar with the deal tell ESPN. Machado helped turn around the franchise. He'll stay as the ascendant Padres seek their first championship. Is Manny Machado the highest-paid player in the MLB? In terms of Average Annual Value (AAV), Max Scherzer and Justin Verlander, both of the New York Mets, are the highest-paid players in MLB. Both have an estimated AAV of $43. 33 million. They are followed by New York Yankees captain Aaron Judge, who recently signed a nine-year contract with an AAV of $40 million. Machado is preceded by a few other players on this list, namely LA Angels' Anthony Rendon and Mike Trout, Yankees' Gerrit Cole, and Texas Rangers shortstop Corey Seager, among others. His current AAV places him among the top-15 highest-paid players in the MLB, alongside the likes of the Yankees' Giancarlo Stanton and Detroit Tigers' DH Miguel Cabrera.
english
After Navi Mumbai, now Thane Municipal Corporation (TMC) announced a 24-hour water cut in several parts of the city today, January 18. This move has been taken in order to carry out essential works in the TMC’s water supply scheme and STEM authority. The water supply will remain completely shut till 9 am tomorrow, January 19. It will be affected in these areas: Ghodbunder Road, Lokmanya Nagar, Vartak Nagar, Saket, Ritu Park, Thane central Jail, Gandhi Nagar, Rustamjee complex, Siddanchal complex, Indira Nagar, Srinagar, Samta Nagar, Siddheshwar, Eternity Mall, and some parts of Mumbra and Kalwa. Explaining further, the civic body officials stated that due to the shutdown, water supply is likely to be at low pressure for the next 1 to 2 days until it is fully restored. The essential works include the removal of leaks, installation of vacuum air valves, connecting the new 1,168 mm new water pipeline at Indira Nagar to the main water channel and carrying out essential daily maintenance and repair work in the water supply. Officials clarified that they had given prior notification in these areas about the water cut and asked the residents to store adequate water. These works are crucial for the smooth supply of water in future.
english
["api_500px","ar-drone","ar-drone-browserified","ardrone-autonomy","ardrone-autonomy-withsim","asterisk-ami","blogdown","buffy","felix-metrics","formidable","formidable7","graphite","hub-namespace","hubjs-fork","lazy-socket","limo","mtfe_orm","mysql","mysql-ali","mysql-alpha","mysql-clusterfarm","mysql-mmx","mysql-robin","mysql2","node-google-prediction","nodejs-multisock","nomo","patillades_formidable_fork","pubsubway","reskit","sql-query","sqlstring","streamchunker","url2image","urun","vn-mysql","vogievetsky-mysql2"]
json
<reponame>hashnfv/hashnfv-moon<gh_stars>0 # Copyright 2014 IBM Corp. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from keystone.common import controller from keystone.common import dependency from keystone import notifications @dependency.requires('policy_api', 'catalog_api', 'endpoint_policy_api') class EndpointPolicyV3Controller(controller.V3Controller): collection_name = 'endpoints' member_name = 'endpoint' def __init__(self): super(EndpointPolicyV3Controller, self).__init__() notifications.register_event_callback( 'deleted', 'endpoint', self._on_endpoint_delete) notifications.register_event_callback( 'deleted', 'service', self._on_service_delete) notifications.register_event_callback( 'deleted', 'region', self._on_region_delete) notifications.register_event_callback( 'deleted', 'policy', self._on_policy_delete) def _on_endpoint_delete(self, service, resource_type, operation, payload): self.endpoint_policy_api.delete_association_by_endpoint( payload['resource_info']) def _on_service_delete(self, service, resource_type, operation, payload): self.endpoint_policy_api.delete_association_by_service( payload['resource_info']) def _on_region_delete(self, service, resource_type, operation, payload): self.endpoint_policy_api.delete_association_by_region( payload['resource_info']) def _on_policy_delete(self, service, resource_type, operation, payload): self.endpoint_policy_api.delete_association_by_policy( payload['resource_info']) @controller.protected() def create_policy_association_for_endpoint(self, context, policy_id, endpoint_id): """Create an association between a policy and an endpoint.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_endpoint(endpoint_id) self.endpoint_policy_api.create_policy_association( policy_id, endpoint_id=endpoint_id) @controller.protected() def check_policy_association_for_endpoint(self, context, policy_id, endpoint_id): """Check an association between a policy and an endpoint.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_endpoint(endpoint_id) self.endpoint_policy_api.check_policy_association( policy_id, endpoint_id=endpoint_id) @controller.protected() def delete_policy_association_for_endpoint(self, context, policy_id, endpoint_id): """Delete an association between a policy and an endpoint.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_endpoint(endpoint_id) self.endpoint_policy_api.delete_policy_association( policy_id, endpoint_id=endpoint_id) @controller.protected() def create_policy_association_for_service(self, context, policy_id, service_id): """Create an association between a policy and a service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.endpoint_policy_api.create_policy_association( policy_id, service_id=service_id) @controller.protected() def check_policy_association_for_service(self, context, policy_id, service_id): """Check an association between a policy and a service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.endpoint_policy_api.check_policy_association( policy_id, service_id=service_id) @controller.protected() def delete_policy_association_for_service(self, context, policy_id, service_id): """Delete an association between a policy and a service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.endpoint_policy_api.delete_policy_association( policy_id, service_id=service_id) @controller.protected() def create_policy_association_for_region_and_service( self, context, policy_id, service_id, region_id): """Create an association between a policy and region+service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.catalog_api.get_region(region_id) self.endpoint_policy_api.create_policy_association( policy_id, service_id=service_id, region_id=region_id) @controller.protected() def check_policy_association_for_region_and_service( self, context, policy_id, service_id, region_id): """Check an association between a policy and region+service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.catalog_api.get_region(region_id) self.endpoint_policy_api.check_policy_association( policy_id, service_id=service_id, region_id=region_id) @controller.protected() def delete_policy_association_for_region_and_service( self, context, policy_id, service_id, region_id): """Delete an association between a policy and region+service.""" self.policy_api.get_policy(policy_id) self.catalog_api.get_service(service_id) self.catalog_api.get_region(region_id) self.endpoint_policy_api.delete_policy_association( policy_id, service_id=service_id, region_id=region_id) @controller.protected() def get_policy_for_endpoint(self, context, endpoint_id): """Get the effective policy for an endpoint.""" self.catalog_api.get_endpoint(endpoint_id) ref = self.endpoint_policy_api.get_policy_for_endpoint(endpoint_id) # NOTE(henry-nash): since the collection and member for this class is # set to endpoints, we have to handle wrapping this policy entity # ourselves. self._add_self_referential_link(context, ref) return {'policy': ref} # NOTE(henry-nash): As in the catalog controller, we must ensure that the # legacy_endpoint_id does not escape. @classmethod def filter_endpoint(cls, ref): if 'legacy_endpoint_id' in ref: ref.pop('legacy_endpoint_id') return ref @classmethod def wrap_member(cls, context, ref): ref = cls.filter_endpoint(ref) return super(EndpointPolicyV3Controller, cls).wrap_member(context, ref) @controller.protected() def list_endpoints_for_policy(self, context, policy_id): """List endpoints with the effective association to a policy.""" self.policy_api.get_policy(policy_id) refs = self.endpoint_policy_api.list_endpoints_for_policy(policy_id) return EndpointPolicyV3Controller.wrap_collection(context, refs)
python
<reponame>DmitryBogomolov/aws-cloudformation-sample<filename>xawscf/pattern/statemachine.py import json from ..utils.loader import Custom from ..utils.helper import get_full_name, make_output from .base_resource import BaseResource from .role import Role class StateMachineRole(Role): PRINCIPAL_SERVICE = Custom('!Sub', 'states.${AWS::Region}.amazonaws.com') def get_role_name(name): return name + 'Role' class StateMachine(BaseResource): TEMPLATE = \ ''' Type: AWS::StepFunctions::StateMachine Properties: DefinitionString: !Sub null ''' TYPE = 'statemachine' def __init__(self, *args): super().__init__(*args) self.full_name = get_full_name(self.name, self.root) def _dump(self, template, parent_template): super()._dump(template, parent_template) name = self.name role_name = get_role_name(name) StateMachineRole(role_name, { 'statement': self.get('role_statement', []) }, self.root).dump(parent_template) outputs = parent_template['Outputs'] outputs[name] = make_output(Custom('!GetAtt', name + '.Name')) outputs[name + 'Arn'] = make_output(Custom('!Ref', name)) template['DependsOn'].append(role_name) def _dump_properties(self, properties): properties['StateMachineName'] = self.full_name properties['RoleArn'] = Custom('!GetAtt', get_role_name(self.name) + '.Arn') definition = self.get('definition') if isinstance(definition, dict): definition = json.dumps(definition, indent=2) properties['DefinitionString'].value = [ definition, self.get('definition_args', {}) ]
python
<gh_stars>1-10 #include "dx/dx10/DX10Buffer.hpp" #ifdef _WIN32 #include <Safeties.hpp> #include "dx/dx10/DX10GraphicsInterface.hpp" #include "dx/dx10/DX10RenderingContext.hpp" #include "TauEngine.hpp" #if TAU_RTTI_CHECK #define CTX() \ if(!RTT_CHECK(context, DX10RenderingContext)) \ { TAU_THROW(IncorrectContextException); } \ auto& ctx = reinterpret_cast<DX10RenderingContext&>(context) #else #define CTX() \ auto& ctx = reinterpret_cast<DX10RenderingContext&>(context) #endif bool DX10VertexBuffer::beginModification(IRenderingContext&) noexcept { if(canReWrite(_usage)) { #if TAU_BUFFER_SAFETY ++_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount > 1) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyBegin); return false; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { _currentMapping = bufferAccess; return true; } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } return false; } void DX10VertexBuffer::endModification(IRenderingContext&) noexcept { if(_currentMapping) { _d3dBuffer->Unmap(); _currentMapping = null; #if TAU_BUFFER_SAFETY --_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount < 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyEnd); } #endif #endif } } void DX10VertexBuffer::modifyBuffer(const uSys offset, const uSys size, const void* const data) noexcept { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_MODIFY_WITHOUT_BEGIN if(_modificationLockCount <= 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedWithoutBegin); return; } #endif #endif if(_currentMapping) { ::std::memcpy(reinterpret_cast<u8*>(_currentMapping) + offset, data, size); } } void DX10VertexBuffer::fillBuffer(IRenderingContext&, const void* const data) noexcept { if(DX10VertexBuffer::canReWrite(_usage)) { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_FILLED_WHILE_MODIFYING if(_modificationLockCount > 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::FilledWhileModifying); return; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { ::std::memcpy(bufferAccess, data, _bufferSize); _d3dBuffer->Unmap(); } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } } bool DX10IndexBuffer::beginModification(IRenderingContext&) noexcept { if(DX10VertexBuffer::canReWrite(_usage)) { #if TAU_BUFFER_SAFETY ++_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount > 1) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyBegin); return false; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { _currentMapping = bufferAccess; return true; } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } return false; } void DX10IndexBuffer::endModification(IRenderingContext&) noexcept { if(_currentMapping) { _d3dBuffer->Unmap(); _currentMapping = null; #if TAU_BUFFER_SAFETY --_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount < 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyEnd); } #endif #endif } } void DX10IndexBuffer::modifyBuffer(const uSys offset, const uSys size, const void* const data) noexcept { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_MODIFY_WITHOUT_BEGIN if(_modificationLockCount <= 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedWithoutBegin); return; } #endif #endif if(_currentMapping) { ::std::memcpy(reinterpret_cast<u8*>(_currentMapping) + offset, data, size); } } void DX10IndexBuffer::fillBuffer(IRenderingContext&, const void* const data) noexcept { if(DX10VertexBuffer::canReWrite(_usage)) { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_FILLED_WHILE_MODIFYING if(_modificationLockCount > 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::FilledWhileModifying); return; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { ::std::memcpy(bufferAccess, data, _bufferSize); _d3dBuffer->Unmap(); } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } } void DX10UniformBuffer::bind(IRenderingContext& context, const EShader::Stage stage, const u32 index) const noexcept { #if TAU_BUFFER_SAFETY ++_uniformBindLockCount; #if TAU_BUFFER_SAFETY_UNIFORM_DOUBLE_BIND if(_uniformBindLockCount > 1) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleUniformBufferBind); } #endif #endif CTX(); switch(stage) { case EShader::Stage::Vertex: ctx.d3dDevice()->VSSetConstantBuffers(index, 1, &_d3dBuffer); break; case EShader::Stage::Geometry: ctx.d3dDevice()->GSSetConstantBuffers(index, 1, &_d3dBuffer); break; case EShader::Stage::Pixel: ctx.d3dDevice()->PSSetConstantBuffers(index, 1, &_d3dBuffer); break; default: break; } } void DX10UniformBuffer::unbind(IRenderingContext& context, const EShader::Stage stage, const u32 index) const noexcept { #if TAU_BUFFER_SAFETY --_uniformBindLockCount; #if TAU_BUFFER_SAFETY_UNIFORM_DOUBLE_UNBIND if(_uniformBindLockCount < 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleUniformBufferUnbind); } #endif #endif CTX(); switch(stage) { case EShader::Stage::Vertex: ctx.d3dDevice()->VSSetConstantBuffers(index, 0, null); break; case EShader::Stage::Geometry: ctx.d3dDevice()->GSSetConstantBuffers(index, 0, null); break; case EShader::Stage::Pixel: ctx.d3dDevice()->PSSetConstantBuffers(index, 0, null); break; default: break; } } void DX10UniformBuffer::fastUnbind() const noexcept { #if TAU_BUFFER_SAFETY --_uniformBindLockCount; #if TAU_BUFFER_SAFETY_UNIFORM_DOUBLE_UNBIND if(_uniformBindLockCount < 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleUniformBufferUnbind); } #endif #endif } bool DX10UniformBuffer::beginModification(IRenderingContext&) noexcept { if(DX10VertexBuffer::canReWrite(_usage)) { #if TAU_BUFFER_SAFETY ++_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount > 1) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyBegin); return false; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { _currentMapping = bufferAccess; return true; } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } return false; } void DX10UniformBuffer::endModification(IRenderingContext&) noexcept { if(_currentMapping) { _d3dBuffer->Unmap(); _currentMapping = null; #if TAU_BUFFER_SAFETY --_modificationLockCount; #if TAU_BUFFER_SAFETY_DOUBLE_MODIFY_BEGIN if(_modificationLockCount < 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::DoubleModifyEnd); } #endif #endif } } void DX10UniformBuffer::modifyBuffer(const uSys offset, const uSys size, const void* const data) noexcept { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_MODIFY_WITHOUT_BEGIN if(_modificationLockCount <= 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedWithoutBegin); return; } #endif #endif if(_currentMapping) { ::std::memcpy(reinterpret_cast<u8*>(_currentMapping) + offset, data, size); } } void DX10UniformBuffer::fillBuffer(IRenderingContext&, const void* const data) noexcept { if(DX10VertexBuffer::canReWrite(_usage)) { #if TAU_BUFFER_SAFETY #if TAU_BUFFER_SAFETY_FILLED_WHILE_MODIFYING if(_modificationLockCount > 0) { TAU_THROW(BufferSafetyException, BufferSafetyException::FilledWhileModifying); return; } #endif #endif void* bufferAccess; const HRESULT h = _d3dBuffer->Map(D3D10_MAP_WRITE_DISCARD, 0, &bufferAccess); if(!FAILED(h)) { ::std::memcpy(bufferAccess, data, _bufferSize); _d3dBuffer->Unmap(); } } else { #if TAU_BUFFER_SAFETY_MODIFIED_STATIC_BUFFER TAU_THROW(BufferSafetyException, BufferSafetyException::ModifiedStaticBuffer); #endif } } DX10VertexBuffer* DX10BufferBuilder::build(const VertexBufferArgs& args, Error* const error) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } DX10VertexBuffer* const buffer = new(::std::nothrow) DX10VertexBuffer(args.usage, args.bufferSize(), args.descriptor.build(), d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } DX10VertexBuffer* DX10BufferBuilder::build(const VertexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } DX10VertexBuffer* const buffer = allocator.allocateT<DX10VertexBuffer>(args.usage, args.bufferSize(), args.descriptor.build(), d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } CPPRef<IVertexBuffer> DX10BufferBuilder::buildCPPRef(const VertexBufferArgs& args, Error* const error) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const CPPRef<DX10VertexBuffer> buffer(new(::std::nothrow) DX10VertexBuffer(args.usage, args.bufferSize(), args.descriptor.build(), d3dBuffer)); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableRef<IVertexBuffer> DX10BufferBuilder::buildTauRef(const VertexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const NullableRef<DX10VertexBuffer> buffer(allocator, args.usage, args.bufferSize(), args.descriptor.build(), d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableStrongRef<IVertexBuffer> DX10BufferBuilder::buildTauSRef(const VertexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const NullableStrongRef<DX10VertexBuffer> buffer(allocator, args.usage, args.bufferSize(), args.descriptor.build(), d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } DX10IndexBuffer* DX10BufferBuilder::build(const IndexBufferArgs& args, Error* const error) const noexcept { DXIndexBufferArgs dxArgs; if(!processArgs(args, &dxArgs, error)) { return null; } DX10IndexBuffer* const buffer = new(::std::nothrow) DX10IndexBuffer(args.usage, args.indexSize, args.bufferSize(), dxArgs.indexSize, dxArgs.d3dBuffer); if(!buffer) { dxArgs.d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } DX10IndexBuffer* DX10BufferBuilder::build(const IndexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { DXIndexBufferArgs dxArgs; if(!processArgs(args, &dxArgs, error)) { return null; } DX10IndexBuffer* const buffer = allocator.allocateT<DX10IndexBuffer>(args.usage, args.indexSize, args.bufferSize(), dxArgs.indexSize, dxArgs.d3dBuffer); if(!buffer) { dxArgs.d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } CPPRef<IIndexBuffer> DX10BufferBuilder::buildCPPRef(const IndexBufferArgs& args, Error* const error) const noexcept { DXIndexBufferArgs dxArgs; if(!processArgs(args, &dxArgs, error)) { return null; } const CPPRef<DX10IndexBuffer> buffer(new(::std::nothrow) DX10IndexBuffer(args.usage, args.indexSize, args.bufferSize(), dxArgs.indexSize, dxArgs.d3dBuffer)); if(!buffer) { dxArgs.d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableRef<IIndexBuffer> DX10BufferBuilder::buildTauRef(const IndexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { DXIndexBufferArgs dxArgs; if(!processArgs(args, &dxArgs, error)) { return null; } const NullableRef<DX10IndexBuffer> buffer(allocator, args.usage, args.indexSize, args.bufferSize(), dxArgs.indexSize, dxArgs.d3dBuffer); if(!buffer) { dxArgs.d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableStrongRef<IIndexBuffer> DX10BufferBuilder::buildTauSRef(const IndexBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { DXIndexBufferArgs dxArgs; if(!processArgs(args, &dxArgs, error)) { return null; } const NullableStrongRef<DX10IndexBuffer> buffer(allocator, args.usage, args.indexSize, args.bufferSize(), dxArgs.indexSize, dxArgs.d3dBuffer); if(!buffer) { dxArgs.d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } DX10UniformBuffer* DX10BufferBuilder::build(const UniformBufferArgs& args, Error* const error) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } DX10UniformBuffer* const buffer = new(::std::nothrow) DX10UniformBuffer(args.usage, args.bufferSize, d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } DX10UniformBuffer* DX10BufferBuilder::build(const UniformBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } DX10UniformBuffer* const buffer = allocator.allocateT<DX10UniformBuffer>(args.usage, args.bufferSize, d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } CPPRef<IUniformBuffer> DX10BufferBuilder::buildCPPRef(const UniformBufferArgs& args, Error* const error) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const CPPRef<DX10UniformBuffer> buffer(new(::std::nothrow) DX10UniformBuffer(args.usage, args.bufferSize, d3dBuffer)); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableRef<IUniformBuffer> DX10BufferBuilder::buildTauRef(const UniformBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const NullableRef<DX10UniformBuffer> buffer(allocator, args.usage, args.bufferSize, d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } NullableStrongRef<IUniformBuffer> DX10BufferBuilder::buildTauSRef(const UniformBufferArgs& args, Error* const error, TauAllocator& allocator) const noexcept { ID3D10Buffer* d3dBuffer; if(!processArgs(args, &d3dBuffer, error)) { return null; } const NullableStrongRef<DX10UniformBuffer> buffer(allocator, args.usage, args.bufferSize, d3dBuffer); if(!buffer) { d3dBuffer->Release(); ERROR_CODE_N(Error::SystemMemoryAllocationFailure); } ERROR_CODE_V(Error::NoError, buffer); } bool DX10BufferBuilder::processArgs(const VertexBufferArgs& args, ID3D10Buffer** const d3dBuffer, Error* const error) const noexcept { ERROR_CODE_COND_F(args.usage == static_cast<EBuffer::UsageType>(0), Error::UsageIsUnset); ERROR_CODE_COND_F(args.elementCount == 0, Error::BufferSizeIsZero); D3D10_BUFFER_DESC bufferDesc; bufferDesc.ByteWidth = args.bufferSize(); bufferDesc.Usage = DX10VertexBuffer::getDXUsage(args.usage); bufferDesc.BindFlags = D3D10_BIND_VERTEX_BUFFER; bufferDesc.CPUAccessFlags = DX10VertexBuffer::getDXAccess(args.usage); bufferDesc.MiscFlags = 0; if(args.initialBuffer) { D3D10_SUBRESOURCE_DATA initialBuffer; initialBuffer.pSysMem = args.initialBuffer; initialBuffer.SysMemPitch = 0; initialBuffer.SysMemSlicePitch = 0; const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, &initialBuffer, d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } else { const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, NULL, d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } return true; } bool DX10BufferBuilder::processArgs(const IndexBufferArgs& args, DXIndexBufferArgs* const dxArgs, Error* const error) const noexcept { ERROR_CODE_COND_F(args.usage == static_cast<EBuffer::UsageType>(0), Error::UsageIsUnset); ERROR_CODE_COND_F(args.elementCount == 0, Error::BufferSizeIsZero); dxArgs->indexSize = DX10IndexBuffer::dxIndexSize(args.indexSize); D3D10_BUFFER_DESC bufferDesc; bufferDesc.ByteWidth = args.bufferSize(); bufferDesc.Usage = DX10VertexBuffer::getDXUsage(args.usage); bufferDesc.BindFlags = D3D10_BIND_INDEX_BUFFER; bufferDesc.CPUAccessFlags = DX10VertexBuffer::getDXAccess(args.usage); bufferDesc.MiscFlags = 0; if(args.initialBuffer) { D3D10_SUBRESOURCE_DATA initialBuffer; initialBuffer.pSysMem = args.initialBuffer; initialBuffer.SysMemPitch = 0; initialBuffer.SysMemSlicePitch = 0; const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, &initialBuffer, &dxArgs->d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } else { const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, NULL, &dxArgs->d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } return true; } bool DX10BufferBuilder::processArgs(const UniformBufferArgs& args, ID3D10Buffer** const d3dBuffer, Error* const error) const noexcept { ERROR_CODE_COND_F(args.usage == static_cast<EBuffer::UsageType>(0), Error::UsageIsUnset); ERROR_CODE_COND_F(args.bufferSize == 0, Error::BufferSizeIsZero); D3D10_BUFFER_DESC bufferDesc; bufferDesc.ByteWidth = args.bufferSize; bufferDesc.Usage = DX10VertexBuffer::getDXUsage(args.usage); bufferDesc.BindFlags = D3D10_BIND_CONSTANT_BUFFER; bufferDesc.CPUAccessFlags = D3D10_CPU_ACCESS_WRITE; bufferDesc.MiscFlags = 0; if(args.initialBuffer) { D3D10_SUBRESOURCE_DATA initialBuffer; initialBuffer.pSysMem = args.initialBuffer; initialBuffer.SysMemPitch = 0; initialBuffer.SysMemSlicePitch = 0; const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, &initialBuffer, d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } else { const HRESULT h = _gi.d3d10Device()->CreateBuffer(&bufferDesc, NULL, d3dBuffer); ERROR_CODE_COND_F(FAILED(h), Error::DriverMemoryAllocationFailure); } return true; } D3D10_USAGE DX10VertexBuffer::getDXUsage(const EBuffer::UsageType usage) noexcept { switch(usage) { case EBuffer::UsageType::StreamDraw: case EBuffer::UsageType::StreamCopy: case EBuffer::UsageType::DynamicDraw: case EBuffer::UsageType::DynamicCopy: return D3D10_USAGE_DYNAMIC; case EBuffer::UsageType::StaticRead: case EBuffer::UsageType::StreamRead: case EBuffer::UsageType::DynamicRead: return D3D10_USAGE_STAGING; case EBuffer::UsageType::StaticCopy: case EBuffer::UsageType::StaticDraw: default: return D3D10_USAGE_DEFAULT; } } D3D10_CPU_ACCESS_FLAG DX10VertexBuffer::getDXAccess(const EBuffer::UsageType usage) noexcept { switch(usage) { case EBuffer::UsageType::StreamDraw: case EBuffer::UsageType::StreamCopy: case EBuffer::UsageType::DynamicDraw: case EBuffer::UsageType::DynamicCopy: return D3D10_CPU_ACCESS_WRITE; case EBuffer::UsageType::StaticRead: case EBuffer::UsageType::StreamRead: case EBuffer::UsageType::DynamicRead: return D3D10_CPU_ACCESS_READ; case EBuffer::UsageType::StaticCopy: case EBuffer::UsageType::StaticDraw: default: return static_cast<D3D10_CPU_ACCESS_FLAG>(0); } } bool DX10VertexBuffer::canReWrite(const EBuffer::UsageType usage) noexcept { switch(usage) { case EBuffer::UsageType::StreamDraw: case EBuffer::UsageType::StreamCopy: case EBuffer::UsageType::DynamicDraw: case EBuffer::UsageType::DynamicCopy: return true; case EBuffer::UsageType::StaticDraw: case EBuffer::UsageType::StaticCopy: case EBuffer::UsageType::StaticRead: case EBuffer::UsageType::DynamicRead: case EBuffer::UsageType::StreamRead: default: return false; } } DXGI_FORMAT DX10IndexBuffer::dxIndexSize(const EBuffer::IndexSize indexSize) noexcept { switch(indexSize) { case EBuffer::IndexSize::Uint32: return DXGI_FORMAT_R32_UINT; case EBuffer::IndexSize::Uint16: return DXGI_FORMAT_R16_UINT; default: return static_cast<DXGI_FORMAT>(0); } } #endif
cpp
68 May God rise up. May those who hate Him be divided. And may they run away from Him. 2 Drive them away like smoke in the wind. Let the sinful be destroyed before God like a candle melts by the fire. 3 But let those who are right and good be glad. Let them be happy before God. Yes, let them be full of joy. 4 Sing to God. Sing praises to His name. Make a road for Him Who goes through the deserts. The Lord is His name. Be full of joy before Him. 5 God in His holy house is a father to those who have no father. And He keeps the women safe whose husbands have died. 6 God makes a home for those who are alone. He leads men out of prison into happiness and well-being. But those who fight against Him live in an empty desert. 7 O God, when You went out before Your people, when You walked through the desert, 8 the earth shook. The heavens poured down rain before God. And Sinai shook before God, the God of Israel. 9 You sent a heavy rain, O God. You brought life back to Your promised land when it was dry. 10 Your people made it their home. O God, You gave the poor what they needed because You are good. 11 The Lord gives the Word. And the women who tell the good news are many. 12 The kings of armies run. They run away. And she who stays at home divides the riches. 13 When you lie down among the sheep, you are like the wings of a dove covered with silver, and the end of its wings with shining gold. 14 When the All-powerful divided the kings there, snow was falling in Zalmon. 15 A mountain of God is the mountain of Bashan. A mountain of many high tops is the mountain of Bashan. 16 O mountains of many high tops, why do you look with jealousy at the mountain which God has chosen for His home? For sure, the Lord will live there forever. 17 The war-wagons of God are 20,000, even thousands of thousands. The Lord is among them, as at Sinai, in the holy place. 18 You have gone up on high. You have taken those who were held with You. You have received gifts of men, even among those who fought against You. So the Lord God may live there with them. 24 They have seen Your people walking together, O God. They have seen the people of my God and King walking into the holy place. 25 Those who sing are in front. Those who play music are behind. And young women who beat timbrels are between. 26 Give thanks to God in the meetings of worship. Give thanks to the Lord, you who are of the family of Israel. 27 There is young Benjamin leading them. The rulers of Judah are among the people, and the rulers of Zebulun, and the rulers of Naphtali. 28 Your God has called for your strength. Show Your strength, O God, Who has acted for us. 29 Kings will bring gifts to You because of Your holy house at Jerusalem. 30 Speak sharp words to the wild animals in the tall grass by the river, the group of bulls with the calves of the people. Walk on those who desire pieces of silver. Divide the people who have joy in war. 31 Princes will come from Egypt. Ethiopia will hurry to put her hands out to God. 32 Sing to God, O nations of the earth. Sing praises to the Lord. 33 Sing to Him Who sits upon the heavens, the heavens of old. Listen, He sends out His voice, His powerful voice. 34 Tell of the power of God. His great power is over Israel and His strength is in the sky. 35 O God, You are honored with fear as You come from Your holy place. The God of Israel Himself gives strength and power to His people. Honor and thanks be to God!
english
Valorant is all set to receive a new map, Fracture, on September 8 2021, along with the Episode 3 Act 2 battlepass. Going by the pattern of content release in Valorant, the community and the fans mostly expected a new agent in Episode 3 Act 2. However, the developers brought a new map instead this time, diving more into Valorant’s mirror verse theory. There have been multiple teasers over the past couple of months hinting at the Valorant’s seventh map. The map was first teased in the The Year One Anthem video on the occasion of Valorant’s one-year completion. Now, after several hints, Fracture will arrive in the game, exposing the players to the Mirror Earth Kingdom. After Breeze in Episode 2 Act 3, Valorant will receive another map, Fracture in Episode 3 Act 2. Fracture follows an H-shape layout design with a visual split dividing the map into two halves. Like Split and Icebox, the map also features a zipline. There is a long zip line connecting underneath through the central collider. However, the defender spawn is in the middle of the map, with the attacker spawn on both sides. The A site of the map is completely abandoned and deserted, whereas the B site contrasts with a green land covered with grass, a science facility covered with leaves and greenery. Apart from the map features and other mechanical elements, one of the key things to look forward to in Fracture is the further revelation of the Valorant’s mirror verse lore theory. It is the first-ever map in Valorant with interactive lore elements. The Duality cinematic was the first time when players were introduced to the Mirror Agents and the Mirror Earth Kingdom. Since then, there has been no more information or expansion of the Valorant lore. Finally, Fracture will give some context to the mirror verse theory, with some interactive narrative objects present in the map. Fracture is certainly an interesting map, shaping up the Valorant’s lore perspective. Fans will get to experience interactive visual storytelling, questioning the events that took place on the map.
english
<filename>dataset_gen/tools/deb_unpacker.py<gh_stars>10-100 # Metadata and file extractor for DEB packages # Extracts files from a Debian package and saves metadata to JSON # meta data extracted includes: 1)software name, 2)version number, 3)architecture, 4)package file structure, 5)md5 file hashes 6)control info import os import re import argparse import os.path import json import logging import subprocess import threading import multiprocessing from helpers.call_cmd import call_cmd class Data(): def __init__(self): self.software = "unknown" self.version = "unknown" self.filestructure = "unknown" self.controlinfo = "unknown" self.package = "unknown" self.iso = "unknown" self.architecture = "unknown" class UnpackDebianFiles(): def main(self, filename, deb, arch, result): try: data = self.extract_info(filename, deb, arch, result) finally: self.threadLimiter.release() def extract_info(self, line, deb, arch, result): o = Data() o.iso = deb["iso"] o.architecture = arch o.package = line.strip() candidates = self.create_candidates(o.package) o.software = candidates[0].split("/")[-1] o.version = candidates[1] fstruct = self.extract_filesysinfo(o.package) hash_control = self.extract_files(o.package) # extraction probably failed if hash_control == []: result.append(o) return if hash_control[0] != "N/A": o.filestructure = self.extract_hash(fstruct, hash_control[0]) o.controlinfo = hash_control[1] result.append(o) return def create_candidates(self, filename): rpart = filename.rpartition('.') if rpart[0] == "": name_no_ext = rpart[2] else: name_no_ext = rpart[0] candidates = [] candidates = candidates + name_no_ext.rsplit('_', 1) return candidates def extract_hash(self, fstruct, hashlist): for line in hashlist.split('\n')[:-2]: split = line.split() for item in fstruct: if '/' + split[1] == item['path']: item['hash'] = split[0] return fstruct def dump_json(self, data_object): return json.dumps(data_object, default=lambda x: x.__dict__) def extract_filesysinfo(self, debpackagefilename): filesysinfo = debpackagefilename + "_filesysinfo.txt" try: cmd = "dpkg-deb -c " + debpackagefilename + " > " + filesysinfo call_cmd(cmd=cmd, shell=True, verbose=self.verbose) except subprocess.CalledProcessError as e: return try: f = open(filesysinfo, "r") fstruct = f.read() f.close() os.remove(filesysinfo) except: return filelist = fstruct.split("\n")[:-2] result = [] for x in filelist: data = {} results = x.split() data["size"] = results[2] data["date"] = results[3] data["path"] = " ".join(results[5:]) if data["path"].startswith("."): data["path"] = data["path"][1:] data['hash'] = '' result.append(data) return result def extract_files(self, debpackagefilename): ret = -1 try: cmd = ["dpkg-deb", "-R", debpackagefilename, debpackagefilename + "_extraction"] ret = call_cmd(cmd=cmd, shell=False, verbose=self.verbose) except subprocess.CalledProcessError as e: return [] if ret != 0: return [] md5_path = debpackagefilename + "_extraction/DEBIAN/md5sums" control_path = debpackagefilename + "_extraction/DEBIAN/control" hash_control = [] if os.path.exists(md5_path): f = open(md5_path, "r", errors='replace') hash_control.append(f.read()) f.close() else: hash_control.append("N/A") if os.path.exists(control_path): g = open(control_path, "r", errors='replace') hash_control.append(g.read()) g.close() return hash_control def init(self, thread_count, verbose, input_json, output_path): self.threadLimiter = threading.BoundedSemaphore(int(thread_count)) self.verbose = verbose self.input_json = input_json self.output_path = output_path def unpack_debian_files(self): try: with open(self.input_json) as f: json_data = json.load(f) except IOError: logging.error("Failed to open JSON file: " + input_json) exit(1) with multiprocessing.Manager() as manager: results = {} result = manager.list() threads = [] for arch in json_data: for deb in json_data[arch]: self.threadLimiter.acquire() deb_path = deb["deb_path"] t = threading.Thread(target=self.main, args=( deb_path, deb, arch, result)) t.start() threads.append(t) if arch not in results: results[arch] = [] for t in threads: t.join() count_of_debs = 0 for deb in result: count_of_debs += 1 # remove architecture from the Data object # so it is not repeated for no reason arch = deb.architecture del deb.__dict__['architecture'] results[arch].append(deb) logging.debug("Succesfully extracted " + str(count_of_debs) + " debians") with open(self.output_path, 'w', encoding="utf-8") as result_file: result_file.write(self.dump_json(results))
python
A beer can thrown at the bus would have been more than welcome compared to the three explosions that rocked Dortmund – explosions that damaged the bus on its way to Signal Iduna Park and injured defender Marc Bartra. The Spanish defender was the only player to suffer an injury in the incident and had to undergo surgery on a broken wrist and remove debris from his arm. Everyone remembers when Manchester United’s team bus was attacked on its way to West Ham’s Boleyn Ground in May last season. Videos from inside the bus were widely circulated and, while there was a hint of fear in the ranks as the bus’ windows held firm, there was also bemusement. Jesse Lingard clowning around in the chaos is an everlasting image. But this was no beer can. This was a pre-planned explosion meant to cause significant harm rather than make rival players feel unwelcome by greeting them with a hostile atmosphere. Following the initial confusion, it was the club that kept it together and tweeted relevant information to the fans. They assured fans at the stadium that they were in no immediate danger while asking them to remain calm and wait for an update on the game. Soon, the update everyone was waiting for was tweeted 15 minutes before actual kickoff time: the match had been cancelled and postponed to Wednesday with an earlier kickoff time. The tickets bought would remain valid. Travelling Monaco fans also chanted “Dortmund! Dortmund!” and home fans reciprocated by offering them beds for the night. Terrorism had been defeated as football stood defiantly in the face of misplaced religious extremism. Everyone moved on. Or had they? After what was arguably the most intense and traumatic experience most players and staff had experienced in their lives, they were simply asked to get back to their job – play football. The very next day. In a congested football calendar, there was simply no other date to play this fixture – at least that is what UEFA thought of the whole situation. The fixture couldn’t be played on a weekend because both teams obviously had league commitments. What about next week? Oh no, the second leg of the quarter-final was scheduled for then. Tickets had already been sold. Besides, both Dortmund and Monaco were scheduled to play in the semi-finals of the DFB-Pokal and Coupe de France the following week – and there was no chance they could delay it any further than that because the Champions League semi-finals had to be played soon after. So UEFA did the next best thing and scheduled the game for Wednesday night. Did they even consider Thursday night at the very least or would that have made the Europa League a sideshow (it already is, UEFA!)? Surely both the Bundesliga and Ligue 1 would have allowed the teams to play their league games a day later so they had adequate time to recuperate? “The decision to play the UEFA Champions League match between Borussia Dortmund and AS Monaco FC on Wednesday at 18.45 CET was made on Tuesday night at the BvB Stadion Dortmund in cooperation and complete agreement with clubs and authorities. UEFA washed their hands off the matter and put the onus on the clubs. Which begs the question: why weren’t the players consulted? Defender Sokratis Papastathopoulos summed it up best when he attacked UEFA for scheduling the game the very next day without consulting the players and staff. Anyone who watched the game would have immediately understood that the German side were simply not up for it. Less than 24 hours after the incident, there they were on the pitch with a depleted squad. It wasn’t just Dortmund who were emotionally broken after the explosions. Monaco were also not in the right frame of mind following the attack on the bus. He admitted that the players’ concentration was not at the desired levels. Without Bartra, Dortmund had to force Sven Bender into the centre-back role. This was Bender’s third game of the season – he had last played 45 minutes in December. He lasted only 45 minutes on Wednesday night too as a poor outing highlighted by an own goal saw him taken off at half-time. Monaco certainly looked the more focused team in the first half and the 2-0 lead was certainly well deserved. But Tuchel’s changes at half-time looked to have breathed some life into the game as the German side fought back. But Kylian Mbappe capitalised on a weak Lukasz Piszczek pass to nick the ball and score when he was one-on-one with the keeper. It almost killed the tie but Dortmund’s resilience did see them get one back and almost find an equaliser in stoppage time as the match ended 3-2 in favour of Monaco. While the world rejoiced at the fact that football had kicked terrorism in the butt, it was only after the game that the truth dawned on everyone. In the post-match interview by Jan Aage Fjortoft, Dortmund’s Nuri Sahin was very honest and forthright when he was asked to describe the last 24 hours and what went on behind the scenes. Sahin described how he had seen incidents like this unfold on TV – especially in his country Turkey where an Istanbul nightclub was attacked on New Years’ eve – and that he wouldn’t wish for this to happen to anyone. But what he said soon after was what struck a chord with everyone watching. That tells you the about the levels of concentration – or the lack of it – the Dortmund players had following the attack. As a player representing the club, they have a huge responsibility on their shoulders and one that is not easily shirked. As fans, we can easily take a decision on whether or not we go for a game. As a player, that is never an option. Dortmund claimed the game was a sellout but that was a number based on the tickets sold. A look at the crowd from various camera angles showed a number of empty seats. The quarter-final tie is still alive. But when football should have taken a backseat, the higher-ups’ decision saw convenience trump empathy. While standing up to terrorism in such a manner takes immense courage, a lack of compassion from those not directly involved leaves a nasty taste in the mouth.
english
<gh_stars>1-10 --- title: "Intro to React" author: Watson & Crick date: 2019-07-10 hero_image: ../static/bali-15.jpg --- this is a react intro post React es un framework de js, boludo fafafa lalala
markdown
import { Component, Output, EventEmitter } from '@angular/core'; import { Meal } from './meal.model'; @Component({ selector: 'new-meal', template: ` <h1>New<img src="resources/img/apple.png" alt="Caramel Apple">Meal</h1> <div> <label>Name:</label> <input #newName> </div> <div> <label>Meal Description:</label> <input #newDescription> </div> <div> <label>Calories:</label> <select #newCalorie> <option value=">500">>500</option> <option value="300-500" selected="selected">300-500</option> <option value="<300"><300</option> </select> <button (click)=" addClicked(newName.value, newDescription.value, newCalorie.value); newName.value=''; newDescription.value=''; newCalorie.value=''; ">Add</button> </div> ` }) export class NewMealComponent { @Output() newMealSender = new EventEmitter(); addClicked(name: string, description: string, calorie: number) { var newMealToAdd: Meal = new Meal(name, description, calorie); this.newMealSender.emit(newMealToAdd); } }
typescript
[["aerosur", "asuncion", "paraguay", "santa cruz", "bolivia", "0"], ["aerosur", "yacuiba", "bolivia", "tarija", "bolivia", "0"], ["aerosur", "cochabamba", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "cochabamba", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "cochabamba", "bolivia", "tarija", "bolivia", "0"], ["aerosur", "cochabamba", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "cobija", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "cobija", "bolivia", "riberalta", "bolivia", "0"], ["aerosur", "cobija", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "cuzco", "peru", "la paz", "bolivia", "0"], ["aerosur", "buenos aires", "argentina", "santa cruz", "bolivia", "0"], ["aerosur", "sao paulo", "brazil", "santa cruz", "bolivia", "0"], ["aerosur", "guayaramer\u00c3\u00adn", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "lima", "peru", "la paz", "bolivia", "0"], ["aerosur", "lima", "peru", "santa cruz", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "cochabamba", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "cobija", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "cuzco", "peru", "0"], ["aerosur", "la paz", "bolivia", "lima", "peru", "0"], ["aerosur", "la paz", "bolivia", "rerrenabaque", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "sucre", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "tarija", "bolivia", "0"], ["aerosur", "la paz", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "madrid", "spain", "santa cruz", "bolivia", "0"], ["aerosur", "miami", "united states", "santa cruz", "bolivia", "0"], ["aerosur", "puerto suarez", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "rerrenabaque", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "rerrenabaque", "bolivia", "san borja", "bolivia", "0"], ["aerosur", "riberalta", "bolivia", "cobija", "bolivia", "0"], ["aerosur", "riberalta", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "salta", "argentina", "santa cruz", "bolivia", "0"], ["aerosur", "sucre", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "sucre", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "san borja", "bolivia", "rerrenabaque", "bolivia", "0"], ["aerosur", "san borja", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "trinidad", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "tarija", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "cochabamba", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "cobija", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "guayaramer\u00c3\u00adn", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "riberalta", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "san borja", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "trinidad", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "tarija", "bolivia", "yacuiba", "bolivia", "0"], ["aerosur", "tarija", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "tarija", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "tarija", "bolivia", "santa cruz", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "asuncion", "paraguay", "0"], ["aerosur", "santa cruz", "bolivia", "cochabamba", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "buenos aires", "argentina", "0"], ["aerosur", "santa cruz", "bolivia", "sao paulo", "brazil", "0"], ["aerosur", "santa cruz", "bolivia", "lima", "peru", "0"], ["aerosur", "santa cruz", "bolivia", "la paz", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "madrid", "spain", "0"], ["aerosur", "santa cruz", "bolivia", "miami", "united states", "0"], ["aerosur", "santa cruz", "bolivia", "puerto suarez", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "salta", "argentina", "0"], ["aerosur", "santa cruz", "bolivia", "sucre", "bolivia", "0"], ["aerosur", "santa cruz", "bolivia", "tarija", "bolivia", "0"]]
json
{ "directions": [ "Whisk garlic, balsamic vinegar, grill seasoning, oregano, salt, and pepper together in a large bowl. Add steak and toss to evenly coat. Cover bowl with plastic wrap and marinate in the refrigerator, 8 hours to overnight.", "Heat olive oil in a skillet over medium-high heat until sizzling; add meat. Cook uncovered until liquids have drained from the steak, 3 to 5 minutes. Reduce heat to medium; cover.", "Cook steak for 15 minutes; uncover and flip. Add water to the skillet as needed to keep steak moist. Continue cooking until slightly firm, hot, and lightly pink in the center, about 15 minutes. An instant-read thermometer inserted into the center should read 140 degrees F (60 degrees C). Transfer meat to serving plate; reserve liquid in the skillet.", "Place onion in the skillet with reserved liquids; cook and stir until softened, about 5 minutes. Reduce heat to low; add mushrooms. Cook and stir until mushrooms are warmed through, about 5 minutes. Pour onion mixture over steak." ], "ingredients": [ "Marinade:", "3 cloves garlic, minced", "2 tablespoons balsamic vinegar, or to taste", "2 tablespoons grill seasoning", "2 tablespoons chopped fresh oregano", "salt and ground black pepper to taste", "1 pound skirt steak", "2 tablespoons olive oil, or as needed", "1/4 cup water, or as needed", "1 onion, chopped", "1 (8 ounce) can mushrooms, drained" ], "language": "en-US", "source": "allrecipes.com", "tags": [], "title": "Tender Juicy Skirt Steak (Churrasco)", "url": "http://allrecipes.com/recipe/254976/tender-juicy-skirt-steak-churrasco/" }
json
<reponame>hackape/ian-vscode-extension export * from './pathExists'; export * from './getMonorepoConfig'
typescript
{ "contacts": [ { "name": "<NAME>", "email": "", "designation": "", "mobile": [] } ], "guideStarURL": "http://www.guidestarindia.org/Summary.aspx?CCReg=5620", "name": "ORGANISATION FOR INTEGRATED DEVELOPMENT (OFID)", "primaryEmail": "<EMAIL>", "organisationType": [ "Direct Service" ], "telephone": [ "919437514369" ], "mainAddrress": { "state": "Odisha", "address": [ "At: <NAME>,", "P.O.: <NAME>,", "Via: Purushottampur,", "Ganjam", "Odisha", "761041" ] }, "briefDescription": "ORGANISATION FOR INTEGRATED DEVELOPMENT IS REGISTERED IN GANJAM, ORISSAa.\tTo promote integrated development of the people through systematic, time-bound and result-oriented programmes with specific aims and plans and to help people, irrespective of caste, creed, community, religion and sex, by promoting education, health & legal awareness and to improve the socio-economic conditions to make the people self-reliant.b.\tTo educate the people to be aware of their problems and solve those with the available resources through their organised and collective action and to enhance their self-esteem, knowledge and attitude to attain sustainable development.c.\tTo promote savings habit and cooperation among the people by facilitating formation of SHGs.d.\tTo establish and maintain health clinics, , hospitals, etc. to provide healthcare to all with special attention to women’s reproductive health, infant care and child survival. e.\tTo organise programmes to create awareness of child care, family health, family planning, sanitation, antenatal care and nutrition education etc. f.\tTo establish and maintain Homes for the old people, Homes for the destitute women, Orphanages and Drug De-addiction centres etc. g.\tTo provide intensive trainings in the selected vocations to needy persons, especially women, for their development.h.\tTo rehabilitate the destitute, helpless and disbanded women and their dependent children and orphans by providing vocational training.i.\tTo provide education facilities to all, especially the deprived sections of the society.j.\tTo undertake various programmes for rural development m.\tTo undertake rehabilitation of child labourers, bonded labourers, mentally retarded & differently abled children. n.\tTo undertake various development activities pertaining to women.o.\tTo undertakethe promotion of various developmental programmes to educate the youth, especially the women, on family welfare, population control measures etc.p.\tTo work for the popularization of various health schemes of Union and State Government.q.\tTo respect at all times the religious and cultural feelings of all the people while promoting the aims & objectives of the Society and to unfailingly avoid political activities.", "yearOfEstablishment": "2002" }
json
<reponame>sfbrigade/sf-admin-code<filename>sections/39.7.json {"text":"\n     This Chapter shall be enforceable by the City and any beneficially interested person. Any enforcement action shall be limited to injunctive relief, including specific performance. As set forth in Section \n39.8, there shall be no monetary damages for any violation of this Chapter.\n(Added by Ord. 227-12, File No. 120812, App. 11/7/2012, Eff. 12/7/2012) \n(Former Sec. 39.7 added by Ord. 401-96, App. 10/21/96; amended by Ord. 274-97, App. 7/3/97; Ord. 2-00, File No. 992000, App. 1/13/2000; repealed by Ord. 171-03, File No. 030422, App. 7/3/2003) \n\n","heading":{"title":"39","chaptersection":"7","identifier":"39.7","catch_text":"ENFORCEABILITY."}}
json
{ "type": "service_account", "project_id": "studied-handler-280013", "private_key_id": "<KEY>", "private_key": "-----<KEY>", "client_email": "<EMAIL>", "client_id": "113993375111009833442", "auth_uri": "https://accounts.google.com/o/oauth2/auth", "token_uri": "https://oauth2.googleapis.com/token", "auth_provider_x509_cert_url": "https://www.googleapis.com/oauth2/v1/certs", "client_x509_cert_url": "https://www.googleapis.com/robot/v1/metadata/x509/gladesheet%40studied-handler-280013.iam.gserviceaccount.com" }
json
Islamabad: PTV Home will telecast “Dirlis: Ertugrul” , a Turkish historical fiction and adventure television series, from the first of Ramazan. The drama series will be aired daily at 9:10pm on PTV Home. The series has been dubbed in Urdu on the directives of Prime Minister Imran Khan so that Pakistani viewers can understand it. The world famous Turkish historical television series was created by Mehmet Bozdag. Turkish actor Engin Altan Duzyatan starred as Ertugrul in the drama, which was filmed in Riva, a village Beykoz. Turkish state tv TRT 1 first broadcast the series in December 2014. The serial revolves around the life of Ertugrul, the father of Ottoman Empire Osman 1.
english
From election to election, the party that succeeded in controlling West Java changed. Based on historical data, no single party can emerge victorious in two consecutive elections in this province. During the 2009 election, the Democratic Party took over West Java (22.3 percent of the vote, 15 districts/cities). Then, in the 2014 election, West Java returned to being dominated by the PDI-P (19.6 percent of the vote, 16 districts/cities). In the 2019 election, West Java fell to the Gerindra Party (17.6 percent of the vote, 10 districts/cities). When winning back West Java in 2014, PDI-P controlled the northern to eastern parts of the province. However, their vote acquisition in 2014 decreased significantly compared to 1999. In the 2019 election, PDI-P only controlled six regencies/cities in the eastern region with a decreased vote acquisition of 14.3 percent.Participants of the 1999 election were joined by 48 political parties. Various preparations were made, including the preparation of ballot papers and ballot boxes. Only the Prosperous Justice Party (PKS) managed to dominate the areas previously controlled by nationalist parties. PKS's domination in Jabar is quite unique. In the 2004 election, PKS won three cities, namely Depok, Bekasi, and Bandung.Ideological Contestation Versus Political Pragmatism headtopics.comBaca lebih lajut: Arriving in Hangzhou, the National Badminton Team is confident of winning gold at the Asian GamesThe Indonesian badminton team has arrived in Hangzhou, China, and is preparing to compete in the 2022 Hangzhou Asian Games. Apart from holding training, the team coach also asked the athletes not to be too burdened and to maintain their physical condition. Mother's Prayer, Edgar Xavier Marvelo's Secret Move to Win 'Changquan' SilverMother's prayers were the secret behind Edgar Xavier Marvelo's success in winning wushu silver from the chanquan number in the 2022 Asian Games. Edgar was also listed as Indonesia's first silver donor at the 19th Asian Games. Now Quartararo can sleep soundlyFabio Quartararo can sleep soundly after winning the podium in the main MotoGP race in India. He last finished on the podium in a major race in the United States, 10 series ago. Maudy Effrosina, immediately won the roleMaudy Effrosina admitted that she was nervous about being offered to star in 'KKN di Desa Penari 2: Badarawuhi' because her predecessor film set a record for winning the largest number of viewers. Bezzecchi Liberation RaceMarco Bezzecchi can focus on racing again after deciding to stay at VR46 for MotoGP 2024, at the end of last August. Since then, he has continued to perform solidly until winning in India, his liberation race series. by political parties in the elections from 1999 to 2019 appears to be very dynamic. The winning party in West Java elections always changes. In the 1999 Legislative Election, West Java was dominated by the Indonesian Democratic Party of Struggle (PDI-P) with 32.2 percent of the votes. PDI-P won in 17 districts/cities. However, in the 2004 Election, the Golkar Party won West Java (27.9 percent of the votes, 18 districts/cities). PDI-P only received 17.5 percent of the votes. During the 2009 election, the Democratic Party took over West Java (22.3 percent of the vote, 15 districts/cities). Then, in the 2014 election, West Java returned to being dominated by the PDI-P (19.6 percent of the vote, 16 districts/cities). In the 2019 election, West Java fell to the Gerindra Party (17.6 percent of the vote, 10 districts/cities). The PDI-P has been quite successful in dominating West Java with two victories, namely in 1999 and 2014. In the 1999 elections, the strength of the PDI-P was widespread, conquering most of the western, northern, and eastern regions of West Java. However, when Golkar won West Java in the following election, the PDI-P's area of influence was only left with three districts: Subang, Majalengka, and Cirebon. When winning back West Java in 2014, PDI-P controlled the northern to eastern parts of the province. However, their vote acquisition in 2014 decreased significantly compared to 1999. In the 2019 election, PDI-P only controlled six regencies/cities in the eastern region with a decreased vote acquisition of 14.3 percent.Participants of the 1999 election were joined by 48 political parties. Various preparations were made, including the preparation of ballot papers and ballot boxes. Here we can see parties vying for sympathy and support. West Java has become a dynamic battleground for political parties. Despite the changing ownership of the region by parties, most of West Java is a nationalist base. In the two elections that followed, namely in 2009 and 2014, PKS did not win a single region in West Java. It was only in the 2019 elections that PKS managed to once again win several cities, namely Depok, Bekasi, Bogor, Bandung, and Cimahi.The fluctuating results of the elections in West Java are due to the heterogeneous, dynamic, and adaptive nature of the people in the region. An Anthropology lecturer at Padjadjaran University (Unpad), Rina Hermawati, stated that the Sundanese people are not homogeneous in terms of ethnicity, and the same goes for the Islamic practices followed by the West Javanese community. "There are different regions in West Java, including West Priangan, East Priangan, Sumedang, and others, each with their own unique characteristics. The Islamic religion embraced by the people also varies, with some following modern Islam while others adhere to traditional Islam. All of these factors greatly influence the political choices of the people in West Java," explained Rina. In addition to the heterogeneous characteristics of society, politics in West Java is also considered to be constantly changing due to factors such as disloyal voters. Firman Manan, a political science professor at Unpad, said that voters in West Java are not loyal. This is due to the low affiliation of voters with political parties, as well as the more individualistic nature of society.A farmer smiled as he crossed the village road from his farmland in Kampung Gandok, Suntenjaya Village, Lembang, West Bandung Regency, West Java on Tuesday (29/1/2019). This hard road facilitates farmers' access to farming land and marketing. Changes in political choices are also influenced by public figures."The people of West Java always look to the figure, namely a popular and religious figure," said Firman. In the digital technology era with the now-improving internet penetration, the public's political choices are also influenced by conversations on social media. Social media has become one of the references for the public to recognize and support the figures they will choose. For example, Ridwan Kamil's victory in occupying the number one seat in West Java cannot be separated from the branding of Ridwan Kamil's image on social media. Going forward, Firman stated that the political dynamics in West Java will be determined by the power of three factors, namely social media, figures, and popularity.Although politics in West Java is constantly changing, according to Firman, the political landscape in West Java can essentially be classified into six clusters. Firstly, there is the megapolitan cluster, which covers the area around Jakarta, including the cities and regencies of Bekasi, Bogor, and Depok. This cluster is characterized by urban voters and encompasses the upper-middle class. The domination of PKS is quite strong here. Secondly, the Bandung Raya cluster, comprising of the City and Regency of Bandung, West Bandung Regency, Cimahi City, and Sumedang. This cluster also represents an urban and middle to upper-class voting area. Here, the supporters of PKS are dominant, targeting educated urban communities from the beginning. Thirdly, the Karawangan cluster covers the upper part of West Java, namely Karawang, Subang, and Purwakarta. This cluster is characterized by dominant nationalist voters. Fourthly, the Cirebonan or Ciayumajakuning cluster, which includes the City and Regency of Cirebon, Indramayu, Majalengka, and Kuningan, also tends to be nationalist in general.A motorcyclist passes through Sukabumi Park, Menteng, Central Jakarta, on Friday (28/07/2023). The park, located in the middle of a residential area, is currently undergoing revitalization. Fifth, there is the West Priangan cluster that includes the cities and districts of Sukabumi and Cianjur. This area is known for its rural voters. Sixth, there is the East Priangan cluster that includes Garut, the cities and districts of Tasikmalaya, Banjar, and Ciamis. This area is known for its Muslim traditional and conservative communities with pockets of Islamic boarding schools. With this mapping, it is not surprising to see pockets of party loyalists, where party power remains unshakable for five elections. For example, the PDI-P has its loyal territory in Majalengka and Cirebon regencies. Golkar loyal territory is in Purwakarta regency. The city or regency of Tasikmalaya has been a stronghold for the loyalists of the United Development Party (PPP) for four consecutive elections. However, in the 2019 elections, PPP was unable to win a single district/city in West Java. Tasikmalaya was handed over to the Gerindra Party. The Democratic Party was unable to establish a strong presence in West Java. As the winner of the 2009 elections in West Java, the Democratic Party mainly dominated the western part of the region and had limited influence in the eastern part. However, in subsequent elections, the Democratic Party was unable to win any district or city in West Java. Regarding Gerindra Party, despite winning West Java in the 2019 Election, their domination is not particularly significant. The control of regions by the four major parties tends to be evenly distributed. Gerindra is only slightly ahead. There are many factors that actually cause the society's preferences to change. The movement and direction of development also influence politics when the implemented programs meet the interests of pragmatic society. "The pragmatic society also makes political dynamics in West Java more dynamic, especially since 30 or more national strategic projects are located in West Java," said Firman.Aerial photo of the National Strategic Project (PSN) that has been utilized at the Cibitung Interchange in West Cikarang, Bekasi Regency, West Java, on Sunday (10/9/2023). Economic improvement and welfare continue to be a major issue in West Java. This is because 3.9 million West Java residents (7.6 percent) still live in poverty. With the dynamic territorial control where the people tend to be disloyal, West Java remains a fiercely contested area in the 2024 elections. Gerindra Party faces a test, whether its power will be replaced according to the existing pattern. Or, can Gerindra break the pattern and become the first party in West Java to win two consecutive elections?
english
<filename>app/src/main/java/com/camect/android/example/fragments/MethodListFragment.java package com.camect.android.example.fragments; import android.content.Context; import android.content.DialogInterface; import android.os.Bundle; import android.text.Editable; import android.text.TextUtils; import android.text.TextWatcher; import android.view.LayoutInflater; import android.view.View; import android.view.ViewGroup; import android.widget.Button; import android.widget.EditText; import android.widget.TextView; import androidx.annotation.NonNull; import androidx.annotation.Nullable; import androidx.appcompat.app.AlertDialog; import androidx.fragment.app.Fragment; import androidx.lifecycle.ViewModelProvider; import androidx.recyclerview.widget.DividerItemDecoration; import androidx.recyclerview.widget.LinearLayoutManager; import androidx.recyclerview.widget.RecyclerView; import com.camect.android.example.R; import com.camect.android.example.util.AsyncTask; import com.camect.android.example.viewmodels.CamectViewModel; import com.camect.android.library.CamectSDK; import com.camect.android.library.model.HomeInfo; import java.util.ArrayList; import java.util.concurrent.ThreadPoolExecutor; public class MethodListFragment extends Fragment implements OnItemClickListener { public static MethodListFragment newInstance() { return new MethodListFragment(); } private final ArrayList<Method<?>> mMethods = new ArrayList<>(); private Button mButton; private CamectViewModel mCamectViewModel; private ThreadPoolExecutor mExecutor; private void buildMethodList() { mMethods.clear(); mMethods.add(new Method<String>("Get 24 Hr Access Token") { @Override protected String doInBackground(Void... voids) { String accessToken = CamectSDK.getInstance().getAccessToken(24 * 3600); if (TextUtils.isEmpty(accessToken)) { accessToken = "FAILED"; } return accessToken; } }); mMethods.add(new Method<HomeInfo>("Get Home Info") { @Override protected HomeInfo doInBackground(Void... voids) { if (mCamectViewModel.getHomeInfo() == null) { HomeInfo homeInfo = CamectSDK.getInstance().getHomeInfo(); mCamectViewModel.setHomeInfo(homeInfo); } return mCamectViewModel.getHomeInfo(); } }); mMethods.add(new Method<Void>("Set Home Name") { @Override protected Void doInBackground(Void... voids) { return null; } @Override protected void onPostExecute(Void result) { showNamePrompt(); // reset this task so it can run again reset(); } }); mMethods.add(new Method<HomeInfo>("Set Mode to HOME") { @Override protected HomeInfo doInBackground(Void... voids) { if (CamectSDK.getInstance().setMode(CamectSDK.Mode.HOME)) { mCamectViewModel.getHomeInfo().setMode(CamectSDK.Mode.HOME.getValue()); } return mCamectViewModel.getHomeInfo(); } }); mMethods.add(new Method<HomeInfo>("Set Mode to AWAY") { @Override protected HomeInfo doInBackground(Void... voids) { if (CamectSDK.getInstance().setMode(CamectSDK.Mode.AWAY)) { mCamectViewModel.getHomeInfo().setMode(CamectSDK.Mode.AWAY.getValue()); } return mCamectViewModel.getHomeInfo(); } }); mMethods.add(new Method<Void>("Set Alerts For Home") { @Override protected Void doInBackground(Void... voids) { return null; } @Override protected void onPostExecute(Void result) { ObjectAlertChooserDialogFragment.newInstance(null) .show(getChildFragmentManager(), null); reset(); } }); mMethods.add(new Method<Void>("List Cameras") { @Override protected Void doInBackground(Void... voids) { if (mCamectViewModel.getAllCameras() == null || mCamectViewModel.getAllCameras().size() == 0) { mCamectViewModel.setCameras(CamectSDK.getInstance().getCameras()); } return null; } @Override protected void onPostExecute(Void result) { getFragmentManager().beginTransaction() .replace(R.id.container, CameraListPagerFragment.newInstance()) .addToBackStack("cameras") .commit(); // reset this task so it can run again reset(); } }); } @Override public View onCreateView(LayoutInflater inflater, ViewGroup container, Bundle savedInstanceState) { return inflater.inflate(R.layout.fragment_list, container, false); } @Override public void onDestroyView() { mExecutor.shutdownNow(); super.onDestroyView(); } @Override public void onItemClick(View view, int position, long id) { mMethods.get(position).executeNow(); } @Override public void onViewCreated(@NonNull View view, @Nullable Bundle savedInstanceState) { mCamectViewModel = new ViewModelProvider(requireActivity()) .get(CamectViewModel.class); mExecutor = AsyncTask.newSingleThreadExecutor(); buildMethodList(); DividerItemDecoration decoration = new DividerItemDecoration(requireContext(), DividerItemDecoration.VERTICAL); RecyclerView recyclerView = view.findViewById(R.id.list); recyclerView.addItemDecoration(decoration); recyclerView.setLayoutManager(new LinearLayoutManager(requireContext())); recyclerView.setAdapter(new MethodListAdapter(requireContext(), this)); } private void showNamePrompt() { View view = LayoutInflater.from(requireContext()).inflate(R.layout.dialog_name_input, null, false); final EditText editText = view.findViewById(R.id.text_input_box); editText.setText(mCamectViewModel.getHomeInfo().getName()); editText.addTextChangedListener(new TextWatcher() { @Override public void afterTextChanged(Editable s) { mButton.setEnabled(!TextUtils.isEmpty(s.toString().trim())); } @Override public void beforeTextChanged(CharSequence s, int start, int count, int after) { // not used; here to complete the interface } @Override public void onTextChanged(CharSequence s, int start, int before, int count) { // not used; here to complete the interface } }); AlertDialog.Builder builder = new AlertDialog.Builder(requireContext()); builder.setCancelable(false) .setTitle("Set Home Name") .setView(view) .setNegativeButton(android.R.string.cancel, null) .setPositiveButton(android.R.string.ok, (dialog, which) -> new AsyncTask<Void, Void, HomeInfo>(mExecutor) { @Override protected HomeInfo doInBackground(Void... voids) { String name = editText.getText().toString().trim(); if (CamectSDK.getInstance().setHomeName(name)) { mCamectViewModel.getHomeInfo().setName(name); } return mCamectViewModel.getHomeInfo(); } @Override protected void onPostExecute(HomeInfo homeInfo) { ModelInspectorDialogFragment.newInstance("Set Home Info", homeInfo.toString()) .show(getChildFragmentManager(), null); } }.executeNow()); AlertDialog alert = builder.create(); alert.setOnShowListener(dialog -> { String name = editText.getText().toString().trim(); mButton = alert.getButton(DialogInterface.BUTTON_POSITIVE); mButton.setEnabled(!TextUtils.isEmpty(name)); }); alert.show(); } private static class MethodViewHolder extends RecyclerView.ViewHolder { public final TextView mName; public MethodViewHolder(View view, final OnItemClickListener listener) { super(view); mName = view.findViewById(R.id.method_name); if (listener != null) { view.setOnClickListener(v -> listener.onItemClick(view, getAdapterPosition(), getItemId())); } } } public abstract class Method<T> extends AsyncTask<Void, Void, T> { private final String mName; private Method(String name) { super(mExecutor); mName = name; } public String getName() { return mName; } @Override protected void onPostExecute(T result) { ModelInspectorDialogFragment.newInstance(mName, result.toString()) .show(getChildFragmentManager(), null); // reset this task so it can run again reset(); } } private class MethodListAdapter extends RecyclerView.Adapter<MethodViewHolder> { private final LayoutInflater mInflater; private final OnItemClickListener mListener; public MethodListAdapter(Context context, OnItemClickListener listener) { mInflater = LayoutInflater.from(context); mListener = listener; } @Override public int getItemCount() { return mMethods.size(); } @Override public void onBindViewHolder(@NonNull MethodViewHolder holder, int position) { Method<?> method = mMethods.get(position); holder.mName.setText(method.getName()); } @NonNull @Override public MethodViewHolder onCreateViewHolder(@NonNull ViewGroup parent, int viewType) { View view = mInflater.inflate(R.layout.list_item_method, parent, false); return new MethodViewHolder(view, mListener); } } }
java
<reponame>SohrabAmin/Free-Room {"id": "FCS292H1S20201", "code": "FCS292H1S", "name": "Special Topics in French Cultural Studies I", "description": "The relation of French popular culture to society. For more information, see http://www.french.utoronto.ca/undergraduate/courses/french_cultural_studies", "division": "Faculty of Arts and Science", "department": "French", "prerequisites": "None", "exclusions": "", "level": 200, "campus": "UTSG", "term": "2020 Winter", "breadths": [1], "meeting_sections": [{"code": "L9901", "instructors": ["<NAME>"], "times": [], "size": 110, "enrolment": 110}]}
json
{ "id": 5009, "cites": 1, "cited_by": 12, "reference": [ "Feenberg, <NAME>. and <NAME>. Income Inequality and the Incomes of Very High Income Tax Payers: Evidence from Tax Returns, in <NAME>, ed., Tax Policy and the Economy 2. Cambridge: MIT Press, 1993, pp. 145-77." ] }
json
In Nagaland, a total of 124 nomination papers were filed on Monday. Sitting Chief Minister Neiphiu Rio, Deputy Chief Minister Y Patton, and state BJP president Temjen Imna Along were among the 124 candidates who filed their nominations. In Meghalaya, 176 party heavyweights, independents, and first-timers filed their nominations on Monday. Meghalaya TMC chief Charles Pyngrope, Opposition leader Mukul Sangma, TMC MLA Dikkanchi D Shira, United Democratic Party (UDP) chief Metbah Lyngdoh, UDP general secretary Jemino Mawthoh, People’s Democratic Party (PDF) leader Banteidor Lyngdoh, Meghalaya BJP chief Ernest Mawrie and BJP spokesperson Mariahom Kharkrang were among those who filed their nominations. Download The Economic Times News App to get Daily Market Updates & Live Business News. Subscribe to The Economic Times Prime and read the ET ePaper online.
english
Apple is said to be considering a move that will bring Mac mini production to the U.S., through manufacturing partner Foxconn, according to supply chain sources speaking to Digitimes. Foxconn already has an estimated 15 “operating bases” in the U.S. according to Digitimes. Indeed two of those at least include factories in California and Texas that finish assembly of partially assembled products, and while Foxconn officially denied plans earlier in the year to expand to Detroit, it did note that it has multiple U.S.-based facilities already in place. Part of the production effort will involve Foxconn’s push to outfit some of those facilities with more automated workers, something else we heard the manufacturer was planning for a future ramp-up back in November. More automated production lines would help Apple get around the limitations it has cited in the past for failing to do more production at home in the U.S.: costs, and getting production facilities up to its exacting standards. The Mac mini is a good candidate for Apple getting its feet wet once again with U.S. production for a number of reasons. First, like the Mac Pro which was first suggested as the likely target for Apple’s $100 million investment in U.S. production, it doesn’t have a screen. Eliminating display components from the equation represents a significant cost savings in terms of shipping components, and it likely has other benefits, too. Displays on most Mac and mobile device models are now integrated tightly with other components including the glass and other internals, so having production facilities near to display partners just makes sense in case things go wrong or need adjustment once limited trial or full production has already begun. The Mac mini is also a Mac with relatively low shipping volume: Digitimes predicts 1.4 million units total for 2012. While Apple doesn’t break out individual Mac sales figures, that would make for a relatively small chunk of the 18.1 million Macs it sold during fiscal 2012. It’s small enough to be manageable for what is essentially a trial run, while also being large enough to represent a serious undertaking, where producing the niche and aging Mac Pro would’ve been a symbolic gesture, at best. This is Digitimes, which has a spotty track record, so be wary of its veracity as usual, but remember also that the publication has proven in the past to have significant upstream supply chain access, too, and Apple CEO Tim Cook is on the record saying the production of one Mac line in particular will move stateside in 2013.
english
<filename>SOLVER/src/preloop/utilities/PreloopFFTW.cpp // PreloopFFTW.cpp // created by Kuangdai on 21-Sep-2016 // perform FFT using fftw #include "PreloopFFTW.h" int PreloopFFTW::sNmax = 0; std::vector<fftw_plan> PreloopFFTW::sR2CPlans; std::vector<fftw_plan> PreloopFFTW::sC2RPlans; std::vector<RDColX> PreloopFFTW::sR2C_RMats; std::vector<CDColX> PreloopFFTW::sR2C_CMats; std::vector<RDColX> PreloopFFTW::sC2R_RMats; std::vector<CDColX> PreloopFFTW::sC2R_CMats; void PreloopFFTW::checkAndInit(int nr) { if (nr <= sNmax) { return; } int xx = 1; for (int NR = sNmax + 1; NR <= nr; NR++) { int NC = NR / 2 + 1; int n[] = {NR}; sR2C_RMats.push_back(RDColX(NR, 1)); sR2C_CMats.push_back(CDColX(NC, 1)); sC2R_RMats.push_back(RDColX(NR, 1)); sC2R_CMats.push_back(CDColX(NC, 1)); double *r2c_r = &(sR2C_RMats[NR - 1](0, 0)); ComplexD *r2c_c = &(sR2C_CMats[NR - 1](0, 0)); sR2CPlans.push_back(fftw_plan_many_dft_r2c( 1, n, xx, r2c_r, n, 1, NR, reinterpret_cast<fftw_complex*>(r2c_c), n, 1, NC, FFTW_ESTIMATE)); double *c2r_r = &(sC2R_RMats[NR - 1](0, 0)); ComplexD *c2r_c = &(sC2R_CMats[NR - 1](0, 0)); sC2RPlans.push_back(fftw_plan_many_dft_c2r( 1, n, xx, reinterpret_cast<fftw_complex*>(c2r_c), n, 1, NC, c2r_r, n, 1, NR, FFTW_ESTIMATE)); } sNmax = nr; } void PreloopFFTW::finalize() { for (int i = 0; i < sNmax; i++) { fftw_destroy_plan(sR2CPlans[i]); fftw_destroy_plan(sC2RPlans[i]); } sNmax = 0; } void PreloopFFTW::computeR2C(int nr) { checkAndInit(nr); fftw_execute(sR2CPlans[nr - 1]); double inv_nr = one / (double)nr; sR2C_CMats[nr - 1] *= inv_nr; } void PreloopFFTW::computeC2R(int nr) { checkAndInit(nr); fftw_execute(sC2RPlans[nr - 1]); } bool PreloopFFTW::isLuckyNumber(int n, bool forceOdd) { int num = n; // We always hope to use even numbers that are generally faster, // but the Nyquist frequency sometimes causes trouble. // force odd if (forceOdd && num % 2 == 0) { return false; } // use even when n > 10 if (!forceOdd && num % 2 != 0 && num > 10) { return false; } for (int i = 2; i <= num; i++) { while(num % i == 0) { num /= i; if (i > 13) { return false; } } } num = n; int e = 0; while(num % 11 == 0) { num /= 11; e++; } num = n; int f = 0; while(num % 13 == 0) { num /= 13; f++; } if (e + f > 1) { return false; } return true; } int PreloopFFTW::nextLuckyNumber(int n, bool forceOdd) { while(true) { if (isLuckyNumber(n, forceOdd)) { return n; } n++; } }
cpp
{"Target":"css/academic.min.1bcfb251c53da9811552e33f4fbd28a5.css","MediaType":"text/css","Data":{"Integrity":"md5-G8+yUcU9qYEVUuM/T70opQ=="}}
json
<filename>tools/webpack/getResolve.js import { isDevelopment } from 'config/env'; export default function getResolve() { return { extensions: ['.js', '.json'], ...(!isDevelopment ? { alias: { react: 'preact-compat', 'react-dom': 'preact-compat', }, } : {}), }; }
javascript
The three men arrested in connection with the blast near the Bharatiya Janata Party office in Malleswaram on April 17 were produced before the magistrate here Tuesday and taken into police custody for further investigation. The suspects, Peer Mohideen (39), Basheer Sunnati (30) and Kichan Buhari (38), all Tirunelveli natives, were picked up from Tamil Nadu. Mohideen used to frequent Bangalore in connection with his tea leaves business, police said. Sunnati has a real estate business in Chennai. City Police Commissioner Raghavendra H. Auradkar confirmed the arrest but refused to say more. But senior police officials said Mohideen and Sunnati are suspected to have executed the blast while Buhari was their handler, providing logistical support. “We suspect that others also were involved in the blasts and efforts are on to nab them soon,” a police source said. The police are ascertaining their motive and whether they are linked to a terrorist group. The city police, who verified call details obtained through mobile phone towers around the blast site, analysed the data and zeroed in on the suspects.
english
<reponame>UnderMybrella/blaseball-feed-schemas<gh_stars>0 {"id":"8fe665b6-c862-4a28-9143-84aef14b7825","playerTags":[],"teamTags":["eb67ae5e-c4bf-46ca-bbbc-425cd34182ff"],"gameTags":[],"metadata":{"title":"Ground Broken","votes":806544,"renoId":"8"},"created":"2021-03-11T16:37:47.575Z","season":12,"tournament":-1,"type":56,"day":72,"phase":5,"category":1,"description":"The Moist Talkers break ground on Gleek Arena, selecting to build the Boreal prefab.\nAnother flag is planted!","nuts":0}
json
Is a 12-inch MacBook on the horizon – and if so, uh, why? Do we really still need a 12-inch MacBook? Rumors have been floating around in recent weeks that Apple might be reintroducing a 12-inch MacBook into its product lineup, but where there's hope for some who may revel in this news, not everyone is convinced about its accuracy. As reported by iMore, the initial murmurs started to appear just after the M2 chip was announced at WWDC 2022 by Bloomberg's Mark Gurman. Gurman predicted that the smaller form-factor MacBook would launch either at the end of 2023 or early 2024, though no information was provided about the nature of the laptop itself. As such, we have no idea if this speculated release would be a MacBook Air, a more powerful MacBook Pro, or something new entirely. In fact, another prominent analyst has come forward to dash any expectations regarding the prediction. Ross Young of Display Supply Chain Consultants claimed in a tweet made available only to his super followers that "We are skeptical on a 12-inch MacBook at this point. Apple's strategy for notebooks is currently 13-inches and larger. Companies in the MacBook Pro display supply chain we talked to are not aware of it". This goes without saying, but while both of these analysts are respected in their fields and have provided accurate information in the past, take all of this with a pinch of salt. The prediction game can be wildly inaccurate and Apple loves to keep a tight lid on its product launches so either party has the potential to be incorrect in this instance. Despite the lack of real evidence, this wouldn't be an unusual move from the tech giant as it has previously released several 12-inch laptops over the years, as both high-end and entry-level devices, though we have to question its place among the current lineup of Apple hardware. Analysis: Is the 12-inch MacBook even relevant these days? The 12-inch MacBook was fantastic when it was released back in 2015. Petite laptops were nowhere near as commonplace back then, which made its offering of a small, lightweight and stylish laptop more unusual and exciting. Apple would even have an easier time making the design work these days now that it has its own Silicon, allowing devices such as the MacBook Air 2020 to run a fanless design thanks to the first M1 chip, so it's not inconceivable to put that in a 12-inch MacBook too. Thing is, if a new 12-inch MacBook is targeting everyday users, why not simply buy the 2022 M2 MacBook Air? It's not much bigger at 13.6-inches and has every bit of the slimline charm of its smaller predecessor. It also feels unlikely that a 12-inch MacBook Pro would see much use as a display that small would be restrictive to the folks who typically need a Pro device - graphic designers, audio engineers, and so on. Heck, even the iPad is equipped with Apple Silicon these days. It's a bit more restrictive running iPadOS rather than macOS, but if you really value form factor over capability then this is an ideal solution when paired with the Magic Keyboard. Apple isn't a stranger to making hardware with niche appeal though. The current version of the Mac Pro desktop computer was so wildly expensive at launch that it was never going to permeate the mainstream market for everyday consumers, and the Studio Display appears to have been created solely with creatives in mind. I won't speculate, but I will express a hopeful outcome - It's time for a 2-in-1 MacBook. I'm a fan of what the iPad and iPad Pro offer to artists and content creators, especially when paired with the Apple Pencil, so a macOS-powered hybrid device would genuinely have the potential to replace several of my devices, provided it can connect to more than a single display. A girl can dream, but I refuse to speculate. We will simply have to keep watching for further evidence to support the existence (or lack thereof) of a new 12-inch MacBook in the coming months. Get the hottest deals available in your inbox plus news, reviews, opinion, analysis, deals and more from the TechRadar team. Jess is a former TechRadar Computing writer, where she covered all aspects of Mac and PC hardware, including PC gaming and peripherals. She has been interviewed as an industry expert for the BBC, and while her educational background was in prosthetics and model-making, her true love is in tech and she has built numerous desktop computers over the last 10 years for gaming and content creation. Jess is now a journalist at The Verge.
english
Gaining extra weight and not having a perfect jawline or chubby cheeks can make people feel insecure about their appearance. Some skip their meals and start hitting the gym, while others make abrupt changes to their diet. But there is also a procedure where people remove excess fat from their face to tone their cheeks; buccal fat removal surgery. Read this article to learn about buccal fat surgery, who can get it, and the complications post-surgery. Buccal fat refers to the fat which is present between your cheekbones and jawbones. You might feel like your face is excessively round or full if you have larger buccal fat pads and can also make you look older than your age or overweight. Buccal fat surgery is also known as cheek reduction or buccal lipectomy. It involves removing fatty tissue from your face and changing its shape. It aids in improving cheekbone definition, reduces the width of round faces, and gives the face a chiselled appearance. The surgery is a non-invasive procedure which is generally completed within 30 minutes to one hour. It takes two to three weeks for the healing process post-surgery. Its costs range from ₹ 25,000 to ₹ 40,000 in India and can also exceed in some prominent cities. - You can opt for a buccal fat surgery if your physical health is good and you have a healthy weight. - People with narrow faces are not recommended to try this procedure as it can cause sunken cheeks as you age. - You have a small rounded fat mass in the cheek, known as pseudoherniation, which is caused due to weak buccal fat pad. - People who do not smoke. - People who are old should avoid this surgery as your cheeks naturally lose fat as you age. Buccal fat surgery is generally safe but can cause some complications too. Some of them are as follows: You can opt for cheek reduction if you have a fuller face and want to remove the excess fats. The surgery can change the contouring of the shape but not the size of the cheeks, as it depends on your jaw size. You should also note that buccal fat diminishes with age, and removing it can result in a skinny look with age. Moreover, it is not a substitute for weight loss as it cannot change the shape of your whole face. Patients should consume liquid or semi-liquid food after the surgery. Visit your doctor immediately if you have shortness of breath, chest pain, severe pain, signs of infection, or an abnormal heartbeat.
english
<gh_stars>10-100 { "directions": [ "Bring a large pot of lightly salted water to a boil. Add pasta and cook for 8 to 10 minutes or until al dente. Drain, then return to pot.", "Heat oil in a large heavy skillet over medium heat. Saute garlic for 5 minutes. Toss in prawns, and cook for 5 minutes on each side. Remove prawns, and set aside. Remove garlic slices, and discard.", "Pour oil over pasta in pot, and toss to evenly coat. Sprinkle 3/4 of the Parmesan cheese onto pasta, and stir until evenly distributed. Transfer to serving dish. Arrange prawns on top, then sprinkle with remaining Parmesan and parsley." ], "ingredients": [ "8 ounces fusilli (spiral) pasta", "6 tablespoons olive oil", "2 cloves garlic, sliced", "8 tiger prawns, peeled and deveined", "1/2 cup grated Parmesan cheese, divided", "1 teaspoon chopped fresh parsley" ], "language": "en-US", "source": "allrecipes.com", "tags": [], "title": "Garlic Pasta with Prawns", "url": "http://allrecipes.com/recipe/46360/garlic-pasta-with-prawns/" }
json
Tokyo, Japan, May 31 – Japan killed 122 pregnant minke whales during a highly controversial annual whaling expedition that Tokyo defends as scientific research but conservationists call “gruesome and unnecessary”. The four-month expedition in the Antarctic ended in March after the fleet killed 333 minke whales, according to a report submitted by Japanese authorities to the International Whaling Commission (IWC) last month. Of those, 122 were pregnant, according to the Japanese report, with dozens more immature whales among those killed. Humane Society International, a conservationist group, called the figures “a shocking statistic and sad indictment on the cruelty of Japan’s whale hunt”. “It is further demonstration, if needed, of the truly gruesome and unnecessary nature of whaling operations, especially when non-lethal surveys have been shown to be sufficient for scientific needs,” said the group’s senior program manager, Alexia Wellbelove. Japan is a signatory to the International Whaling Commission, which has maintained a moratorium on hunting whales since 1986. But Tokyo exploits a loophole allowing whales to be killed for “scientific research” and claims it is trying to prove the population is large enough to sustain a return to commercial hunting. It makes no secret of the fact that meat from the expeditions ends up on dinner tables. Japan’s Fisheries Agency defended the hunt, saying it was not targeting pregnant whales. “We catch whales totally at random,” said Yuki Morita, an official in charge of whaling at the agency. “The IWC scientific committee recognises the number of whales we hunt is at the level that is necessary for research, but not above the level that would hurt the conservation of the stock,” he added. “We’d like to stress here that the high ratio of pregnant females is noteworthy… This shows there are many mature females, suggesting we can expect growth in resources stock,” he told AFP. Japan has hunted whales for centuries, and their meat was a key source of protein in the immediate post-World War II years when the country was desperately poor. But consumption has declined significantly in recent decades, with much of the population saying they rarely or never eat whale meat. The submission to the IWC contains details of the scientific research that Tokyo says justifies the continued expeditions, including measurements on the weight and blubber thickness and stomach contents of the killed whales, and data on sighting locations. In 2014, the International Court of Justice ordered Tokyo to end the Antarctic hunt, saying it found permits issued by Japan were “not for purposes of scientific research”. Tokyo cancelled the hunt the following year, but resumed it in 2016, also killing around 300 minke whales. Earlier this year, the fisheries ministry confirmed it was studying a possible upgrade to its ageing lead whaling ship, and Prime Minister Shinzo Abe vowed to continue the hunts. “We will pursue all possibilities in order to resume commercial whaling, including opportunities at the September meeting of the IWC,” he told parliament, when asked to comment on the nation’s policy.
english
:root { --code: #222; --code-fg: #ddd; --background: #333; --foreground: #ccc; --footer-text: #888; --border: #444; --table1: #444; --table2: #4a4a4a; --table3: #555; --table4: #5a5a5a; --input: #444; --input-border: #555; --link: #0088ce; --link-hover: #00559e; --link-visited: #ad6ed1; --list-link: #d73a3a; --diff-bg-red: #433; --diff-bg-green: #343; --diff-bg-chg: #333; } body { background: var(--background); color: var(--foreground); border: 1px solid var(--border); } input { background: var(--input); color: var(--foreground); border: 1px solid var(--input-border); padding: 3px; } a, a[href], a.header, div.page_header a:visited { color: var(--link); } a:hover, a:active, a[href]:hover, a[href]:active, div.page_header a:hover { color: var(--link-hover); } a:visited, a[href]:visited, div.page_nav a:visited { color: var(--link-visited); } a.list { color: var(--foreground); } a.list:hover, .list a:hover { color: var(--list-link); } div.page_header, div.page_footer { background-color: var(--table1); } div.page_footer_text { color: var(--footer-text); } div.title, a.title, span.title { color: var(--foreground); background-color: var(--table1); } a.title:hover, span.title:hover { background-color: var(--table3); } div.index_include, div.page_path { border-color: var(--border); } div.page_body { background-color: var(--code); } div.diff.header { background-color: var(--table1); border-color: var(--table3); } div.diff.extended_header { background-color: var(--code); } div.diff.chunk_header span.chunk_info { background-color: inherit; } div.diff.chunk_header a, div.diff.chunk_header { color: #f0e; } td.pre, div.pre, div.diff, div.chunk_block.ctx div div.diff.ctx { color: var(--code-fg); } div.chunk_block.rem div.old div.diff.rem { background-color: var(--diff-bg-red); } div.chunk_block.add div.new div.diff.add { background-color: var(--diff-bg-green); } div.chunk_block.chg div div.diff { background-color: var(--diff-bg-chg); } div.diff.add span.marked { background-color: #282; color: #8e8; } span.refs span.head { border-color: #383; background-color: #383; border-radius: 2px; } span.refs span.remote { border-color: #f0e; background-color: #f0e; border-radius: 2px; } tr.dark, table.blame .dark:hover { background-color: var(--table1); } tr.light, table.blame .light:hover { background-color: var(--table2); } tr.light:hover, tr.dark:hover { background-color: var(--table3); } /*****************************/ /* Code Highlighting */ /*****************************/ .kwa { color:#cda869; } .kwb { color:#cda869; } .kwc { color:#cda869; } .kwd { color:#cda869; } .str { color:#8f9d6a; } .num { color:#cf6a4c; font-weight:700 } /*! Highlight: Twilight (CodeMirror Reference)*/ /* .CodeMirror { background:#141414; color:#ccc } .highlight { background:#141414; color:#ccc } .listingblock pre.prettyprint { background:#141414; color:#ccc } .listingblock pre:not(.highlight) { background:#141414; color:#ccc } .highlight .pun { color:#ccc; } .highlight .pln { color:#ccc; } .highlight .hll { background-color:#ffc; } .highlight .com { color:#5f5a60 ;font-style:italic } .highlight .cm { color:#5f5a60 ;font-style:italic } .highlight .c { color:#5f5a60 ;font-style:italic } .highlight .err { border:#b22518 } .highlight .k { color:#cda869; } .highlight .key { color:#cda869; } .highlight .kwd { color:#cda869; } .highlight .cp { color:#5f5a60; } .highlight .c1 { color:#5f5a60; font-style:italic } .highlight .cs { color:#5f5a60; font-style:italic } .highlight .gd { background:#420e09 } .highlight .ge { font-style:italic } .highlight .gr { background:#b22518 } .highlight .gh { color:navy; font-weight:700} .highlight .gi { background: #253b22} .highlight .gp { font-weight:700 } .highlight .gs { font-weight:700 } .highlight .gu { color:purple;font-weight:700} .highlight .kd { color:#e9df8f} .highlight .kp { color:#9b703f} .highlight .na { color:#f9ee98} .highlight .nb { color:#cda869} .highlight .nc { color:#9b859d; font-weight:700} .highlight .no { color:#9b859d} .highlight .nd { color:#7587a6} .highlight .ni { color:#cf6a4c; font-weight:700 } .highlight .nf { color:#9b703f; font-weight:700 } .highlight .nn { color:#9b859d; font-weight:700 } .highlight .nt { color:#cda869; font-weight:700 } .highlight .nv { color:#7587a6} .highlight .ow { color:#aa22ff; font-weight:700 } .highlight .w { color:#141414; } .highlight .mf { color:#cf6a4c; } .highlight .mh { color:#cf6a4c; } .highlight .mi { color:#cf6a4c; } .highlight .mo { color:#cf6a4c; } .highlight .str { color:#8f9d6a; } .highlight .sb { color:#8f9d6a; } .highlight .sc { color:#8f9d6a; } .highlight .sd { color:#8f9d6a; font-style:italic } .highlight .s2 { color:#8f9d6a; } .highlight .se { color:#f9ee98; font-weight:700 } .highlight .sh { color:#8f9d6a; } .highlight .si { color:#daefa3; font-weight:700 } .highlight .sx { color:#8f9d6a; } .highlight .sr { color:#e9c062; } .highlight .s1 { color:#8f9d6a; } .highlight .ss { color:#cf6a4c; } .highlight .bp { color:#0aa; } .highlight .vc { color:#7587a6; } .highlight .vg { color:#7587a6; } .highlight .vi { color:#7587a6; } .highlight .il { color:#099; } *
css
<gh_stars>0 from .boost import VCBooster
python
<reponame>Ryebread4/Rustionary {"word":"contestation","definition":"1. The act of contesting; emulation; rivalry; strife; dispute. \"Loverlike contestation.\" Milton. After years spent in domestic, unsociable contestations, she found means to withdraw. Clarendon. 2. Proof by witness; attestation; testimony. [Obs.] A solemn contestation ratified on the part of God. Barrow."}
json
<reponame>zondahuman/serialize-svr package com.abin.lee.serialize.avro; import org.apache.avro.io.*; import org.apache.avro.specific.SpecificDatumReader; import org.apache.avro.specific.SpecificDatumWriter; import org.apache.commons.collections.iterators.ObjectArrayIterator; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; /** * Created by abin on 2018/4/15 18:05. * serialize-svr * com.abin.lee.serialize.avro * https://unmi.cc/apache-avro-serializing-deserializing/#more-7488 * https://blog.csdn.net/hua245942641/article/details/50724360 * 貌似还有些问题 */ public class AvroSerialize { public static byte[] serialize(Class<?> clazz, Object obj) throws IOException { DatumWriter datumWriter = new SpecificDatumWriter(clazz); ByteArrayOutputStream outputStream = new ByteArrayOutputStream(); BinaryEncoder binaryEncoder = EncoderFactory.get().directBinaryEncoder(outputStream, null); datumWriter.write(obj, binaryEncoder); return outputStream.toByteArray(); } public static <T> T deserialize(Class<T> clazz, byte[] bytes) throws IOException { DatumReader datumReader = new SpecificDatumReader(clazz.getClass()); BinaryDecoder binaryEncoder = DecoderFactory.get().directBinaryDecoder(new ByteArrayInputStream(bytes), null); return (T)datumReader.read(clazz.getClass(), binaryEncoder); } }
java
<reponame>DeeSnow97/node-that { "name": "that-thing", "version": "1.0.0", "description": "That thing you were looking for", "main": "index.js", "scripts": { "test": "standard" }, "author": "b3nsn0w", "license": "WTFPL", "devDependencies": { "standard": "^10.0.2" }, "dependencies": { "standard": "^10.0.2" }, "repository": { "type": "git", "url": "git+https://github.com/DeeSnow97/node-that.git" }, "keywords": [ "that", "thing", "this" ], "bugs": { "url": "https://github.com/DeeSnow97/node-that/issues" }, "homepage": "https://github.com/DeeSnow97/node-that#readme" }
json
describe("traversing", function() { "use strict"; var link; beforeEach(function() { jasmine.sandbox.set("<div><b></b><b></b><i></i><a id='test'><strong></strong><em></em></a><b></b><i></i><i></i></div>"); link = DOM.find("#test"); }); function _forIn(obj, callback, thisPtr) { for (var prop in obj) { callback.call(thisPtr, obj[prop], prop, obj); } } describe("next, prev, closest", function() { it("should return an appropriate element", function() { var expectedResults = { next: "b", prev: "i" }; _forIn(expectedResults, function(tagName, methodName) { expect(link[methodName]()).toHaveTag(tagName); }); }); it("should search for the first matching element if selector exists", function() { expect(link.next("i")).toHaveTag("i"); expect(link.prev("b")).toHaveTag("b"); }); }); describe("closest", function() { it("searches for the first matching element if selector exists", function() { expect(link.closest("body")).toHaveTag("body"); }); // it("returns direct parent when no selector specified", function() { // expect(DOM.find("html").closest()).toBe(DOM); // expect(DOM.closest()[0]).toBeUndefined(); // }); }); it("should return empty element if value is not found", function() { var unknownEl = link.find("unknown"); expect(unknownEl.next()[0]).toBeUndefined(); expect(unknownEl.prev()[0]).toBeUndefined(); expect(unknownEl.child(0)[0]).toBeUndefined(); }); it("should throw error if arguments are invalid", function() { expect(function() { link.child({}) }).toThrow(); expect(function() { link.child(function() {}) }).toThrow(); expect(function() { link.next({}) }).toThrow(); expect(function() { link.prev(function() {}) }).toThrow(); }); describe("children, nextAll, prevAll", function() { it("should return an appropriate collection of elements", function() { var expectedResults = { children: "strong em".split(" "), nextAll: "b i i".split(" "), prevAll: "i b b".split(" ") }, isOK = function(methodName) { return function(el, index) { expect(el).toHaveTag(expectedResults[methodName][index]); }; }; _forIn(expectedResults, function(tagName, methodName) { for (var arr = link[methodName](), i = 0, n = arr.length; i < n; ++i) { isOK(arr[i]); } }); }); it("should filter matching elements by optional selector", function() { var filters = { children: "em", nextAll: "i", prevAll: "i" }, haveTag = function(tagName) { return function(el) { expect(el).toHaveTag(tagName); }; }; _forIn(filters, function(tagName, methodName) { for (var arr = link[methodName](tagName), i = 0, n = arr.length; i < n; ++i) { haveTag(tagName); } }); }); it("should return empty element if value is not found", function() { var unknownEl = link.find("unknown"); expect(unknownEl.nextAll().length).toBe(0); expect(unknownEl.prevAll().length).toBe(0); expect(unknownEl.children().length).toBe(0); }); it("should throw error if arguments are invalid", function() { expect(function() { link.children({}) }).toThrow(); expect(function() { link.children(function() {}) }).toThrow(); expect(function() { link.nextAll({}) }).toThrow(); expect(function() { link.prevAll(function() {}) }).toThrow(); }); }); });
javascript
{"artist_id":"ARREGHI1187FB47C48","artist_latitude":null,"artist_location":"","artist_longitude":null,"artist_name":"<NAME>","duration":315.34975,"num_songs":1,"song_id":"SOKEZWF12AB0185C39","title":"Heroina Madness","year":2004}
json
"""Test the mulled BioContainers image name generation.""" import pytest from nf_core.modules import MulledImageNameGenerator @pytest.mark.parametrize( "specs, expected", [ (["foo==0.1.2", "bar==1.1"], [("foo", "0.1.2"), ("bar", "1.1")]), (["foo=0.1.2", "bar=1.1"], [("foo", "0.1.2"), ("bar", "1.1")]), ], ) def test_target_parsing(specs, expected): """Test that valid specifications are correctly parsed into tool, version pairs.""" assert MulledImageNameGenerator.parse_targets(specs) == expected @pytest.mark.parametrize( "specs", [ ["foo<0.1.2", "bar==1.1"], ["foo=0.1.2", "bar>1.1"], ], ) def test_wrong_specification(specs): """Test that unexpected version constraints fail.""" with pytest.raises(ValueError, match="expected format"): MulledImageNameGenerator.parse_targets(specs) @pytest.mark.parametrize( "specs", [ ["foo==0a.1.2", "bar==1.1"], ["foo==0.1.2", "bar==1.b1b"], ], ) def test_noncompliant_version(specs): """Test that version string that do not comply with PEP440 fail.""" with pytest.raises(ValueError, match="PEP440"): MulledImageNameGenerator.parse_targets(specs) @pytest.mark.parametrize( "specs, expected", [ ( [("chromap", "0.2.1"), ("samtools", "1.15")], "mulled-v2-1f09f39f20b1c4ee36581dc81cc323c70e661633:bd74d08a359024829a7aec1638a28607bbcd8a58-0", ), ( [("pysam", "0.16.0.1"), ("biopython", "1.78")], "mulled-v2-3a59640f3fe1ed11819984087d31d68600200c3f:185a25ca79923df85b58f42deb48f5ac4481e91f-0", ), ( [("samclip", "0.4.0"), ("samtools", "1.15")], "mulled-v2-d057255d4027721f3ab57f6a599a2ae81cb3cbe3:13051b049b6ae536d76031ba94a0b8e78e364815-0", ), ], ) def test_generate_image_name(specs, expected): """Test that a known image name is generated from given targets.""" assert MulledImageNameGenerator.generate_image_name(specs) == expected
python
Starting the third day of the second Test at 182/3, England would have hoped to put up a big total on the board before letting Pakistan bat again. But Wahab Riaz and Yasir Shah had other plans for the visitors, as they picked up six quick wickets together to get England all out for just 242 after adding only 60 runs in the morning. Joe Root and Jonny Bairstow who continued batting on day 3, could not continue their partnership as Root fell after just 8 runs to his total. Bairstow, on the other hand, put up a stiff challenge against the Pakistani bowlers, but even he couldn’t help England as wickets fell at the other end. Yasir and Wahab were the reason for England’s downfall as they chipped at the English batting line-up with ease. Each of them picked up four wickets and destroyed all the chances of England trying to get back into the game. Except for Stuart Broad who scored 15 runs, the middle and lower order crumbled with a combined score of a paltry 9 runs. Root was the first to fall at 206, after which wickets fell at regular intervals with four wickets falling in the span of 11 runs. Wahab with his pace and Yasir with his craft all but just bowled England out of the match. Pakistan started their second innings with a lead of 136 runs. They lost two early wickets in the form of Shan Masood and Shoaib Malik with just 16 on the board. Despite the early breakthrough, they lifted by the spirited performances by Mohammad Hafeez who scored a half century before getting out. Captain Misbah-ul-Haq and Younis Khan continued their form from the first innings with a 130 run partnership to give extra boost to Pakistan’s lead. Misbah is not out on 87 heading towards a second century in as many innings and Younis is also well set with a score of 71. At the end of days play Pakistan were well placed with the score at 222/3. Pakistan will start day 4 with 358 runs lead and will look to win this before going into the third and final Test. â? ? Brief scores: England: 242 (Joe Root 88, Jonny Bairstow 46; Wahab Riaz 66/4, Yasir Shah 93/4) Pakistan: 222/3 (Misbah 87, Younis 71; Mark Wood 22/2)
english
<filename>fundamentals-of-linux-pathway.json { "title": "Fundamentals of Linux", "description": "Data school course aimed at beginners to learn about the Linux operating system", "courses": [ { "course_id": "fundamentals-of-linux-explore-the-file-system", "title": "Fundamentals of Linux: Explore the file system", "description": "How to use cd, pwd, and ls to explore the file system on a linux server" }, { "course_id": "fundamentals-of-linux-switch-between-users", "title": "Fundamentals of Linux: Switch between users", "description": "How to change user on a linux server" }, { "course_id": "fundamentals-of-linux-environment-variables", "title": "Fundamentals of Linux: Environment variables", "description": "What are environment variables and what is bash profile" }, { "course_id": "fundamentals-of-linux-filesystem-operations", "title": "Fundamentals of Linux: Filesystem operations", "description": "Tutorial to learn navigating the Linux filesystem and the basic operations on files" }, { "course_id": "fundamentals-of-linux-process-management", "title": "Fundamentals of Linux: Process management", "description": "Learn how to manage processes in Linux shell." }, { "course_id": "fundamentals-of-linux-text-manipulation", "title": "Fundamentals of Linux: Unix Scripts For Text Manipulation", "description": "The goal of this course is to present the most common Unix\ntext manipulation commands " }, { "course_id": "fundamentals-of-linux-stream-processing", "title": "Fundamentals of linux: Stream processing and manipulation.", "description": "The goal of this course is to present some common tools for stream processing and manipulation in the Unix environment " }, { "course_id": "fundamentals-of-linux-remote-access", "title": "Fundamentals of Linux: Connect to a remote server", "description": "Use SSH to connect to a remote host" }, { "course_id": "fundamentals-of-linux-package-manager", "title": "Fundamentals of Linux: Package manager", "description": "Tutorial to familiarize with Linux software packages and the tools to manage them." } ] }
json
To illustrate the point further, Jesus told them this story: “A man had two sons. The younger son told his father, ‘I want my share of your estate now before you die.’ So his father agreed to divide his wealth between his sons. “A few days later this younger son packed all his belongings and moved to a distant land, and there he wasted all his money in wild living. About the time his money ran out, a great famine swept over the land, and he began to starve. He persuaded a local farmer to hire him, and the man sent him into his fields to feed the pigs. The young man became so hungry that even the pods he was feeding the pigs looked good to him. But no one gave him anything.
english
//! structural search replace use crate::source_change::SourceFileEdit; use ra_ide_db::RootDatabase; use ra_syntax::ast::make::expr_from_text; use ra_syntax::ast::{AstToken, Comment}; use ra_syntax::{AstNode, SyntaxElement, SyntaxNode}; use ra_text_edit::{TextEdit, TextEditBuilder}; use rustc_hash::FxHashMap; use std::collections::HashMap; use std::str::FromStr; pub use ra_db::{SourceDatabase, SourceDatabaseExt}; use ra_ide_db::symbol_index::SymbolsDatabase; #[derive(Debug, PartialEq)] pub struct SsrError(String); impl std::fmt::Display for SsrError { fn fmt(&self, f: &mut core::fmt::Formatter<'_>) -> core::fmt::Result { write!(f, "Parse error: {}", self.0) } } impl std::error::Error for SsrError {} pub fn parse_search_replace( query: &str, db: &RootDatabase, ) -> Result<Vec<SourceFileEdit>, SsrError> { let mut edits = vec![]; let query: SsrQuery = query.parse()?; for &root in db.local_roots().iter() { let sr = db.source_root(root); for file_id in sr.walk() { dbg!(db.file_relative_path(file_id)); let matches = find(&query.pattern, db.parse(file_id).tree().syntax()); if !matches.matches.is_empty() { edits.push(SourceFileEdit { file_id, edit: replace(&matches, &query.template) }); } } } Ok(edits) } #[derive(Debug)] struct SsrQuery { pattern: SsrPattern, template: SsrTemplate, } #[derive(Debug)] struct SsrPattern { pattern: SyntaxNode, vars: Vec<Var>, } /// represents an `$var` in an SSR query #[derive(Debug, Clone, PartialEq, Eq, Hash)] struct Var(String); #[derive(Debug)] struct SsrTemplate { template: SyntaxNode, placeholders: FxHashMap<SyntaxNode, Var>, } type Binding = HashMap<Var, SyntaxNode>; #[derive(Debug)] struct Match { place: SyntaxNode, binding: Binding, ignored_comments: Vec<Comment>, } #[derive(Debug)] struct SsrMatches { matches: Vec<Match>, } impl FromStr for SsrQuery { type Err = SsrError; fn from_str(query: &str) -> Result<SsrQuery, SsrError> { let mut it = query.split("==>>"); let pattern = it.next().expect("at least empty string").trim(); let mut template = it .next() .ok_or_else(|| SsrError("Cannot find delemiter `==>>`".into()))? .trim() .to_string(); if it.next().is_some() { return Err(SsrError("More than one delimiter found".into())); } let mut vars = vec![]; let mut it = pattern.split('$'); let mut pattern = it.next().expect("something").to_string(); for part in it.map(split_by_var) { let (var, var_type, remainder) = part?; is_expr(var_type)?; let new_var = create_name(var, &mut vars)?; pattern.push_str(new_var); pattern.push_str(remainder); template = replace_in_template(template, var, new_var); } let template = expr_from_text(&template).syntax().clone(); let mut placeholders = FxHashMap::default(); traverse(&template, &mut |n| { if let Some(v) = vars.iter().find(|v| v.0.as_str() == n.text()) { placeholders.insert(n.clone(), v.clone()); false } else { true } }); let pattern = SsrPattern { pattern: expr_from_text(&pattern).syntax().clone(), vars }; let template = SsrTemplate { template, placeholders }; Ok(SsrQuery { pattern, template }) } } fn traverse(node: &SyntaxNode, go: &mut impl FnMut(&SyntaxNode) -> bool) { if !go(node) { return; } for ref child in node.children() { traverse(child, go); } } fn split_by_var(s: &str) -> Result<(&str, &str, &str), SsrError> { let end_of_name = s.find(':').ok_or_else(|| SsrError("Use $<name>:expr".into()))?; let name = &s[0..end_of_name]; is_name(name)?; let type_begin = end_of_name + 1; let type_length = s[type_begin..].find(|c| !char::is_ascii_alphanumeric(&c)).unwrap_or_else(|| s.len()); let type_name = &s[type_begin..type_begin + type_length]; Ok((name, type_name, &s[type_begin + type_length..])) } fn is_name(s: &str) -> Result<(), SsrError> { if s.chars().all(|c| c.is_ascii_alphanumeric() || c == '_') { Ok(()) } else { Err(SsrError("Name can contain only alphanumerics and _".into())) } } fn is_expr(s: &str) -> Result<(), SsrError> { if s == "expr" { Ok(()) } else { Err(SsrError("Only $<name>:expr is supported".into())) } } fn replace_in_template(template: String, var: &str, new_var: &str) -> String { let name = format!("${}", var); template.replace(&name, new_var) } fn create_name<'a>(name: &str, vars: &'a mut Vec<Var>) -> Result<&'a str, SsrError> { let sanitized_name = format!("__search_pattern_{}", name); if vars.iter().any(|a| a.0 == sanitized_name) { return Err(SsrError(format!("Name `{}` repeats more than once", name))); } vars.push(Var(sanitized_name)); Ok(&vars.last().unwrap().0) } fn find(pattern: &SsrPattern, code: &SyntaxNode) -> SsrMatches { fn check( pattern: &SyntaxElement, code: &SyntaxElement, placeholders: &[Var], mut match_: Match, ) -> Option<Match> { match (pattern, code) { (SyntaxElement::Token(ref pattern), SyntaxElement::Token(ref code)) => { if pattern.text() == code.text() { Some(match_) } else { None } } (SyntaxElement::Node(ref pattern), SyntaxElement::Node(ref code)) => { if placeholders.iter().any(|n| n.0.as_str() == pattern.text()) { match_.binding.insert(Var(pattern.text().to_string()), code.clone()); Some(match_) } else { let mut pattern_children = pattern .children_with_tokens() .filter(|element| !element.kind().is_trivia()); let mut code_children = code.children_with_tokens().filter(|element| !element.kind().is_trivia()); let new_ignored_comments = code.children_with_tokens().filter_map(|element| { element.as_token().and_then(|token| Comment::cast(token.clone())) }); match_.ignored_comments.extend(new_ignored_comments); let match_from_children = pattern_children .by_ref() .zip(code_children.by_ref()) .fold(Some(match_), |accum, (a, b)| { accum.and_then(|match_| check(&a, &b, placeholders, match_)) }); match_from_children.and_then(|match_| { if pattern_children.count() == 0 && code_children.count() == 0 { Some(match_) } else { None } }) } } _ => None, } } let kind = pattern.pattern.kind(); let matches = code .descendants() .filter(|n| n.kind() == kind) .filter_map(|code| { let match_ = Match { place: code.clone(), binding: HashMap::new(), ignored_comments: vec![] }; check( &SyntaxElement::from(pattern.pattern.clone()), &SyntaxElement::from(code), &pattern.vars, match_, ) }) .collect(); SsrMatches { matches } } fn replace(matches: &SsrMatches, template: &SsrTemplate) -> TextEdit { let mut builder = TextEditBuilder::default(); for match_ in &matches.matches { builder.replace( match_.place.text_range(), render_replace(&match_.binding, &match_.ignored_comments, template), ); } builder.finish() } fn render_replace( binding: &Binding, ignored_comments: &Vec<Comment>, template: &SsrTemplate, ) -> String { let mut builder = TextEditBuilder::default(); for element in template.template.descendants() { if let Some(var) = template.placeholders.get(&element) { builder.replace(element.text_range(), binding[var].to_string()) } } for comment in ignored_comments { builder.insert(template.template.text_range().end(), comment.syntax().to_string()) } builder.finish().apply(&template.template.text().to_string()) } #[cfg(test)] mod tests { use super::*; use ra_syntax::SourceFile; fn parse_error_text(query: &str) -> String { format!("{}", query.parse::<SsrQuery>().unwrap_err()) } #[test] fn parser_happy_case() { let result: SsrQuery = "foo($a:expr, $b:expr) ==>> bar($b, $a)".parse().unwrap(); assert_eq!(&result.pattern.pattern.text(), "foo(__search_pattern_a, __search_pattern_b)"); assert_eq!(result.pattern.vars.len(), 2); assert_eq!(result.pattern.vars[0].0, "__search_pattern_a"); assert_eq!(result.pattern.vars[1].0, "__search_pattern_b"); assert_eq!(&result.template.template.text(), "bar(__search_pattern_b, __search_pattern_a)"); dbg!(result.template.placeholders); } #[test] fn parser_empty_query() { assert_eq!(parse_error_text(""), "Parse error: Cannot find delemiter `==>>`"); } #[test] fn parser_no_delimiter() { assert_eq!(parse_error_text("foo()"), "Parse error: Cannot find delemiter `==>>`"); } #[test] fn parser_two_delimiters() { assert_eq!( parse_error_text("foo() ==>> a ==>> b "), "Parse error: More than one delimiter found" ); } #[test] fn parser_no_pattern_type() { assert_eq!(parse_error_text("foo($a) ==>>"), "Parse error: Use $<name>:expr"); } #[test] fn parser_invalid_name() { assert_eq!( parse_error_text("foo($a+:expr) ==>>"), "Parse error: Name can contain only alphanumerics and _" ); } #[test] fn parser_invalid_type() { assert_eq!( parse_error_text("foo($a:ident) ==>>"), "Parse error: Only $<name>:expr is supported" ); } #[test] fn parser_repeated_name() { assert_eq!( parse_error_text("foo($a:expr, $a:expr) ==>>"), "Parse error: Name `a` repeats more than once" ); } #[test] fn parse_match_replace() { let query: SsrQuery = "foo($x:expr) ==>> bar($x)".parse().unwrap(); let input = "fn main() { foo(1+2); }"; let code = SourceFile::parse(input).tree(); let matches = find(&query.pattern, code.syntax()); assert_eq!(matches.matches.len(), 1); assert_eq!(matches.matches[0].place.text(), "foo(1+2)"); assert_eq!(matches.matches[0].binding.len(), 1); assert_eq!( matches.matches[0].binding[&Var("__search_pattern_x".to_string())].text(), "1+2" ); let edit = replace(&matches, &query.template); assert_eq!(edit.apply(input), "fn main() { bar(1+2); }"); } fn assert_ssr_transform(query: &str, input: &str, result: &str) { let query: SsrQuery = query.parse().unwrap(); let code = SourceFile::parse(input).tree(); let matches = find(&query.pattern, code.syntax()); let edit = replace(&matches, &query.template); assert_eq!(edit.apply(input), result); } #[test] fn ssr_function_to_method() { assert_ssr_transform( "my_function($a:expr, $b:expr) ==>> ($a).my_method($b)", "loop { my_function( other_func(x, y), z + w) }", "loop { (other_func(x, y)).my_method(z + w) }", ) } #[test] fn ssr_nested_function() { assert_ssr_transform( "foo($a:expr, $b:expr, $c:expr) ==>> bar($c, baz($a, $b))", "fn main { foo (x + value.method(b), x+y-z, true && false) }", "fn main { bar(true && false, baz(x + value.method(b), x+y-z)) }", ) } #[test] fn ssr_expected_spacing() { assert_ssr_transform( "foo($x:expr) + bar() ==>> bar($x)", "fn main() { foo(5) + bar() }", "fn main() { bar(5) }", ); } #[test] fn ssr_with_extra_space() { assert_ssr_transform( "foo($x:expr ) + bar() ==>> bar($x)", "fn main() { foo( 5 ) +bar( ) }", "fn main() { bar(5) }", ); } #[test] fn ssr_keeps_nested_comment() { assert_ssr_transform( "foo($x:expr) ==>> bar($x)", "fn main() { foo(other(5 /* using 5 */)) }", "fn main() { bar(other(5 /* using 5 */)) }", ) } #[test] fn ssr_keeps_comment() { assert_ssr_transform( "foo($x:expr) ==>> bar($x)", "fn main() { foo(5 /* using 5 */) }", "fn main() { bar(5)/* using 5 */ }", ) } }
rust
<filename>root/pli/ms/vinaya/pli-tv-pvr/pli-tv-pvr1.16_root-pli-ms.json { "pli-tv-pvr1.16:0.1": "Parivāra", "pli-tv-pvr1.16:0.2": "Bhikkhuvibhaṅga", "pli-tv-pvr1.16:0.3": "Dutiyabhāga", "pli-tv-pvr1.16:0.4": "8. Samuccayavāra", "pli-tv-pvr1.16:1.1": "Methunaṁ dhammaṁ paṭisevanapaccayā kati āpattiyo āpajjati?", "pli-tv-pvr1.16:1.2": "Methunaṁ dhammaṁ paṭisevanapaccayā catasso āpattiyo āpajjati.", "pli-tv-pvr1.16:1.3": "Akkhāyite sarīre methunaṁ dhammaṁ paṭisevati, āpatti pārājikassa;", "pli-tv-pvr1.16:1.4": "yebhuyyena khāyite sarīre methunaṁ dhammaṁ paṭisevati, āpatti thullaccayassa;", "pli-tv-pvr1.16:1.5": "vaṭṭakate mukhe acchupantaṁ aṅgajātaṁ paveseti, āpatti dukkaṭassa;", "pli-tv-pvr1.16:1.6": "jatumaṭṭhake pācittiyaṁ—", "pli-tv-pvr1.16:1.7": "methunaṁ dhammaṁ paṭisevanapaccayā imā catasso āpattiyo āpajjati.", "pli-tv-pvr1.16:1.8": "Tā āpattiyo catunnaṁ vipattīnaṁ kati vipattiyo bhajanti?", "pli-tv-pvr1.16:1.9": "Sattannaṁ āpattikkhandhānaṁ katihi āpattikkhandhehi saṅgahitā?", "pli-tv-pvr1.16:1.10": "Channaṁ āpattisamuṭṭhānānaṁ katihi samuṭṭhānehi samuṭṭhanti?", "pli-tv-pvr1.16:1.11": "Catunnaṁ adhikaraṇānaṁ katamaṁ adhikaraṇaṁ?", "pli-tv-pvr1.16:1.12": "Sattannaṁ samathānaṁ katihi samathehi sammanti?", "pli-tv-pvr1.16:1.13": "Tā āpattiyo catunnaṁ vipattīnaṁ dve vipattiyo bhajanti—", "pli-tv-pvr1.16:1.14": "siyā sīlavipattiṁ siyā ācāravipattiṁ.", "pli-tv-pvr1.16:1.15": "Sattannaṁ āpattikkhandhānaṁ catūhi āpattikkhandhehi saṅgahitā—", "pli-tv-pvr1.16:1.16": "siyā pārājikāpattikkhandhena, siyā thullaccayāpattikkhandhena, siyā pācittiyāpattikkhandhena, siyā dukkaṭāpattikkhandhena.", "pli-tv-pvr1.16:1.17": "Channaṁ āpattisamuṭṭhānānaṁ ekena samuṭṭhānena samuṭṭhanti—", "pli-tv-pvr1.16:1.18": "kāyato ca cittato ca samuṭṭhanti, na vācato.", "pli-tv-pvr1.16:1.19": "Catunnaṁ adhikaraṇānaṁ, āpattādhikaraṇaṁ.", "pli-tv-pvr1.16:1.20": "Sattannaṁ samathānaṁ, tīhi samathehi sammanti—", "pli-tv-pvr1.16:1.21": "siyā sammukhāvinayena ca, paṭiññātakaraṇena ca, siyā sammukhāvinayena ca tiṇavatthārakena ca …pe…. #1", "pli-tv-pvr1.16:2.1": "Anādariyaṁ paṭicca udake uccāraṁ vā passāvaṁ vā kheḷaṁ vā karaṇapaccayā kati āpattiyo āpajjati?", "pli-tv-pvr1.16:2.2": "Anādariyaṁ paṭicca udake uccāraṁ vā passāvaṁ vā kheḷaṁ vā karaṇapaccayā ekaṁ āpattiṁ āpajjati.", "pli-tv-pvr1.16:2.3": "Dukkaṭaṁ—", "pli-tv-pvr1.16:2.4": "anādariyaṁ paṭicca udake uccāraṁ vā passāvaṁ vā kheḷaṁ vā karaṇapaccayā imaṁ ekaṁ āpattiṁ āpajjati.", "pli-tv-pvr1.16:2.5": "Sā āpatti catunnaṁ vipattīnaṁ kati vipattiyo bhajati?", "pli-tv-pvr1.16:2.6": "Sattannaṁ āpattikkhandhānaṁ katihi āpattikkhandhehi saṅgahitā?", "pli-tv-pvr1.16:2.7": "Channaṁ āpattisamuṭṭhānānaṁ katihi samuṭṭhānehi samuṭṭhāti?", "pli-tv-pvr1.16:2.8": "Catunnaṁ adhikaraṇānaṁ katamaṁ adhikaraṇaṁ?", "pli-tv-pvr1.16:2.9": "Sattannaṁ samathānaṁ katihi samathehi sammati?", "pli-tv-pvr1.16:2.10": "Sā āpatti catunnaṁ vipattīnaṁ ekaṁ vipattiṁ bhajati—", "pli-tv-pvr1.16:2.11": "ācāravipattiṁ.", "pli-tv-pvr1.16:2.12": "Sattannaṁ āpattikkhandhānaṁ ekena āpattikkhandhena saṅgahitā—", "pli-tv-pvr1.16:2.13": "dukkaṭāpattikkhandhena.", "pli-tv-pvr1.16:2.14": "Channaṁ āpattisamuṭṭhānānaṁ ekena samuṭṭhānena samuṭṭhāti—", "pli-tv-pvr1.16:2.15": "kāyato ca cittato ca samuṭṭhāti, na vācato.", "pli-tv-pvr1.16:2.16": "Catunnaṁ adhikaraṇānaṁ, āpattādhikaraṇaṁ.", "pli-tv-pvr1.16:2.17": "Sattannaṁ samathānaṁ tīhi samathehi sammati—", "pli-tv-pvr1.16:2.18": "siyā sammukhāvinayena ca paṭiññātakaraṇena ca, siyā sammukhāvinayena ca tiṇavatthārakena cāti. #2", "pli-tv-pvr1.16:3.1": "Samuccayavāro niṭṭhito aṭṭhamo.", "pli-tv-pvr1.16:3.2": "Aṭṭhapaccayavārā niṭṭhitā.", "pli-tv-pvr1.16:3.3": "Mahāvibhaṅge soḷasamahāvārā niṭṭhitā.", "pli-tv-pvr1.16:4.1": "Bhikkhuvibhaṅgamahāvāro niṭṭhito." }
json
<reponame>danyprasetya8/line-clone<gh_stars>0 import { isValidEmail, isEmptyString } from './validation' const keyExist = k => k !== undefined const isEmpty = v => keyExist(v) && isEmptyString(v) const validateForm = forms => { const errors = [] if (isEmpty(forms.username)) { errors.push('Email / Id must be filled') } if (isEmpty(forms.id)) { errors.push('Id must be filled') } if (isEmpty(forms.name)) { errors.push('Name must be filled') } if (isEmpty(forms.email)) { errors.push('Email must be filled') } else { if (keyExist(forms.email) && !isValidEmail(forms.email)) { errors.push('Not a valid email') } } if (isEmpty(forms.password)) { errors.push('Password must be filled') } if (isEmpty(forms.confirmPassword)) { errors.push('Confirm password must be filled') } if (keyExist(forms.confirmPassword) && forms.password !== forms.confirmPassword) { errors.push('Confirm password is not the same') } if (isEmpty(forms.address)) { errors.push('Address must be filled') } return errors } export { validateForm }
javascript
Washington, May 17: Fitness freaks, here is another reason to hit the gym! Higher levels of physical activity may lower the risk of 13 types of cancer, a new study has claimed. Physical inactivity is common, with an estimated 31 per cent of people worldwide not meeting recommended physical activity levels, researchers said. Any decrease in cancer risk associated with physical activity could be relevant to public health and cancer prevention efforts, they said. Steven C Moore from National Cancer Institute in the US and colleagues pooled data from 12 US and European cohorts with self-reported physical activity (1987-2004). They analysed associations of physical activity with the incidence of 26 kinds of cancer. The study included 1. 4 million participants and 186,932 cancers were identified during a median of 11 years of follow-up. Researchers found that higher levels of physical activity were associated with lower risk of 13 of 26 cancers - esophageal adenocarcinoma (42 per cent), liver (27 per cent), lung (26 percent); kidney (23 per cent), gastric cardia (22 per cent) and endometrial (21 per cent). Regular exercise also led to a lower risk for cancers like myeloid leukemia (20 per cent), myeloma (17 per cent), colon (16 per cent), head and neck (15 per cent), rectal (13 per cent), bladder (13 per cent) and breast (10 per cent). Most of the associations remained regardless of body size or smoking history, the study found. Overall, a higher level of physical activity was associated with a 7 per cent lower risk of total cancer, researchers said. "These findings support promoting physical activity as a key component of population-wide cancer prevention and control efforts," researchers said. The findings were published in the journal JAMA.
english
{ "name": "jsdoc-vue-component", "version": "2.2.4", "description": "A simple plugin for jsdoc (`pase vue SFC info to description`)", "main": "index.js", "repository": { "type": "git", "url": "git+https://github.com/ccqgithub/jsdoc-vue-component.git" }, "author": "", "license": "MIT", "bugs": { "url": "https://github.com/ccqgithub/jsdoc-vue-component/issues" }, "homepage": "https://github.com/ccqgithub/jsdoc-vue-component#readme", "dependencies": { "escodegen": "^1.9.0", "espree": "^3.5.2", "indent-string": "^3.2.0", "strip-indent": "^2.0.0", "vue-template-compiler": "2.4.2" } }
json
#include<bits/stdc++.h> using namespace std; inline int solveMeFirst(int a, int b){ return a + b; } int main() { int num1, num2, sum; cin >> num1 >> num2; sum = solveMeFirst(num1, num2); cout << sum; return 0; }
cpp
Last month, the Ludhiana administration discussed the vaccination drive for teenagers during a meeting held with school principals. It was decided that special camps would be organised in schools to cover maximum children between the ages 15 and 18. However, the surge in cases across states in the country prompted the shutdown of schools. Over the past few weeks, the education authorities and school management in many cities and towns of Punjab have expressed concern over the reopening of schools with the state polls scheduled in the coming days. A few school principals, who spoke to a leading news daily, have demanded that schools be allowed to reopen for vaccinated students with the parents’ consent. The Federation of Punjab Schools Association president Jagjit Singh Dhuri said that the government has failed both parents and schools and schools must be reopened with 50 per cent student strength. He also said that it is unfair to ask the authorities to shut the schools after all the financial crisis endured by the department. The principal of a CBSE school at BRS Nagar touched upon the fact that the Covid-19 outbreak, in most countries, has been declared as ‘endemic. ’ Further stating that it is important that every nation be prepared to fight the virus, the principal said that reopening schools is required for the development of children. According to Private School Association president Thakur Anand, behavioural changes have been seen in children which are leading to anxiety, panic attacks, stress, depression, and personality disorder. He slammed the government for letting the education institutes face the brunt of the political rallies. On behalf of the management, he asserted that they should be exempt from paying road tax, insurance, electricity bills, property tax, and water tax if the government is forcing the schools to remain shut. Preschool Association of Ludhiana president Amanpreet Singh, also principal of ABC Magical World at Bhai Randhir Singh Nagar, questioned the government for allowing political rallies with huge crowds and long processions; restaurants, bars, cinema houses, malls to open with 50 per cent capacity and even temples, gurdwaras, mosques, and churches to open with 100 per cent capacity.
english
<reponame>Djinoo/beefy-app<gh_stars>0 { "Disclaimer": "This project is in Beta. Use with caution and DYOR.", "Network-Error": "Network error", "Vault-Wallet": "Wallet", "Vault-Balance": "Balance", "Vault-APY": "APY", "Vault-APYDaily": "Daily", "Vault-TVL": "TVL", "Vault-Approving": "Approve...", "Vault-ApproveButton": "Approve", "Vault-DepositButton": "Deposit", "Vault-DepositButtonAll": "Deposit All", "Vault-Deposited": "Deposited", "Vault-Withdrawing": "Withdraw...", "Vault-WithdrawButton": "Withdraw", "Vault-WithdrawButtonAll": "Withdraw All", "Vault-WithdrawFee": "There is a 0.1% withdrawal fee on all vaults", "Vault-MainTitle": "Vaults", "Vault-SecondTitle": "Deposit & Earn money", "Hide-Zero-Balances": "Hide Zero Balances", "Retired-Vaults": "Retired Vaults", "Filters-Platform": "Platform", "Filters-Asset": "Asset", "Filters-Sort": "Sort by", "Filters-All": "All", "Filters-Default": "Default" }
json
Sonu Sood shares that Jackie Chan felt really special during his recent India tour; also says that his fitness was the reason he was chosen for Kung Fu Yoga. Actor Sonu Sood is happy with the response Kung Fu Yoga is garnering in India. The 43 year-old actor says, it was a huge effort to get superstar Jackie Chan to India for the film’s promotion. And he is content with the end result. “When we were shooting the film, I had discussed with Jackie and Stanley Tong (director) that we must promote the film in India. But when the time came, I realised how tough it was to bring him and his team here in a chartered flight from Beijing. Then, there were events to attend and of course, I had the whole industry waiting to meet Jackie. To do this for a Hindi film, it takes five days but we only had 10 hours,” he elucidates. Sonu says, the Chinese actor totally enjoyed his stay here. “While, I was driving him to the Film City, every time there was a green signal, he used to get excited and say, ‘Oh finally we have a green signal, let’s go’. He also kept on telling me that let’s leave his whole entourage (that was following him) behind and zoom ahead,” Sonu recalls. He says even though Jackie was starving during the promotion but he never complained. “Just before flying back, he held my hands and said Sonu we have promoted the film around the globe but India was the most special,” he says.
english
{ "db.username": "nucleus", "db.password": "<PASSWORD>", "db.url": "jdbc:postgresql://localhost:5432/nucleus", "batch.size": 1000, "thread.pool.size": 5, "interval.between.jobs.seconds": 600, "http.timeout.seconds": 30, "db.pool.size": 8 }
json
<reponame>ffalt/jamberry<filename>src/app/modules/tab-portal/index.ts<gh_stars>0 export * from './tab-portal.interfaces'; export * from './tab-portal-outlet';
typescript
<gh_stars>10-100 { "directions": [ "In a large pot over medium heat, combine the cranberry juice, sugar and water. Bring to a boil and stir until sugar is dissolved.", "Place the cinnamon, allspice and clove into a cheesecloth bag and tie shut. Add to the liquid and simmer 20 minutes.", "Remove spice bag. Just before serving, add fruit juices. Serve steaming hot." ], "ingredients": [ "2 cups cranberry juice", "2/3 cup white sugar", "7 cups water", "2 cinnamon sticks", "1 tablespoon whole allspice", "1 tablespoon whole cloves", "2 (20 ounce) cans pineapple juice", "6 fluid ounces frozen concentrated fruit punch", "1 (6 ounce) can frozen orange juice concentrate" ], "language": "en-US", "source": "allrecipes.com", "tags": [], "title": "Wassail II", "url": "http://allrecipes.com/recipe/15667/wassail-ii/" }
json
Lagan Laagi Thare Pyaar Ki Lyrics in Hindi sung by Haiyat Khan. This song is written and music composed by Swaroop Khan. |▶︎ See music video of Lagan Laagi Thare Pyaar Ki Song on Hitz Music YouTube channel for your reference and song details.
english
<gh_stars>0 { "result": { "item": "simpleprocessors:processorsocket" }, "type": "forge:ore_shapeless", "ingredients": [ { "item": "simpleprocessors:solderingiron" }, { "type": "forge:ore_dict", "ore": "dustRedstone" }, { "type": "forge:ore_dict", "ore": "stone" } ] }
json
<reponame>ageofambrosia/expresso .aboutButton { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #7892c2), color-stop(1, #476e9e)); background:-moz-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-webkit-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-o-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-ms-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:linear-gradient(to bottom, #7892c2 5%, #476e9e 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#7892c2', endColorstr='#476e9e',GradientType=0); background-color:#7892c2; -moz-border-radius:28px; -webkit-border-radius:28px; border-radius:28px; border:1px solid #4e6096; display:inline-block; cursor:pointer; color:#ffffff; font-size:17px; padding:14px 20px; text-decoration:none; text-shadow:0px 1px 0px #283966; position: absolute; } .aboutButton:hover { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #476e9e), color-stop(1, #7892c2)); background:-moz-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-webkit-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-o-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-ms-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:linear-gradient(to bottom, #476e9e 5%, #7892c2 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#476e9e', endColorstr='#7892c2',GradientType=0); background-color:#476e9e; } .aboutButton:active { position:relative; top:1px; } .buttons { position: absolute; top: 0; right: 0; padding-top: 10px; padding-right: 30px; } html { font-family: 'K2D'; } .Button { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #7892c2), color-stop(1, #476e9e)); background:-moz-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-webkit-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-o-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:-ms-linear-gradient(top, #7892c2 5%, #476e9e 100%); background:linear-gradient(to bottom, #7892c2 5%, #476e9e 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#7892c2', endColorstr='#476e9e',GradientType=0); background-color:#7892c2; -moz-border-radius:28px; -webkit-border-radius:28px; border-radius:28px; border:1px solid #4e6096; display:inline-block; cursor:pointer; color:#ffffff; font-size:17px; padding:14px 20px; text-decoration:none; text-shadow:0px 1px 0px #283966; position: absolute; } .Button:hover { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #476e9e), color-stop(1, #7892c2)); background:-moz-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-webkit-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-o-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:-ms-linear-gradient(top, #476e9e 5%, #7892c2 100%); background:linear-gradient(to bottom, #476e9e 5%, #7892c2 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#476e9e', endColorstr='#7892c2',GradientType=0); background-color:#476e9e; } .Button:active { position:relative; top:1px; } .disabledButton { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #8f8f8f), color-stop(1, #686868)); background:-moz-linear-gradient(top, #8f8f8f 5%, #686868 100%); background:-webkit-linear-gradient(top, #8f8f8f 5%, #686868 100%); background:-o-linear-gradient(top, #8f8f8f 5%, #686868 100%); background:-ms-linear-gradient(top, #8f8f8f 5%, #686868 100%); background:linear-gradient(to bottom, #8f8f8f 5%, #686868 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#8f8f8f', endColorstr='#686868',GradientType=0); background-color:#8f8f8f; -moz-border-radius:28px; -webkit-border-radius:28px; border-radius:28px; border:1px solid #616161; display:inline-block; cursor:pointer; color:#ffffff; font-family:Arial; font-size:17px; padding:16px 31px; text-decoration:none; text-shadow:0px 1px 0px #393939; } .disabledButton:hover { background:-webkit-gradient(linear, left top, left bottom, color-stop(0.05, #686868), color-stop(1, #8f8f8f)); background:-moz-linear-gradient(top, #686868 5%, #8f8f8f 100%); background:-webkit-linear-gradient(top, #686868 5%, #8f8f8f 100%); background:-o-linear-gradient(top, #686868 5%, #8f8f8f 100%); background:-ms-linear-gradient(top, #686868 5%, #8f8f8f 100%); background:linear-gradient(to bottom, #686868 5%, #8f8f8f 100%); filter:progid:DXImageTransform.Microsoft.gradient(startColorstr='#686868', endColorstr='#8f8f8f',GradientType=0); background-color:#686868; }
css