repo_name
stringlengths
7
65
path
stringlengths
5
187
copies
stringclasses
483 values
size
stringlengths
4
7
content
stringlengths
805
1.02M
license
stringclasses
14 values
mozilla/normandy
contract-tests/v3_api/test_group_update.py
1
1210
import uuid from support.assertions import assert_valid_schema from urllib.parse import urljoin def test_group_update(conf, requests_session, headers): # Create a new group data = {"name": str(uuid.uuid4())} response = requests_session.post( urljoin(conf.getoption("server"), "/api/v3/group/"), headers=headers, data=data ) assert response.status_code == 201 assert_valid_schema(response.json()) group_data = response.json() group_id = group_data["id"] # Verify group was stored and contains expected data response = requests_session.get( urljoin(conf.getoption("server"), "/api/v3/group/{}/".format(group_id)), headers=headers ) group_data = response.json() assert response.status_code == 200 assert_valid_schema(response.json()) # Use the update to change the name updated_data = {"name": str(uuid.uuid4())} response = requests_session.put( urljoin(conf.getoption("server"), "/api/v3/group/{}/".format(group_id)), headers=headers, data=updated_data, ) assert response.status_code == 200 assert_valid_schema(response.json()) assert response.json()["name"] == updated_data["name"]
mpl-2.0
mozilla/normandy
normandy/conftest.py
1
3099
from django.core.management import call_command from django.db import connection from django.db.migrations.executor import MigrationExecutor import pytest import requests_mock from graphene.test import Client as GrapheneClient from rest_framework.test import APIClient from normandy.schema import schema as normandy_schema from normandy.base.tests import UserFactory from normandy.recipes import geolocation as geolocation_module from normandy.recipes.tests import fake_sign @pytest.fixture def api_client(): """Fixture to provide a DRF API client.""" user = UserFactory(is_superuser=True) client = APIClient() client.force_authenticate(user=user) return client @pytest.fixture def gql_client(): """Fixture to provide a Graphene client.""" client = GrapheneClient(normandy_schema) return client @pytest.fixture def geolocation(): """Fixture to load geolocation data.""" geolocation_module.load_geoip_database() if geolocation_module.geoip_reader is None: pytest.skip() else: return geolocation_module @pytest.fixture def mocked_autograph(mocker): mocked = mocker.patch("normandy.recipes.models.Autographer") mocked.return_value.sign_data.side_effect = fake_sign return mocked @pytest.fixture def mocked_remotesettings(mocker): return mocker.patch("normandy.recipes.models.RemoteSettings") @pytest.fixture def rs_settings(settings): settings.REMOTE_SETTINGS_URL = "https://remotesettings.example.com/v1" settings.REMOTE_SETTINGS_USERNAME = "normandy" settings.REMOTE_SETTINGS_PASSWORD = "n0rm4ndy" return settings @pytest.fixture() def migrations(transactional_db): """ This fixture returns a helper object to test Django data migrations. Based on: https://gist.github.com/bennylope/82a6088c02fefdd47e18f3c04ec167af """ class Migrator(object): def migrate(self, app, to): migration = [(app, to)] executor = MigrationExecutor(connection) executor.migrate(migration) return executor.loader.project_state(migration).apps def reset(self): call_command("migrate", no_input=True) return Migrator() @pytest.fixture def requestsmock(): """Return a context where requests are all mocked. Usage:: def test_something(requestsmock): requestsmock.get( 'https://example.com/path' content=b'The content' ) # Do stuff that involves requests.get('http://example.com/path') """ with requests_mock.mock() as m: yield m @pytest.fixture def storage(settings): settings.DEFAULT_FILE_STORAGE = "normandy.base.storage.NormandyInMemoryStorage" from django.core.files.storage import default_storage yield default_storage dirs_to_delete = ["/"] while len(dirs_to_delete) > 0: dir_path = dirs_to_delete.pop() paths, new_dirs = default_storage.listdir(dir_path) dirs_to_delete.extend(new_dirs) for path in paths: default_storage.delete(path)
mpl-2.0
mozilla/normandy
normandy/base/tests/test_checks.py
1
1449
import pytest from django.core.checks.registry import run_checks from normandy.base import checks as base_checks from normandy.recipes import checks as recipe_checks, geolocation as geolocation_module @pytest.mark.django_db def test_run_checks_happy_path(): errors = set(e.id for e in run_checks()) expected = set() # If geolocation isn't enabled, expect that their is a geolocation warning geolocation_module.load_geoip_database() if geolocation_module.geoip_reader is None: expected.add(recipe_checks.ERROR_GEOIP_DB_NOT_AVAILABLE) assert errors == expected @pytest.mark.django_db def test_run_checks_all_things_that_can_go_wrong(settings): settings.CDN_URL = "http://cdn.example.com" settings.APP_SERVER_URL = "http://app.example.com" settings.OIDC_REMOTE_AUTH_HEADER = "Y_HTTP_HEADER" settings.OIDC_LOGOUT_URL = None settings.USE_OIDC = True errors = run_checks() assert errors != [] error_ids = [x.id for x in errors] assert base_checks.ERROR_MISCONFIGURED_CDN_URL_SLASH in error_ids assert base_checks.ERROR_MISCONFIGURED_CDN_URL_HTTPS in error_ids assert base_checks.ERROR_MISCONFIGURED_APP_SERVER_URL_SLASH in error_ids assert base_checks.ERROR_MISCONFIGURED_APP_SERVER_URL_HTTPS in error_ids assert base_checks.ERROR_MISCONFIGURED_OIDC_LOGOUT_URL in error_ids assert base_checks.WARNING_MISCONFIGURED_OIDC_REMOTE_AUTH_HEADER_PREFIX in error_ids
mpl-2.0
mozilla/normandy
normandy/base/tests/test_auth_backends.py
1
3238
import pytest from django.test import RequestFactory from normandy.base.auth_backends import ( INFO_LOGIN_SUCCESS, LoggingModelBackend, EmailOnlyRemoteUserBackend, WARNING_LOGIN_FAILURE, ) from normandy.base.tests import UserFactory, Whatever class TestLoggingModelBackend(object): @pytest.fixture def mock_logger(self, mocker): return mocker.patch("normandy.base.auth_backends.logger") @pytest.fixture def mock_authenticate(self, mocker): return mocker.patch("normandy.base.auth_backends.ModelBackend.authenticate") def test_log_failure_username(self, mock_logger, mock_authenticate): mock_authenticate.return_value = None request = RequestFactory().get("/") user = LoggingModelBackend().authenticate( request, username="fakeuser", password="does.not.exist" ) assert user is None mock_logger.warning.assert_called_with( Whatever.contains("fakeuser"), extra={"code": WARNING_LOGIN_FAILURE} ) def test_log_failure_no_username(self, mock_logger, mock_authenticate): mock_authenticate.return_value = None request = RequestFactory().get("/") user = LoggingModelBackend().authenticate(request, password="does.not.exist") assert user is None mock_logger.warning.assert_called_with( Whatever.contains("no username provided"), extra={"code": WARNING_LOGIN_FAILURE} ) def test_log_success(self, mock_logger, mock_authenticate): mock_authenticate.return_value = True request = RequestFactory().get("/") user = LoggingModelBackend().authenticate( request, username="fakeuser", password="does.not.exist" ) assert user mock_logger.info.assert_called_with( Whatever.contains("fakeuser"), extra={"code": INFO_LOGIN_SUCCESS} ) @pytest.mark.django_db class TestEmailOnlyRemoteUserBackend(object): @pytest.fixture def backend(self): return EmailOnlyRemoteUserBackend() def test_it_works(self, backend): user = backend.authenticate(request=None, remote_user="test@example.com") assert user is not None assert not user.is_anonymous def test_it_requires_an_email(self, backend): user = backend.authenticate(request=None, remote_user="not_an_email") assert user is None def test_it_adds_an_email_to_the_user(self, backend): email = "test@example.com" user = backend.authenticate(request=None, remote_user=email) assert user.email == email def test_existing_user(self, backend): email = "test@example.com" existing_user = UserFactory(username=email, email=email) logged_in_user = backend.authenticate(request=None, remote_user=email) assert existing_user == logged_in_user assert logged_in_user.email == email def test_existing_user_no_email(self, backend): email = "test@example.com" existing_user = UserFactory(username=email, email="") logged_in_user = backend.authenticate(request=None, remote_user=email) assert existing_user == logged_in_user assert logged_in_user.email == email
mpl-2.0
mozilla/normandy
normandy/recipes/tests/test_signing.py
1
17765
import base64 import os from datetime import datetime, timedelta from unittest.mock import MagicMock, call from django.core.exceptions import ImproperlyConfigured import pytest import pytz from pyasn1.type import useful as pyasn1_useful from pyasn1_modules import rfc5280 from normandy.base.tests import Whatever from normandy.recipes import signing @pytest.fixture def mock_logger(mocker): return mocker.patch("normandy.recipes.signing.logger") class TestAutographer(object): test_settings = { "URL": "https://autograph.example.com/", "HAWK_ID": "hawk id", "HAWK_SECRET_KEY": "hawk secret key", } def test_it_checks_settings(self, settings): """Test that each required key is required individually""" # Leave out URL settings.AUTOGRAPH_URL = None settings.AUTOGRAPH_HAWK_ID = "hawk id" settings.AUTOGRAPH_HAWK_SECRET_KEY = "hawk secret key" with pytest.raises(ImproperlyConfigured) as exc: signing.Autographer() assert "AUTOGRAPH_URL" in str(exc.value) # Leave out HAWK_ID settings.AUTOGRAPH_URL = "https://autograph.example.com" settings.AUTOGRAPH_HAWK_ID = None settings.AUTOGRAPH_HAWK_SECRET_KEY = "hawk secret key" with pytest.raises(ImproperlyConfigured) as exc: signing.Autographer() assert "AUTOGRAPH_HAWK_ID" in str(exc.value) # Leave out HAWK_SECRET_KEY settings.AUTOGRAPH_URL = "https://autograph.example.com" settings.AUTOGRAPH_HAWK_ID = "hawk id" settings.AUTOGRAPH_HAWK_SECRET_KEY = None with pytest.raises(ImproperlyConfigured) as exc: signing.Autographer() assert "AUTOGRAPH_HAWK_SECRET_KEY" in str(exc.value) # Include everything settings.AUTOGRAPH_URL = "https://autograph.example.com" settings.AUTOGRAPH_HAWK_ID = "hawk id" settings.AUTOGRAPH_HAWK_SECRET_KEY = "hawk secret key" # assert doesn't raise signing.Autographer() def test_it_interacts_with_autograph_correctly(self, settings, mock_logger): settings.AUTOGRAPH_URL = "https://autograph.example.com" settings.AUTOGRAPH_HAWK_ID = "hawk id" settings.AUTOGRAPH_HAWK_SECRET_KEY = "hawk secret key" autographer = signing.Autographer() autographer.session = MagicMock() autographer.session.post.return_value.json.return_value = [ { "content-signature": ( 'x5u="https://example.com/fake_x5u_1";p384ecdsa=fake_signature_1' ), "x5u": "https://example.com/fake_x5u_1", "hash_algorithm": "sha384", "ref": "fake_ref_1", "signature": "fake_signature_1", }, { "content-signature": ( 'x5u="https://example.com/fake_x5u_2";p384ecdsa=fake_signature_2' ), "x5u": "https://example.com/fake_x5u_2", "hash_algorithm": "sha384", "ref": "fake_ref_2", "signature": "fake_signature_2", }, ] url = self.test_settings["URL"] + "sign/data" foo_base64 = base64.b64encode(b"foo").decode("utf8") bar_base64 = base64.b64encode(b"bar").decode("utf8") # Assert the correct data is returned assert autographer.sign_data([b"foo", b"bar"]) == [ { "timestamp": Whatever(), "signature": "fake_signature_1", "x5u": "https://example.com/fake_x5u_1", }, { "timestamp": Whatever(), "signature": "fake_signature_2", "x5u": "https://example.com/fake_x5u_2", }, ] # Assert that logging happened mock_logger.info.assert_has_calls( [ call(Whatever.contains("2"), extra={"code": signing.INFO_RECEIVED_SIGNATURES}), call(Whatever.contains("fake_ref_1")), call(Whatever.contains("fake_ref_2")), ] ) # Assert the correct request was made assert autographer.session.post.called_once_with( [ url, [ {"template": "content-signature", "input": foo_base64}, {"template": "content-signature", "input": bar_base64}, ], ] ) class TestVerifySignaturePubkey(object): # known good data data = '{"action":"console-log","arguments":{"message":"telemetry available"},"enabled":true,"filter_expression":"telemetry != undefined","id":1,"last_updated":"2017-01-02T11:32:07.687408Z","name":"mython\'s system addon test","revision_id":"6dc874ded7d14af9ef9c147c5d2ceef9d15b56ca933681e574cd96a50b75946e"}' # noqa signature = "Prb0Jnb3icT0g_hZkgEyuzTlWrsTYrURXy6mzDTDh9WmqXdQBS05cV1mL1GUBbKIgcs02V8P1JvYR4LaJmSUvfZWYS66Jo8AdhcfEikoEZ0jvlLtiJDFxpT4eh2pk3XW" # noqa pubkey = "MHYwEAYHKoZIzj0CAQYFK4EEACIDYgAEVEKiCAIkwRg1VFsP8JOYdSF6a3qvgbRPoEK9eTuLbrB6QixozscKR4iWJ8ZOOX6RPCRgFdfVDoZqjFBFNJN9QtRBk0mVtHbnErx64d2vMF0oWencS1hyLW2whgOgOz7p" # noqa def test_known_good_signature(self): assert signing.verify_signature_pubkey(self.data, self.signature, self.pubkey) def test_raises_nice_error_for_too_short_signatures_bad_padding(self): signature = "a_too_short_signature" with pytest.raises(signing.WrongSignatureSize): signing.verify_signature_pubkey(self.data, signature, self.pubkey) def test_raises_nice_error_for_too_short_signatures_good_base64(self): signature = "aa==" with pytest.raises(signing.WrongSignatureSize): signing.verify_signature_pubkey(self.data, signature, self.pubkey) def test_raises_nice_error_for_wrong_signature(self): # change the signature, but keep it a valid signature signature = self.signature.replace("s", "S") with pytest.raises(signing.SignatureDoesNotMatch): signing.verify_signature_pubkey(self.data, signature, self.pubkey) class TestVerifySignatureX5U(object): def test_happy_path(self, mocker): mock_verify_x5u = mocker.patch("normandy.recipes.signing.verify_x5u") mock_der_encode = mocker.patch("normandy.recipes.signing.der_encode") mock_verify_signature_pubkey = mocker.patch( "normandy.recipes.signing.verify_signature_pubkey" ) data = "abc" signature = "signature" x5u = "http://example.com/cert" cert_contents = b"cert_contents" encoded_cert_contents = base64.b64encode(cert_contents).decode() mock_der_encode.return_value = cert_contents public_key = "public_key" cert = {"tbsCertificate": {"subjectPublicKeyInfo": public_key}} mock_verify_x5u.return_value = cert ret = signing.verify_signature_x5u(data, signature, x5u) mock_verify_x5u.assert_called_with(x5u) mock_der_encode.assert_called_with(public_key) mock_verify_signature_pubkey.assert_called_with(data, signature, encoded_cert_contents) assert ret == mock_verify_signature_pubkey.return_value class TestExtractCertsFromPem(object): def test_empty(self): assert signing.extract_certs_from_pem("") == [] def test_real_certs(self): path = os.path.join(os.path.dirname(__file__), "data", "test_certs.pem") with open(path) as f: certs = signing.extract_certs_from_pem(f.read()) assert len(certs) == 3 def test_incomplete_cert(self): bad_data = "-----BEGIN CERTIFICATE-----\nMIIGXTCCBEWgAwIBAgIEAQAACjANBgkq" with pytest.raises(signing.CertificateParseError) as exc: signing.extract_certs_from_pem(bad_data) assert "Unexpected end of input." in str(exc.value) def test_not_a_cert(self): bad_data = "hello world" with pytest.raises(signing.CertificateParseError) as exc: signing.extract_certs_from_pem(bad_data) assert 'Unexpected input "hello world"' in str(exc.value) class TestParseCertsFromDer(object): def test_real_certs(self): path = os.path.join(os.path.dirname(__file__), "data", "test_certs.pem") with open(path) as f: ders = signing.extract_certs_from_pem(f.read()) certs = [signing.parse_cert_from_der(der) for der in ders] # Quick spot check on the CommonName value of the subjects of the certs # If these are correct, the entire objects were probably parsed correctly assert ( certs[0]["tbsCertificate"]["subject"]["rdnSequence"][4][0]["value"] == b"\x13&normandy.content-signature.mozilla.org" ) assert ( certs[1]["tbsCertificate"]["subject"]["rdnSequence"][3][0]["value"] == b"\x13'Mozilla Signing Services Intermediate 1" ) assert ( certs[2]["tbsCertificate"]["subject"]["rdnSequence"][3][0]["value"] == b"\x13\x16root-ca-production-amo" ) class TestCheckValidity(object): def test_it_works(self): now = datetime.utcnow().replace(tzinfo=pytz.utc) not_before = now - timedelta(days=1) not_after = now + timedelta(days=1) assert signing.check_validity(not_before, not_after, None) def test_not_yet_valid(self): now = datetime.utcnow().replace(tzinfo=pytz.utc) not_before = now + timedelta(days=1) not_after = now + timedelta(days=2) with pytest.raises(signing.CertificateNotYetValid): signing.check_validity(not_before, not_after, None) def test_expired(self): now = datetime.utcnow().replace(tzinfo=pytz.utc) not_before = now - timedelta(days=2) not_after = now - timedelta(days=1) with pytest.raises(signing.CertificateExpired): signing.check_validity(not_before, not_after, None) def test_expiring_early_ok(self): now = datetime.utcnow().replace(tzinfo=pytz.utc) not_before = now - timedelta(days=1) not_after = now + timedelta(days=3) expire_early = timedelta(days=2) assert signing.check_validity(not_before, not_after, expire_early) def test_expiring_early_not_ok(self): now = datetime.utcnow().replace(tzinfo=pytz.utc) not_before = now - timedelta(days=1) not_after = now + timedelta(days=1) expire_early = timedelta(days=2) with pytest.raises(signing.CertificateExpiringSoon): signing.check_validity(not_before, not_after, expire_early) class TestVerifyX5u(object): def _fake_cert(self, not_before=None, not_after=None): fake_cert = rfc5280.Certificate() fake_cert["tbsCertificate"] = rfc5280.TBSCertificate() fake_cert["tbsCertificate"]["validity"] = rfc5280.Validity() if not_before: fake_cert["tbsCertificate"]["validity"]["notBefore"] = rfc5280.Time() fake_cert["tbsCertificate"]["validity"]["notBefore"][ "utcTime" ] = pyasn1_useful.UTCTime.fromDateTime(not_before) if not_after: fake_cert["tbsCertificate"]["validity"]["notAfter"] = rfc5280.Time() fake_cert["tbsCertificate"]["validity"]["notAfter"][ "utcTime" ] = pyasn1_useful.UTCTime.fromDateTime(not_after) return fake_cert def test_it_works(self, mocker, settings): settings.CERTIFICATES_CHECK_VALIDITY = True settings.CERTIFICATES_EXPECTED_ROOT_HASH = None settings.CERTIFICATES_EXPECTED_SUBJECT_CN = None settings.X5U_CACHE_TIME = 0 # don't cache, since mocks can't be cached mock_requests = mocker.patch("normandy.recipes.signing.requests") mock_extract_certs_from_pem = mocker.patch( "normandy.recipes.signing.extract_certs_from_pem" ) mock_parse_cert_from_der = mocker.patch("normandy.recipes.signing.parse_cert_from_der") url = "https://example.com/cert.pem" now = datetime.now() not_before = now - timedelta(days=1) not_after = now + timedelta(days=1) mock_extract_certs_from_pem.return_value = ["a", "b"] mock_parse_cert_from_der.return_value = self._fake_cert( not_before=not_before, not_after=not_after ) assert signing.verify_x5u(url) == mock_parse_cert_from_der.return_value assert mock_requests.get.called_once_with(url) body = mock_requests.get.return_value.content.decode.return_value assert mock_extract_certs_from_pem.called_once_with(body) assert mock_parse_cert_from_der.called_twice() def test_invalid_dates(self, mocker, settings): settings.CERTIFICATES_CHECK_VALIDITY = True settings.CERTIFICATES_EXPECTED_ROOT_HASH = None settings.CERTIFICATES_EXPECTED_SUBJECT_CN = None settings.X5U_CACHE_TIME = 0 # don't cache, since mocks can't be cached mock_requests = mocker.patch("normandy.recipes.signing.requests") mock_extract_certs_from_pem = mocker.patch( "normandy.recipes.signing.extract_certs_from_pem" ) mock_parse_cert_from_der = mocker.patch("normandy.recipes.signing.parse_cert_from_der") url = "https://example.com/cert.pem" now = datetime.now().replace(tzinfo=pytz.UTC) not_before = now - timedelta(days=2) not_after = now - timedelta(days=1) mock_extract_certs_from_pem.return_value = ["a"] mock_parse_cert_from_der.return_value = self._fake_cert( not_before=not_before, not_after=not_after ) with pytest.raises(signing.CertificateExpired): signing.verify_x5u(url) assert mock_requests.get.called_once_with(url) body = mock_requests.get.return_value.content.decode.return_value assert mock_extract_certs_from_pem.called_once_with(body) assert mock_parse_cert_from_der.called_once_with( mock_extract_certs_from_pem.return_value[0] ) def test_mixed_timestamp_format(self, mocker): # The certificate used for testing expired on 2018-04-24. This test is # only concerned with the parsing of the dates, so mock the call to the # validate function and assert about the values of the dates. mock_requests = mocker.patch("normandy.recipes.signing.requests") mock_check_validity = mocker.patch("normandy.recipes.signing.check_validity") path = os.path.join(os.path.dirname(__file__), "data", "mixed_timestamps_certs.pem") with open(path, "rb") as f: mock_requests.get.return_value.content = f.read() assert signing.verify_x5u("https://example.com/cert.pem") assert mock_check_validity.mock_calls == [ call( datetime(2017, 12, 25, tzinfo=pytz.UTC), datetime(2018, 4, 24, tzinfo=pytz.UTC), None, ), call( datetime(2017, 5, 4, 0, 12, 39, tzinfo=pytz.UTC), datetime(2019, 5, 4, 0, 12, 39, tzinfo=pytz.UTC), None, ), call( datetime(2015, 3, 17, 22, 53, 57, tzinfo=pytz.UTC), datetime(2025, 3, 14, 22, 53, 57, tzinfo=pytz.UTC), None, ), ] def test_it_checks_cert_root(self, mocker, settings): path = os.path.join(os.path.dirname(__file__), "data", "test_certs.pem") with open(path) as f: cert_pem = f.read() settings.CERTIFICATES_CHECK_VALIDITY = False settings.CERTIFICATES_EXPECTED_ROOT_HASH = "CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF:EE:CO:FF" settings.CERTIFICATES_EXPECTED_SUBJECT_CN = None mock_requests = mocker.patch("normandy.recipes.signing.requests") mock_requests.get.return_value.content.decode.return_value = cert_pem with pytest.raises(signing.CertificateHasWrongRoot): signing.verify_x5u("https://example.com/cert.pem") def test_it_checks_cert_subject(self, mocker, settings): path = os.path.join(os.path.dirname(__file__), "data", "test_certs.pem") with open(path) as f: cert_pem = f.read() settings.CERTIFICATES_CHECK_VALIDITY = False settings.CERTIFICATES_EXPECTED_ROOT_HASH = None settings.CERTIFICATES_EXPECTED_SUBJECT_CN = "wrong.subject.example.com" mock_requests = mocker.patch("normandy.recipes.signing.requests") mock_requests.get.return_value.content.decode.return_value = cert_pem with pytest.raises(signing.CertificateHasWrongSubject): signing.verify_x5u("https://example.com/cert.pem") class TestReadTimestampObject(object): def test_it_reads_utc_time_format(self): dt = datetime(2018, 1, 25, 16, 1, 13, 0, tzinfo=pytz.UTC) obj = rfc5280.Time() obj["utcTime"] = pyasn1_useful.UTCTime.fromDateTime(dt) assert signing.read_timestamp_object(obj) == dt def test_it_reads_general_time_format(self): dt = datetime(2018, 1, 25, 16, 1, 13, 0, tzinfo=pytz.UTC) obj = rfc5280.Time() obj["generalTime"] = pyasn1_useful.GeneralizedTime.fromDateTime(dt) assert signing.read_timestamp_object(obj) == dt def test_it_errors_on_unsupported_formats(self): with pytest.raises(signing.BadCertificate) as exc: signing.read_timestamp_object({"unsupportedTimestamp": b"gibberish"}) assert "Timestamp not in expected format" in str(exc.value) assert "unsupportedTimestamp" in str(exc.value)
mpl-2.0
mozilla/normandy
normandy/recipes/migrations/0005_auto_20180503_2146.py
1
2487
# -*- coding: utf-8 -*- # Generated by Django 1.11.11 on 2018-05-03 21:46 from __future__ import unicode_literals from django.conf import settings from django.db import migrations, models import django.db.models.deletion import django.utils.timezone class Migration(migrations.Migration): dependencies = [("recipes", "0004_auto_20180502_2340")] operations = [ migrations.RemoveField(model_name="approvalrequest", name="revision"), migrations.RemoveField(model_name="recipe", name="approved_revision"), migrations.RemoveField(model_name="recipe", name="latest_revision"), migrations.DeleteModel(name="RecipeRevision"), migrations.RenameModel("TmpRecipeRevision", "RecipeRevision"), migrations.AlterField( model_name="reciperevision", name="action", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, related_name="recipe_revisions", to="recipes.Action", ), ), migrations.AlterField( model_name="reciperevision", name="recipe", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, related_name="revisions", to="recipes.Recipe", ), ), migrations.AlterField( model_name="reciperevision", name="user", field=models.ForeignKey( null=True, on_delete=django.db.models.deletion.SET_NULL, related_name="recipe_revisions", to=settings.AUTH_USER_MODEL, ), ), migrations.RenameField( model_name="approvalrequest", old_name="tmp_revision", new_name="revision" ), migrations.RenameField( model_name="recipe", old_name="approved_tmp_revision", new_name="approved_revision" ), migrations.RenameField( model_name="recipe", old_name="latest_tmp_revision", new_name="latest_revision" ), migrations.AlterField( model_name="approvalrequest", name="revision", field=models.OneToOneField( default=None, on_delete=django.db.models.deletion.CASCADE, related_name="approval_request", to="recipes.RecipeRevision", ), preserve_default=False, ), ]
mpl-2.0
mozilla/normandy
normandy/base/tests/test_utils.py
1
2452
import json from normandy.base.utils import canonical_json_dumps, get_client_ip, sri_hash class TestGetClientIp(object): def test_no_proxies(self, rf, settings): """If there are no proxies, REMOTE_ADDR should be used.""" settings.NUM_PROXIES = 0 client_ip = "1.1.1.1" req = rf.get("/", HTTP_X_FORWARDED_FOR="fake", REMOTE_ADDR=client_ip) assert get_client_ip(req) == client_ip def test_one_proxy(self, rf, settings): """ If there is one proxy, the right-most value in HTTP_X_FORWARDED_FOR should be used. """ settings.NUM_PROXIES = 1 client_ip = "1.1.1.1" nginx_ip = "2.2.2.2" forwarded_for = ", ".join(["fake", client_ip]) req = rf.get("/", HTTP_X_FORWARDED_FOR=forwarded_for, REMOTE_ADDR=nginx_ip) assert get_client_ip(req) == client_ip def test_two_proxies(self, rf, settings): """ If there are two proxies, the second-from-the-right value in HTTP_X_FORWARDED_FOR should be used. """ settings.NUM_PROXIES = 2 client_ip = "1.1.1.1" elb_ip = "2.2.2.2" nginx_ip = "3.3.3.3" forwarded_for = ", ".join(["fake", client_ip, elb_ip]) req = rf.get("/", HTTP_X_FORWARDED_FOR=forwarded_for, REMOTE_ADDR=nginx_ip) assert get_client_ip(req) == client_ip class TestCanonicalJsonDumps(object): def test_it_works(self): data = {"a": 1, "b": 2} assert canonical_json_dumps(data) == '{"a":1,"b":2}' def test_it_works_with_euro_signs(self): data = {"USD": "$", "EURO": "€"} assert canonical_json_dumps(data) == r'{"EURO":"\u20ac","USD":"$"}' def test_it_escapes_quotes_properly(self): data = {"message": 'It "works", I think'} dumped = canonical_json_dumps(data) assert dumped == r'{"message":"It \"works\", I think"}' json.loads(dumped) class TestSRIHash(object): def test_it_works(self): # Pre-generated base64 hash of the string "foobar" expected = "sha384-PJww2fZl501RXIQpYNSkUcg6ASX9Pec5LXs3IxrxDHLqWK7fzfiaV2W/kCr5Ps8G" assert sri_hash(b"foobar") == expected def test_url_safe_works(self): # Pre-generated base64 hash of the string "normandy", urlsafe-ed expected = "sha384-6FydcL0iVnTqXT3rBg6YTrlz0K-mw57n9zxTEmxYG6FIO_vZTMlTWsbkxHchsO65" assert sri_hash(b"normandy", url_safe=True) == expected
mpl-2.0
mozilla/normandy
contract-tests/v3_api/test_approval_request_close.py
1
1496
from support.assertions import assert_valid_schema from support.helpers import new_recipe from urllib.parse import urljoin def test_approval_request_close(conf, requests_session, headers): # Get an action we can work with action_response = requests_session.get( urljoin(conf.getoption("server"), "/api/v3/action/"), headers=headers ) data = action_response.json() action_id = data["results"][0]["id"] # Create a recipe recipe_details = new_recipe(requests_session, action_id, conf.getoption("server"), headers) # Create an approval request response = requests_session.post( urljoin( conf.getoption("server"), "/api/v3/recipe_revision/{}/request_approval/".format( recipe_details["latest_revision_id"] ), ), headers=headers, ) data = response.json() approval_id = data["id"] assert response.status_code != 404 assert_valid_schema(response.json()) # Close the approval request response = requests_session.post( urljoin( conf.getoption("server"), "/api/v3/approval_request/{}/close/".format(approval_id) ), headers=headers, ) assert response.status_code == 204 # Verify that is no longer exists response = requests_session.get( urljoin(conf.getoption("server"), "/api/v3/approval_request/{}/".format(approval_id)), headers=headers, ) assert response.status_code == 404
mpl-2.0
mozilla/normandy
normandy/recipes/migrations/0014_auto_20190228_1128.py
1
2576
# Generated by Django 2.0.13 on 2019-02-28 11:28 import json import sys from urllib.parse import unquote_plus, urlparse from django.db import migrations def get_filename_from_url(url): return unquote_plus(urlparse(url).path.split("/")[-1]) def add_extension_id(apps, schema_editor): Action = apps.get_model("recipes", "Action") RecipeRevision = apps.get_model("recipes", "RecipeRevision") Extension = apps.get_model("studies", "Extension") failures = [] try: action = Action.objects.get(name="opt-out-study") except Action.DoesNotExist: return # Do nothing since there cannot be any recipes using the opt-out-study action revisions = RecipeRevision.objects.filter(action_id=action.id) for revision in revisions: arguments = json.loads(revision.arguments_json) url = arguments.get("addonUrl") filename = get_filename_from_url(url) try: extension = Extension.objects.get(xpi=f"extensions/{filename}") except Extension.DoesNotExist: failures.append( { "filename": filename, "addon_url": arguments.get("addonUrl"), "revision_id": revision.id, "recipe_id": revision.recipe.id, } ) else: arguments["extensionApiId"] = extension.id revision.arguments_json = json.dumps(arguments) revision.save() if failures: for failure in failures: sys.stderr.write(f"{failure}\n") raise Exception("There were failures in this migration.") def remove_extension_id(apps, schema_editor): Action = apps.get_model("recipes", "Action") RecipeRevision = apps.get_model("recipes", "RecipeRevision") try: action = Action.objects.get(name="opt-out-study") except Action.DoesNotExist: return # Do nothing since there cannot be any recipes using the opt-out-study action revisions = RecipeRevision.objects.filter(action_id=action.id) for revision in revisions: arguments = json.loads(revision.arguments_json) if "extensionApiId" in arguments: arguments.pop("extensionApiId") revision.arguments_json = json.dumps(arguments) revision.save() class Migration(migrations.Migration): dependencies = [ ("recipes", "0013_auto_20181018_2049"), ("studies", "0006_extension_hash_algorithm"), ] operations = [migrations.RunPython(add_extension_id, remove_extension_id)]
mpl-2.0
mozilla/normandy
normandy/recipes/tests/test_schema.py
1
2480
import pytest from normandy.base.tests import GQ from normandy.recipes.tests import ActionFactory, ApprovalRequestFactory, RecipeFactory @pytest.mark.django_db class TestQuery(object): def test_resolve_all_action(self, gql_client): a = ActionFactory() res = gql_client.execute(GQ().query.allActions.fields("id")) assert res == {"data": {"allActions": [{"id": str(a.id)}]}} def test_resolve_action_by_id(self, gql_client): a = ActionFactory() res = gql_client.execute(GQ().query.action(id=a.id).fields("name")) assert res == {"data": {"action": {"name": a.name}}} def test_resolve_action_by_name(self, gql_client): a = ActionFactory() res = gql_client.execute(GQ().query.action(name=a.name).fields("id")) assert res == {"data": {"action": {"id": str(a.id)}}} def test_resolve_all_approval_requests(self, gql_client): a = ApprovalRequestFactory() res = gql_client.execute(GQ().query.allApprovalRequests.fields("id")) assert res == {"data": {"allApprovalRequests": [{"id": str(a.id)}]}} def test_resolve_approval_request_by_id(self, gql_client): a = ApprovalRequestFactory() res = gql_client.execute( GQ().query.approvalRequest(id=a.id).fields(GQ().revision.fields("id")) ) assert res == {"data": {"approvalRequest": {"revision": {"id": str(a.revision.id)}}}} def test_resolve_all_recipes(self, gql_client): r = RecipeFactory() res = gql_client.execute(GQ().query.allRecipes.fields("id")) assert res == {"data": {"allRecipes": [{"id": str(r.id)}]}} def test_resolve_recipe_by_id(self, gql_client): r = RecipeFactory() res = gql_client.execute( GQ().query.recipe(id=r.id).fields(GQ().latestRevision.fields("id")) ) assert res == {"data": {"recipe": {"latestRevision": {"id": str(r.latest_revision.id)}}}} def test_resolve_all_recipe_revisions(self, gql_client): r = RecipeFactory() res = gql_client.execute(GQ().query.allRecipeRevisions.fields("id")) assert res == {"data": {"allRecipeRevisions": [{"id": str(r.latest_revision.id)}]}} def test_resolve_recipe_revision_by_id(self, gql_client): r = RecipeFactory() res = gql_client.execute(GQ().query.recipeRevision(id=r.latest_revision.id).fields("id")) assert res == {"data": {"recipeRevision": {"id": str(r.latest_revision.id)}}}
mpl-2.0
developmentseed/landsat-util
setup.py
1
1158
#!/usr/bin/env python # Landsat Util # License: CC0 1.0 Universal try: from setuptools import setup setup_kwargs = {'entry_points': {'console_scripts':['landsat=landsat.landsat:__main__']}} except ImportError: from distutils.core import setup setup_kwargs = {'scripts': ['bin/landsat']} from landsat import __version__ def readme(): with open('README.rst') as f: return f.read() with open('requirements.txt') as fid: INSTALL_REQUIRES = [l.strip() for l in fid.readlines() if l] with open('requirements-dev.txt') as fid: TEST_REQUIRES = [l.strip() for l in fid.readlines() if l] setup( name='landsat-util', version=__version__, description='A utility to search, download and process Landsat 8' + ' satellite imagery', long_description=readme(), author='Development Seed', author_email='info@developmentseed.org', url='https://github.com/developmentseed/landsat-util', packages=['landsat'], include_package_data=True, license='CCO', platforms='Posix; MacOS X; Windows', install_requires=INSTALL_REQUIRES, tests_require=TEST_REQUIRES, **setup_kwargs )
cc0-1.0
developmentseed/landsat-util
landsat/mixins.py
3
2950
# Pansharpened Image Process using Rasterio # Landsat Util # License: CC0 1.0 Universal from __future__ import print_function, division, absolute_import import sys import subprocess from termcolor import colored class VerbosityMixin(object): """ Verbosity Mixin that generates beautiful stdout outputs. """ verbose = False def output(self, value, normal=False, color=None, error=False, arrow=False, indent=None): """ Handles verbosity of this calls. if priority is set to 1, the value is printed if class instance verbose is True, the value is printed :param value: a string representing the message to be printed :type value: String :param normal: if set to true the message is always printed, otherwise it is only shown if verbosity is set :type normal: boolean :param color: The color of the message, choices: 'red', 'green', 'blue' :type normal: String :param error: if set to true the message appears in red :type error: Boolean :param arrow: if set to true an arrow appears before the message :type arrow: Boolean :param indent: indents the message based on the number provided :type indent: Boolean :returns: void """ if error and value and (normal or self.verbose): return self._print(value, color='red', indent=indent) if self.verbose or normal: return self._print(value, color, arrow, indent) return def subprocess(self, argv): """ Execute subprocess commands with proper ouput. This is no longer used in landsat-util :param argv: A list of subprocess arguments :type argv: List :returns: void """ if self.verbose: proc = subprocess.Popen(argv, stderr=subprocess.PIPE) else: proc = subprocess.Popen(argv, stdout=subprocess.PIPE, stderr=subprocess.PIPE) self.output(proc.stderr.read(), error=True) return def exit(self, message): """ outputs an exit message and exits :param message: The message to be outputed :type message: String :returns: void """ self.output(message, normal=True, color="green") sys.exit() def _print(self, msg, color=None, arrow=False, indent=None): """ Print the msg with the color provided. """ if color: msg = colored(msg, color) if arrow: msg = colored('===> ', 'blue') + msg if indent: msg = (' ' * indent) + msg print(msg) return msg
cc0-1.0
rmmh/skybot
plugins/seen.py
3
3063
" seen.py: written by sklnd in about two beers July 2009" from builtins import object import time import unittest from util import hook, timesince def db_init(db): "check to see that our db has the the seen table and return a connection." db.execute( "create table if not exists seen(name, time, quote, chan, " "primary key(name, chan))" ) db.commit() @hook.singlethread @hook.event("PRIVMSG", ignorebots=False) def seeninput(paraml, input=None, db=None, bot=None): db_init(db) db.execute( "insert or replace into seen(name, time, quote, chan)" "values(?,?,?,?)", (input.nick.lower(), time.time(), input.msg, input.chan), ) db.commit() @hook.command def seen(inp, nick="", chan="", db=None, input=None): ".seen <nick> -- Tell when a nickname was last in active in irc" inp = inp.lower() if input.conn.nick.lower() == inp: # user is looking for us, being a smartass return "You need to get your eyes checked." if inp == nick.lower(): return "Have you looked in a mirror lately?" db_init(db) last_seen = db.execute( "select name, time, quote from seen where" " name = ? and chan = ?", (inp, chan) ).fetchone() if last_seen: reltime = timesince.timesince(last_seen[1]) if last_seen[2][0:1] == "\x01": return "%s was last seen %s ago: *%s %s*" % ( inp, reltime, inp, last_seen[2][8:-1], ) else: return "%s was last seen %s ago saying: %s" % (inp, reltime, last_seen[2]) else: return "I've never seen %s" % inp class SeenTest(unittest.TestCase): class Mock(object): def __init__(self, **kwargs): self.__dict__.update(kwargs) def setUp(self): import sqlite3 self.db = sqlite3.connect(":memory:") def seeninput(self, nick, msg, chan="#test"): seeninput(None, db=self.db, input=self.Mock(nick=nick, msg=msg, chan=chan)) def seen(self, inp, nick="bob", chan="#test", bot_nick="skybot"): return seen( inp, nick=nick, chan=chan, db=self.db, input=self.Mock(conn=self.Mock(nick=bot_nick)), ) def test_missing(self): assert "I've never seen nemo" in self.seen("NEMO") def test_seen(self): self.seeninput("nemo", "witty banter") assert "nemo was last seen" in self.seen("nemo") assert "witty banter" in self.seen("nemo") def test_seen_missing_channel(self): self.seeninput("nemo", "msg", chan="#secret") assert "never seen" in self.seen("nemo") def test_seen_ctcp(self): self.seeninput("nemo", "\x01ACTION test lol\x01") assert self.seen("nemo").endswith("ago: *nemo test lol*") def test_snark_eyes(self): assert "eyes checked" in self.seen("skybot", bot_nick="skybot") def test_snark_mirror(self): assert "mirror" in self.seen("bob", nick="bob")
unlicense
rmmh/skybot
test/plugins/test_choose.py
3
1458
from unittest import TestCase from mock import patch from choose import choose class TestChoose(TestCase): def test_choose_one_choice(self): expected = "the decision is up to you" actual = choose("foo") assert expected == actual def test_choose_same_thing(self): expected = "foo" actual = choose("foo, foo, foo") assert expected == actual def test_choose_two_choices(self): actual = choose("foo, bar") assert actual in ["foo", "bar"] def test_choose_choices_space(self): expected_values = ["foo", "bar"] actual = choose("foo bar") assert actual in expected_values def test_choose_strips_whitespace(self): expected_values = ["foo", "bar"] actual = choose(" foo ," " bar ") assert actual in expected_values @patch("random.choice") def test_choose_end_comma_behavior(self, mock_random_choice): mock_random_choice.side_effect = lambda arr: arr[0] expected = "the decision is up to you" actual = choose("foo,") assert actual == expected @patch("random.choice") def test_choose_collapse_commas(self, mock_random_choice): # Should never be an empty string here mock_random_choice.side_effect = lambda arr: arr[1] expected = "bar" actual = choose("foo,,bar") assert actual == expected
unlicense
rmmh/skybot
core/irc.py
3
10652
from __future__ import print_function from builtins import map from builtins import object import re import socket import time import _thread import queue from ssl import wrap_socket, CERT_NONE, CERT_REQUIRED, SSLError DEFAULT_NAME = "skybot" DEFAULT_REALNAME = "Python bot - http://github.com/rmmh/skybot" DEFAULT_NICKSERV_NAME = "nickserv" DEFAULT_NICKSERV_COMMAND = "IDENTIFY %s" def decode(txt): for codec in ("utf-8", "iso-8859-1", "shift_jis", "cp1252"): try: return txt.decode(codec) except UnicodeDecodeError: continue return txt.decode("utf-8", "ignore") def censor(text, censored_strings=None): text = re.sub("[\n\r]+", " ", text) if not censored_strings: return text words = map(re.escape, censored_strings) pattern = "(%s)" % "|".join(words) text = re.sub(pattern, "[censored]", text) return text class crlf_tcp(object): "Handles tcp connections that consist of utf-8 lines ending with crlf" def __init__(self, host, port, timeout=300): self.ibuffer = b"" self.obuffer = b"" self.oqueue = queue.Queue() # lines to be sent out self.iqueue = queue.Queue() # lines that were received self.socket = self.create_socket() self.host = host self.port = port self.timeout = timeout def create_socket(self): return socket.socket(socket.AF_INET, socket.TCP_NODELAY) def run(self): while True: try: self.socket.connect((self.host, self.port)) except socket.timeout: print("timed out connecting to %s:%s" % (self.host, self.port)) time.sleep(60) else: break _thread.start_new_thread(self.recv_loop, ()) _thread.start_new_thread(self.send_loop, ()) def recv_from_socket(self, nbytes): return self.socket.recv(nbytes) def get_timeout_exception_type(self): return socket.timeout def handle_receive_exception(self, error, last_timestamp): if time.time() - last_timestamp > self.timeout: self.iqueue.put(StopIteration) self.socket.close() return True return False def recv_loop(self): last_timestamp = time.time() while True: try: data = self.recv_from_socket(4096) self.ibuffer += data if data: last_timestamp = time.time() else: if time.time() - last_timestamp > self.timeout: self.iqueue.put(StopIteration) self.socket.close() return time.sleep(1) except (self.get_timeout_exception_type(), socket.error) as e: if self.handle_receive_exception(e, last_timestamp): return continue while b"\r\n" in self.ibuffer: line, self.ibuffer = self.ibuffer.split(b"\r\n", 1) self.iqueue.put(decode(line)) def send_loop(self): while True: line = self.oqueue.get().splitlines()[0][:500] print(">>> %s" % line) self.obuffer += line.encode("utf-8", "replace") + b"\r\n" while self.obuffer: sent = self.socket.send(self.obuffer) self.obuffer = self.obuffer[sent:] class crlf_ssl_tcp(crlf_tcp): "Handles ssl tcp connetions that consist of utf-8 lines ending with crlf" def __init__(self, host, port, ignore_cert_errors, timeout=300): self.ignore_cert_errors = ignore_cert_errors crlf_tcp.__init__(self, host, port, timeout) def create_socket(self): return wrap_socket( crlf_tcp.create_socket(self), server_side=False, cert_reqs=CERT_NONE if self.ignore_cert_errors else CERT_REQUIRED, ) def recv_from_socket(self, nbytes): return self.socket.read(nbytes) def get_timeout_exception_type(self): return SSLError def handle_receive_exception(self, error, last_timestamp): return crlf_tcp.handle_receive_exception(self, error, last_timestamp) def zip_channels(channels): channels.sort(key=lambda x: " " not in x) # keyed channels first chans = [] keys = [] for channel in channels: if " " in channel: chan, key = channel.split(" ") chans.append(chan) keys.append(key) else: chans.append(channel) chans = ",".join(chans) if keys: return [chans, ",".join(keys)] else: return [chans] def test_zip_channels(): assert zip_channels(["#a", "#b c", "#d"]) == ["#b,#a,#d", "c"] assert zip_channels(["#a", "#b"]) == ["#a,#b"] class IRC(object): IRC_PREFIX_REM = re.compile(r"(.*?) (.*?) (.*)").match IRC_NOPROFEIX_REM = re.compile(r"()(.*?) (.*)").match IRC_NETMASK_REM = re.compile(r":?([^!@]*)!?([^@]*)@?(.*)").match IRC_PARAM_REF = re.compile(r"(?:^|(?<= ))(:.*|[^ ]+)").findall "handles the IRC protocol" # see the docs/ folder for more information on the protocol def __init__(self, conf): self.conn = None self.nick = DEFAULT_NAME self.user = DEFAULT_NAME self.realname = DEFAULT_REALNAME self.user_mode = None self.server_host = None self.server_port = 6667 self.server_password = None self.nickserv_password = None self.nickserv_name = DEFAULT_NICKSERV_NAME self.nickserv_command = DEFAULT_NICKSERV_COMMAND self.channels = [] self.admins = [] self.censored_strings = [] self.out = queue.Queue() # responses from the server are placed here # format: [rawline, prefix, command, params, # nick, user, host, paramlist, msg] self.set_conf(conf) self.connect() _thread.start_new_thread(self.parse_loop, ()) def set_conf(self, conf): self.nick = conf.get("nick", DEFAULT_NAME) self.user = conf.get("user", DEFAULT_NAME) self.realname = conf.get("realname", DEFAULT_REALNAME) self.user_mode = conf.get("mode", None) self.server_host = conf["server"] self.server_port = conf.get("port", 6667) self.server_password = conf.get("server_password", None) self.nickserv_password = conf.get("nickserv_password", None) self.nickserv_name = conf.get("nickserv_name", DEFAULT_NICKSERV_NAME) self.nickserv_command = conf.get("nickserv_command", DEFAULT_NICKSERV_COMMAND) self.channels = conf.get("channels", []) self.admins = conf.get("admins", []) self.censored_strings = conf.get("censored_strings", []) if self.conn is not None: self.join_channels() def create_connection(self): return crlf_tcp(self.server_host, self.server_port) def connect(self): self.conn = self.create_connection() _thread.start_new_thread(self.conn.run, ()) self.cmd("NICK", [self.nick]) self.cmd("USER", [self.user, "3", "*", self.realname]) if self.server_password: self.cmd("PASS", [self.server_password]) def parse_loop(self): while True: msg = self.conn.iqueue.get() if msg == StopIteration: self.connect() continue if msg.startswith(":"): # has a prefix prefix, command, params = self.IRC_PREFIX_REM(msg).groups() else: prefix, command, params = self.IRC_NOPROFEIX_REM(msg).groups() nick, user, host = self.IRC_NETMASK_REM(prefix).groups() paramlist = self.IRC_PARAM_REF(params) lastparam = "" if paramlist: if paramlist[-1].startswith(":"): paramlist[-1] = paramlist[-1][1:] lastparam = paramlist[-1] self.out.put( [msg, prefix, command, params, nick, user, host, paramlist, lastparam] ) if command == "PING": self.cmd("PONG", paramlist) def join(self, channel): self.cmd("JOIN", channel.split(" ")) # [chan, password] def join_channels(self): if self.channels: # TODO: send multiple join commands for large channel lists self.cmd("JOIN", zip_channels(self.channels)) def msg(self, target, text): self.cmd("PRIVMSG", [target, text]) def cmd(self, command, params=None): if params: params[-1] = ":" + params[-1] params = [censor(p, self.censored_strings) for p in params] self.send(command + " " + " ".join(params)) else: self.send(command) def send(self, str): self.conn.oqueue.put(str) class FakeIRC(IRC): def __init__(self, conf): self.set_conf(conf) self.out = queue.Queue() # responses from the server are placed here self.f = open(fn, "rb") _thread.start_new_thread(self.parse_loop, ()) def parse_loop(self): while True: msg = decode(self.f.readline()[9:]) if msg == "": print("!!!!DONE READING FILE!!!!") return if msg.startswith(":"): # has a prefix prefix, command, params = irc_prefix_rem(msg).groups() else: prefix, command, params = irc_noprefix_rem(msg).groups() nick, user, host = irc_netmask_rem(prefix).groups() paramlist = irc_param_ref(params) lastparam = "" if paramlist: if paramlist[-1].startswith(":"): paramlist[-1] = paramlist[-1][1:] lastparam = paramlist[-1] self.out.put( [msg, prefix, command, params, nick, user, host, paramlist, lastparam] ) if command == "PING": self.cmd("PONG", [params]) def cmd(self, command, params=None): pass class SSLIRC(IRC): def __init__(self, conf): super(SSLIRC, self).__init__(conf=conf) self.server_port = 6697 self.server_ignore_cert = False def set_conf(self, conf): super(SSLIRC, self).set_conf(conf) self.server_port = conf.get("port", 6697) self.server_ignore_cert = conf.get("ignore_cert", False) def create_connection(self): return crlf_ssl_tcp(self.server_host, self.server_port, self.server_ignore_cert)
unlicense
rmmh/skybot
plugins/util/http.py
3
5942
from future.standard_library import hooks from lxml import etree, html import binascii import collections import hmac import json import random import time from hashlib import sha1 from builtins import str from builtins import range try: from http.cookiejar import CookieJar except: from future.backports.http.cookiejar import CookieJar with hooks(): import urllib.request, urllib.parse, urllib.error from urllib.parse import ( quote, unquote, urlencode, urlparse, parse_qsl, quote_plus as _quote_plus, ) from urllib.error import HTTPError, URLError ua_skybot = "Skybot/1.0 https://github.com/rmmh/skybot" ua_firefox = ( "Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.8.1.6) " "Gecko/20070725 Firefox/2.0.0.6" ) ua_internetexplorer = "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)" def get_cookie_jar(): if not hasattr(get_cookie_jar, "memo"): get_cookie_jar.memo = CookieJar() return get_cookie_jar.memo def clear_expired_cookies(): get_cookie_jar().clear_expired_cookies() def get(*args, **kwargs): return open(*args, **kwargs).read().decode("utf-8") def get_html(*args, **kwargs): return html.fromstring(open(*args, **kwargs).read()) def get_xml(*args, **kwargs): return etree.fromstring(open(*args, **kwargs).read()) def get_json(*args, **kwargs): return json.loads(open(*args, **kwargs).read()) def open( url, query_params=None, post_data=None, json_data=None, get_method=None, cookies=False, oauth=False, oauth_keys=None, headers=None, **kwargs ): if query_params is None: query_params = {} query_params.update(kwargs) url = prepare_url(url, query_params) if post_data and isinstance(post_data, collections.Mapping): post_data = urllib.parse.urlencode(post_data) post_data = post_data.encode("UTF-8") if json_data and isinstance(json_data, dict): post_data = json.dumps(json_data).encode("utf-8") request = urllib.request.Request(url, post_data) if json_data: request.add_header("Content-Type", "application/json") if get_method is not None: request.get_method = lambda: get_method if headers is not None: for header_key, header_value in headers.items(): request.add_header(header_key, header_value) if "User-Agent" not in request.headers: request.add_header("User-Agent", ua_skybot) if oauth: nonce = oauth_nonce() timestamp = oauth_timestamp() api_url, req_data = url.split("?") unsigned_request = oauth_unsigned_request( nonce, timestamp, req_data, oauth_keys["consumer"], oauth_keys["access"] ) signature = oauth_sign_request( "GET", api_url, req_data, unsigned_request, oauth_keys["consumer_secret"], oauth_keys["access_secret"], ) header = oauth_build_header( nonce, signature, timestamp, oauth_keys["consumer"], oauth_keys["access"] ) request.add_header("Authorization", header) if cookies: opener = urllib.request.build_opener( urllib.request.HTTPCookieProcessor(get_cookie_jar()) ) else: opener = urllib.request.build_opener() return opener.open(request) def prepare_url(url, queries): if queries: scheme, netloc, path, query, fragment = urllib.parse.urlsplit(str(url)) query = dict(urllib.parse.parse_qsl(query)) query.update(queries) query = urllib.parse.urlencode( dict((to_utf8(key), to_utf8(value)) for key, value in query.items()) ) url = urllib.parse.urlunsplit((scheme, netloc, path, query, fragment)) return url def to_utf8(s): if isinstance(s, str): return s.encode("utf8", "ignore") else: return str(s) def quote_plus(s): return _quote_plus(to_utf8(s)) def oauth_nonce(): return "".join([str(random.randint(0, 9)) for i in range(8)]) def oauth_timestamp(): return str(int(time.time())) def oauth_unsigned_request(nonce, timestamp, req, consumer, token): d = { "oauth_consumer_key": consumer, "oauth_nonce": nonce, "oauth_signature_method": "HMAC-SHA1", "oauth_timestamp": timestamp, "oauth_token": token, "oauth_version": "1.0", } d.update(urllib.parse.parse_qsl(req)) request_items = d.items() # TODO: Remove this when Python 2 is no longer supported. # some of the fields are actual string and others are # a wrapper of str for the python 3 migration. # Convert them all so that they sort correctly. request_items = [(str(k), str(v)) for k, v in request_items] return quote(urllib.parse.urlencode(sorted(request_items, key=lambda key: key[0]))) def oauth_build_header(nonce, signature, timestamp, consumer, token): d = { "oauth_consumer_key": consumer, "oauth_nonce": nonce, "oauth_signature": signature, "oauth_signature_method": "HMAC-SHA1", "oauth_timestamp": timestamp, "oauth_token": token, "oauth_version": "1.0", } header = "OAuth " for x in sorted(d, key=lambda key: key[0]): header += x + '="' + d[x] + '", ' return header[:-1] def oauth_sign_request( method, url, params, unsigned_request, consumer_secret, token_secret ): key = consumer_secret + "&" + token_secret key = key.encode("utf-8", "replace") base = method + "&" + quote(url, "") + "&" + unsigned_request base = base.encode("utf-8", "replace") hash = hmac.new(key, base, sha1) signature = quote(binascii.b2a_base64(hash.digest())[:-1]) return signature def unescape(s): if not s.strip(): return s return html.fromstring(s).text_content()
unlicense
rmmh/skybot
plugins/crowdcontrol.py
3
1132
# crowdcontrol.py by craisins in 2014 # Bot must have some sort of op or admin privileges to be useful import re import time from util import hook # Use "crowdcontrol" array in config # syntax # rule: # re: RegEx. regular expression to match # msg: String. message to display either with kick or as a warning # kick: Integer. 1 for True, 0 for False on if to kick user # ban_length: Integer. (optional) Length of time (seconds) to ban user. (-1 to never unban, 0 to not ban, > 1 for time) @hook.regex(r".*") def crowdcontrol(inp, kick=None, ban=None, unban=None, reply=None, bot=None): inp = inp.group(0) for rule in bot.config.get("crowdcontrol", []): if re.search(rule["re"], inp) is not None: should_kick = rule.get("kick", 0) ban_length = rule.get("ban_length", 0) reason = rule.get("msg") if ban_length != 0: ban() if should_kick: kick(reason=reason) elif "msg" in rule: reply(reason) if ban_length > 0: time.sleep(ban_length) unban()
unlicense
pytube/pytube
pytube/query.py
1
12622
"""This module provides a query interface for media streams and captions.""" from collections.abc import Mapping, Sequence from typing import Callable, List, Optional, Union from pytube import Caption, Stream from pytube.helpers import deprecated class StreamQuery(Sequence): """Interface for querying the available media streams.""" def __init__(self, fmt_streams): """Construct a :class:`StreamQuery <StreamQuery>`. param list fmt_streams: list of :class:`Stream <Stream>` instances. """ self.fmt_streams = fmt_streams self.itag_index = {int(s.itag): s for s in fmt_streams} def filter( self, fps=None, res=None, resolution=None, mime_type=None, type=None, subtype=None, file_extension=None, abr=None, bitrate=None, video_codec=None, audio_codec=None, only_audio=None, only_video=None, progressive=None, adaptive=None, is_dash=None, custom_filter_functions=None, ): """Apply the given filtering criterion. :param fps: (optional) The frames per second. :type fps: int or None :param resolution: (optional) Alias to ``res``. :type res: str or None :param res: (optional) The video resolution. :type resolution: str or None :param mime_type: (optional) Two-part identifier for file formats and format contents composed of a "type", a "subtype". :type mime_type: str or None :param type: (optional) Type part of the ``mime_type`` (e.g.: audio, video). :type type: str or None :param subtype: (optional) Sub-type part of the ``mime_type`` (e.g.: mp4, mov). :type subtype: str or None :param file_extension: (optional) Alias to ``sub_type``. :type file_extension: str or None :param abr: (optional) Average bitrate (ABR) refers to the average amount of data transferred per unit of time (e.g.: 64kbps, 192kbps). :type abr: str or None :param bitrate: (optional) Alias to ``abr``. :type bitrate: str or None :param video_codec: (optional) Video compression format. :type video_codec: str or None :param audio_codec: (optional) Audio compression format. :type audio_codec: str or None :param bool progressive: Excludes adaptive streams (one file contains both audio and video tracks). :param bool adaptive: Excludes progressive streams (audio and video are on separate tracks). :param bool is_dash: Include/exclude dash streams. :param bool only_audio: Excludes streams with video tracks. :param bool only_video: Excludes streams with audio tracks. :param custom_filter_functions: (optional) Interface for defining complex filters without subclassing. :type custom_filter_functions: list or None """ filters = [] if res or resolution: filters.append(lambda s: s.resolution == (res or resolution)) if fps: filters.append(lambda s: s.fps == fps) if mime_type: filters.append(lambda s: s.mime_type == mime_type) if type: filters.append(lambda s: s.type == type) if subtype or file_extension: filters.append(lambda s: s.subtype == (subtype or file_extension)) if abr or bitrate: filters.append(lambda s: s.abr == (abr or bitrate)) if video_codec: filters.append(lambda s: s.video_codec == video_codec) if audio_codec: filters.append(lambda s: s.audio_codec == audio_codec) if only_audio: filters.append( lambda s: ( s.includes_audio_track and not s.includes_video_track ), ) if only_video: filters.append( lambda s: ( s.includes_video_track and not s.includes_audio_track ), ) if progressive: filters.append(lambda s: s.is_progressive) if adaptive: filters.append(lambda s: s.is_adaptive) if custom_filter_functions: filters.extend(custom_filter_functions) if is_dash is not None: filters.append(lambda s: s.is_dash == is_dash) return self._filter(filters) def _filter(self, filters: List[Callable]) -> "StreamQuery": fmt_streams = self.fmt_streams for filter_lambda in filters: fmt_streams = filter(filter_lambda, fmt_streams) return StreamQuery(list(fmt_streams)) def order_by(self, attribute_name: str) -> "StreamQuery": """Apply a sort order. Filters out stream the do not have the attribute. :param str attribute_name: The name of the attribute to sort by. """ has_attribute = [ s for s in self.fmt_streams if getattr(s, attribute_name) is not None ] # Check that the attributes have string values. if has_attribute and isinstance( getattr(has_attribute[0], attribute_name), str ): # Try to return a StreamQuery sorted by the integer representations # of the values. try: return StreamQuery( sorted( has_attribute, key=lambda s: int( "".join( filter(str.isdigit, getattr(s, attribute_name)) ) ), # type: ignore # noqa: E501 ) ) except ValueError: pass return StreamQuery( sorted(has_attribute, key=lambda s: getattr(s, attribute_name)) ) def desc(self) -> "StreamQuery": """Sort streams in descending order. :rtype: :class:`StreamQuery <StreamQuery>` """ return StreamQuery(self.fmt_streams[::-1]) def asc(self) -> "StreamQuery": """Sort streams in ascending order. :rtype: :class:`StreamQuery <StreamQuery>` """ return self def get_by_itag(self, itag: int) -> Optional[Stream]: """Get the corresponding :class:`Stream <Stream>` for a given itag. :param int itag: YouTube format identifier code. :rtype: :class:`Stream <Stream>` or None :returns: The :class:`Stream <Stream>` matching the given itag or None if not found. """ return self.itag_index.get(int(itag)) def get_by_resolution(self, resolution: str) -> Optional[Stream]: """Get the corresponding :class:`Stream <Stream>` for a given resolution. Stream must be a progressive mp4. :param str resolution: Video resolution i.e. "720p", "480p", "360p", "240p", "144p" :rtype: :class:`Stream <Stream>` or None :returns: The :class:`Stream <Stream>` matching the given itag or None if not found. """ return self.filter( progressive=True, subtype="mp4", resolution=resolution ).first() def get_lowest_resolution(self) -> Optional[Stream]: """Get lowest resolution stream that is a progressive mp4. :rtype: :class:`Stream <Stream>` or None :returns: The :class:`Stream <Stream>` matching the given itag or None if not found. """ return ( self.filter(progressive=True, subtype="mp4") .order_by("resolution") .first() ) def get_highest_resolution(self) -> Optional[Stream]: """Get highest resolution stream that is a progressive video. :rtype: :class:`Stream <Stream>` or None :returns: The :class:`Stream <Stream>` matching the given itag or None if not found. """ return self.filter(progressive=True).order_by("resolution").last() def get_audio_only(self, subtype: str = "mp4") -> Optional[Stream]: """Get highest bitrate audio stream for given codec (defaults to mp4) :param str subtype: Audio subtype, defaults to mp4 :rtype: :class:`Stream <Stream>` or None :returns: The :class:`Stream <Stream>` matching the given itag or None if not found. """ return ( self.filter(only_audio=True, subtype=subtype) .order_by("abr") .last() ) def otf(self, is_otf: bool = False) -> "StreamQuery": """Filter stream by OTF, useful if some streams have 404 URLs :param bool is_otf: Set to False to retrieve only non-OTF streams :rtype: :class:`StreamQuery <StreamQuery>` :returns: A StreamQuery object with otf filtered streams """ return self._filter([lambda s: s.is_otf == is_otf]) def first(self) -> Optional[Stream]: """Get the first :class:`Stream <Stream>` in the results. :rtype: :class:`Stream <Stream>` or None :returns: the first result of this query or None if the result doesn't contain any streams. """ try: return self.fmt_streams[0] except IndexError: return None def last(self): """Get the last :class:`Stream <Stream>` in the results. :rtype: :class:`Stream <Stream>` or None :returns: Return the last result of this query or None if the result doesn't contain any streams. """ try: return self.fmt_streams[-1] except IndexError: pass @deprecated("Get the size of this list directly using len()") def count(self, value: Optional[str] = None) -> int: # pragma: no cover """Get the count of items in the list. :rtype: int """ if value: return self.fmt_streams.count(value) return len(self) @deprecated("This object can be treated as a list, all() is useless") def all(self) -> List[Stream]: # pragma: no cover """Get all the results represented by this query as a list. :rtype: list """ return self.fmt_streams def __getitem__(self, i: Union[slice, int]): return self.fmt_streams[i] def __len__(self) -> int: return len(self.fmt_streams) def __repr__(self) -> str: return f"{self.fmt_streams}" class CaptionQuery(Mapping): """Interface for querying the available captions.""" def __init__(self, captions: List[Caption]): """Construct a :class:`Caption <Caption>`. param list captions: list of :class:`Caption <Caption>` instances. """ self.lang_code_index = {c.code: c for c in captions} @deprecated( "This object can be treated as a dictionary, i.e. captions['en']" ) def get_by_language_code( self, lang_code: str ) -> Optional[Caption]: # pragma: no cover """Get the :class:`Caption <Caption>` for a given ``lang_code``. :param str lang_code: The code that identifies the caption language. :rtype: :class:`Caption <Caption>` or None :returns: The :class:`Caption <Caption>` matching the given ``lang_code`` or None if it does not exist. """ return self.lang_code_index.get(lang_code) @deprecated("This object can be treated as a dictionary") def all(self) -> List[Caption]: # pragma: no cover """Get all the results represented by this query as a list. :rtype: list """ return list(self.lang_code_index.values()) def __getitem__(self, i: str): return self.lang_code_index[i] def __len__(self) -> int: return len(self.lang_code_index) def __iter__(self): return iter(self.lang_code_index.values()) def __repr__(self) -> str: return f"{self.lang_code_index}"
unlicense
pytube/pytube
pytube/innertube.py
1
11658
"""This module is designed to interact with the innertube API. This module is NOT intended to be used directly by end users, as each of the interfaces returns raw results. These should instead be parsed to extract the useful information for the end user. """ # Native python imports import json import os import pathlib import time from urllib import parse # Local imports from pytube import request # YouTube on TV client secrets _client_id = '861556708454-d6dlm3lh05idd8npek18k6be8ba3oc68.apps.googleusercontent.com' _client_secret = 'SboVhoG9s0rNafixCSGGKXAT' # Extracted API keys -- unclear what these are linked to. _api_keys = [ 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8', 'AIzaSyCtkvNIR1HCEwzsqK6JuE6KqpyjusIRI30', 'AIzaSyA8eiZmM1FaDVjRy-df2KTyQ_vz_yYM39w', 'AIzaSyC8UYZpvA2eknNex0Pjid0_eTLJoDu6los', 'AIzaSyCjc_pVEDi4qsv5MtC2dMXzpIaDoRFLsxw', 'AIzaSyDHQ9ipnphqTzDqZsbtd8_Ru4_kiKVQe2k' ] _default_clients = { 'WEB': { 'context': { 'client': { 'clientName': 'WEB', 'clientVersion': '2.20200720.00.02' } }, 'api_key': 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8' }, 'ANDROID': { 'context': { 'client': { 'clientName': 'ANDROID', 'clientVersion': '16.20' } }, 'api_key': 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8' }, 'WEB_EMBED': { 'context': { 'client': { 'clientName': 'WEB', 'clientVersion': '2.20210721.00.00', 'clientScreen': 'EMBED' } }, 'api_key': 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8' }, 'ANDROID_EMBED': { 'context': { 'client': { 'clientName': 'ANDROID', 'clientVersion': '16.20', 'clientScreen': 'EMBED' } }, 'api_key': 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8' } } _token_timeout = 1800 _cache_dir = pathlib.Path(__file__).parent.resolve() / '__cache__' _token_file = os.path.join(_cache_dir, 'tokens.json') class InnerTube: """Object for interacting with the innertube API.""" def __init__(self, client='ANDROID', use_oauth=False, allow_cache=True): """Initialize an InnerTube object. :param str client: Client to use for the object. Default to web because it returns the most playback types. :param bool use_oauth: Whether or not to authenticate to YouTube. :param bool allow_cache: Allows caching of oauth tokens on the machine. """ self.context = _default_clients[client]['context'] self.api_key = _default_clients[client]['api_key'] self.access_token = None self.refresh_token = None self.use_oauth = use_oauth self.allow_cache = allow_cache # Stored as epoch time self.expires = None # Try to load from file if specified if self.use_oauth and self.allow_cache: # Try to load from file if possible if os.path.exists(_token_file): with open(_token_file) as f: data = json.load(f) self.access_token = data['access_token'] self.refresh_token = data['refresh_token'] self.expires = data['expires'] self.refresh_bearer_token() def cache_tokens(self): """Cache tokens to file if allowed.""" if not self.allow_cache: return data = { 'access_token': self.access_token, 'refresh_token': self.refresh_token, 'expires': self.expires } if not os.path.exists(_cache_dir): os.mkdir(_cache_dir) with open(_token_file, 'w') as f: json.dump(data, f) def refresh_bearer_token(self, force=False): """Refreshes the OAuth token if necessary. :param bool force: Force-refresh the bearer token. """ if not self.use_oauth: return # Skip refresh if it's not necessary and not forced if self.expires > time.time() and not force: return # Subtracting 30 seconds is arbitrary to avoid potential time discrepencies start_time = int(time.time() - 30) data = { 'client_id': _client_id, 'client_secret': _client_secret, 'grant_type': 'refresh_token', 'refresh_token': self.refresh_token } response = request._execute_request( 'https://oauth2.googleapis.com/token', 'POST', headers={ 'Content-Type': 'application/json' }, data=data ) response_data = json.loads(response.read()) self.access_token = response_data['access_token'] self.expires = start_time + response_data['expires_in'] self.cache_tokens() def fetch_bearer_token(self): """Fetch an OAuth token.""" # Subtracting 30 seconds is arbitrary to avoid potential time discrepencies start_time = int(time.time() - 30) data = { 'client_id': _client_id, 'scope': 'https://www.googleapis.com/auth/youtube' } response = request._execute_request( 'https://oauth2.googleapis.com/device/code', 'POST', headers={ 'Content-Type': 'application/json' }, data=data ) response_data = json.loads(response.read()) verification_url = response_data['verification_url'] user_code = response_data['user_code'] print(f'Please open {verification_url} and input code {user_code}') input('Press enter when you have completed this step.') data = { 'client_id': _client_id, 'client_secret': _client_secret, 'device_code': response_data['device_code'], 'grant_type': 'urn:ietf:params:oauth:grant-type:device_code' } response = request._execute_request( 'https://oauth2.googleapis.com/token', 'POST', headers={ 'Content-Type': 'application/json' }, data=data ) response_data = json.loads(response.read()) self.access_token = response_data['access_token'] self.refresh_token = response_data['refresh_token'] self.expires = start_time + response_data['expires_in'] self.cache_tokens() @property def base_url(self): """Return the base url endpoint for the innertube API.""" return 'https://www.youtube.com/youtubei/v1' @property def base_data(self): """Return the base json data to transmit to the innertube API.""" return { 'context': self.context } @property def base_params(self): """Return the base query parameters to transmit to the innertube API.""" return { 'key': self.api_key, 'contentCheckOk': True, 'racyCheckOk': True } def _call_api(self, endpoint, query, data): """Make a request to a given endpoint with the provided query parameters and data.""" # Remove the API key if oauth is being used. if self.use_oauth: del query['key'] endpoint_url = f'{endpoint}?{parse.urlencode(query)}' headers = { 'Content-Type': 'application/json', } # Add the bearer token if applicable if self.use_oauth: if self.access_token: self.refresh_bearer_token() headers['Authorization'] = f'Bearer {self.access_token}' else: self.fetch_bearer_token() headers['Authorization'] = f'Bearer {self.access_token}' response = request._execute_request( endpoint_url, 'POST', headers=headers, data=data ) return json.loads(response.read()) def browse(self): """Make a request to the browse endpoint. TODO: Figure out how we can use this """ # endpoint = f'{self.base_url}/browse' # noqa:E800 ... # return self._call_api(endpoint, query, self.base_data) # noqa:E800 def config(self): """Make a request to the config endpoint. TODO: Figure out how we can use this """ # endpoint = f'{self.base_url}/config' # noqa:E800 ... # return self._call_api(endpoint, query, self.base_data) # noqa:E800 def guide(self): """Make a request to the guide endpoint. TODO: Figure out how we can use this """ # endpoint = f'{self.base_url}/guide' # noqa:E800 ... # return self._call_api(endpoint, query, self.base_data) # noqa:E800 def next(self): """Make a request to the next endpoint. TODO: Figure out how we can use this """ # endpoint = f'{self.base_url}/next' # noqa:E800 ... # return self._call_api(endpoint, query, self.base_data) # noqa:E800 def player(self, video_id): """Make a request to the player endpoint. :param str video_id: The video id to get player info for. :rtype: dict :returns: Raw player info results. """ endpoint = f'{self.base_url}/player' query = { 'videoId': video_id, } query.update(self.base_params) return self._call_api(endpoint, query, self.base_data) def search(self, search_query, continuation=None): """Make a request to the search endpoint. :param str search_query: The query to search. :rtype: dict :returns: Raw search query results. """ endpoint = f'{self.base_url}/search' query = { 'query': search_query } query.update(self.base_params) data = {} if continuation: data['continuation'] = continuation data.update(self.base_data) return self._call_api(endpoint, query, data) def verify_age(self, video_id): """Make a request to the age_verify endpoint. Notable examples of the types of video this verification step is for: * https://www.youtube.com/watch?v=QLdAhwSBZ3w * https://www.youtube.com/watch?v=hc0ZDaAZQT0 :param str video_id: The video id to get player info for. :rtype: dict :returns: Returns information that includes a URL for bypassing certain restrictions. """ endpoint = f'{self.base_url}/verify_age' data = { 'nextEndpoint': { 'urlEndpoint': { 'url': f'/watch?v={video_id}' } }, 'setControvercy': True } data.update(self.base_data) result = self._call_api(endpoint, self.base_params, data) return result def get_transcript(self, video_id): """Make a request to the get_transcript endpoint. This is likely related to captioning for videos, but is currently untested. """ endpoint = f'{self.base_url}/get_transcript' query = { 'videoId': video_id, } query.update(self.base_params) result = self._call_api(endpoint, query, self.base_data) return result
unlicense
pytube/pytube
tests/test_request.py
1
1820
import socket import os import pytest from unittest import mock from urllib.error import URLError from pytube import request from pytube.exceptions import MaxRetriesExceeded @mock.patch("pytube.request.urlopen") def test_streaming(mock_urlopen): # Given fake_stream_binary = [ os.urandom(8 * 1024), os.urandom(8 * 1024), os.urandom(8 * 1024), None, ] mock_response = mock.Mock() mock_response.read.side_effect = fake_stream_binary mock_response.info.return_value = {"Content-Range": "bytes 200-1000/24576"} mock_urlopen.return_value = mock_response # When response = request.stream("http://fakeassurl.gov/streaming_test") # Then assert len(b''.join(response)) == 3 * 8 * 1024 assert mock_response.read.call_count == 4 @mock.patch('pytube.request.urlopen') def test_timeout(mock_urlopen): exc = URLError(reason=socket.timeout('timed_out')) mock_urlopen.side_effect = exc generator = request.stream('http://fakeassurl.gov/timeout_test', timeout=1) with pytest.raises(MaxRetriesExceeded): next(generator) @mock.patch("pytube.request.urlopen") def test_headers(mock_urlopen): response = mock.Mock() response.info.return_value = {"content-length": "16384"} mock_urlopen.return_value = response response = request.head("http://fakeassurl.gov") assert response == {"content-length": "16384"} @mock.patch("pytube.request.urlopen") def test_get(mock_urlopen): response = mock.Mock() response.read.return_value = "<html></html>".encode("utf-8") mock_urlopen.return_value = response response = request.get("http://fakeassurl.gov") assert response == "<html></html>" def test_get_non_http(): with pytest.raises(ValueError): # noqa: PT011 request.get("file://bad")
unlicense
pytube/pytube
pytube/contrib/playlist.py
1
14204
"""Module to download a complete playlist from a youtube channel.""" import json import logging from collections.abc import Sequence from datetime import date, datetime from typing import Dict, Iterable, List, Optional, Tuple, Union from pytube import extract, request, YouTube from pytube.helpers import cache, DeferredGeneratorList, install_proxy, uniqueify logger = logging.getLogger(__name__) class Playlist(Sequence): """Load a YouTube playlist with URL""" def __init__(self, url: str, proxies: Optional[Dict[str, str]] = None): if proxies: install_proxy(proxies) self._input_url = url # These need to be initialized as None for the properties. self._html = None self._ytcfg = None self._initial_data = None self._sidebar_info = None self._playlist_id = None @property def playlist_id(self): """Get the playlist id. :rtype: str """ if self._playlist_id: return self._playlist_id self._playlist_id = extract.playlist_id(self._input_url) return self._playlist_id @property def playlist_url(self): """Get the base playlist url. :rtype: str """ return f"https://www.youtube.com/playlist?list={self.playlist_id}" @property def html(self): """Get the playlist page html. :rtype: str """ if self._html: return self._html self._html = request.get(self.playlist_url) return self._html @property def ytcfg(self): """Extract the ytcfg from the playlist page html. :rtype: dict """ if self._ytcfg: return self._ytcfg self._ytcfg = extract.get_ytcfg(self.html) return self._ytcfg @property def initial_data(self): """Extract the initial data from the playlist page html. :rtype: dict """ if self._initial_data: return self._initial_data else: self._initial_data = extract.initial_data(self.html) return self._initial_data @property def sidebar_info(self): """Extract the sidebar info from the playlist page html. :rtype: dict """ if self._sidebar_info: return self._sidebar_info else: self._sidebar_info = self.initial_data['sidebar'][ 'playlistSidebarRenderer']['items'] return self._sidebar_info @property def yt_api_key(self): """Extract the INNERTUBE_API_KEY from the playlist ytcfg. :rtype: str """ return self.ytcfg['INNERTUBE_API_KEY'] def _paginate( self, until_watch_id: Optional[str] = None ) -> Iterable[List[str]]: """Parse the video links from the page source, yields the /watch?v= part from video link :param until_watch_id Optional[str]: YouTube Video watch id until which the playlist should be read. :rtype: Iterable[List[str]] :returns: Iterable of lists of YouTube watch ids """ videos_urls, continuation = self._extract_videos( json.dumps(extract.initial_data(self.html)) ) if until_watch_id: try: trim_index = videos_urls.index(f"/watch?v={until_watch_id}") yield videos_urls[:trim_index] return except ValueError: pass yield videos_urls # Extraction from a playlist only returns 100 videos at a time # if self._extract_videos returns a continuation there are more # than 100 songs inside a playlist, so we need to add further requests # to gather all of them if continuation: load_more_url, headers, data = self._build_continuation_url(continuation) else: load_more_url, headers, data = None, None, None while load_more_url and headers and data: # there is an url found logger.debug("load more url: %s", load_more_url) # requesting the next page of videos with the url generated from the # previous page, needs to be a post req = request.post(load_more_url, extra_headers=headers, data=data) # extract up to 100 songs from the page loaded # returns another continuation if more videos are available videos_urls, continuation = self._extract_videos(req) if until_watch_id: try: trim_index = videos_urls.index(f"/watch?v={until_watch_id}") yield videos_urls[:trim_index] return except ValueError: pass yield videos_urls if continuation: load_more_url, headers, data = self._build_continuation_url( continuation ) else: load_more_url, headers, data = None, None, None def _build_continuation_url(self, continuation: str) -> Tuple[str, dict, dict]: """Helper method to build the url and headers required to request the next page of videos :param str continuation: Continuation extracted from the json response of the last page :rtype: Tuple[str, dict, dict] :returns: Tuple of an url and required headers for the next http request """ return ( ( # was changed to this format (and post requests) # between 2021.03.02 and 2021.03.03 "https://www.youtube.com/youtubei/v1/browse?key=" f"{self.yt_api_key}" ), { "X-YouTube-Client-Name": "1", "X-YouTube-Client-Version": "2.20200720.00.02", }, # extra data required for post request { "continuation": continuation, "context": { "client": { "clientName": "WEB", "clientVersion": "2.20200720.00.02" } } } ) @staticmethod def _extract_videos(raw_json: str) -> Tuple[List[str], Optional[str]]: """Extracts videos from a raw json page :param str raw_json: Input json extracted from the page or the last server response :rtype: Tuple[List[str], Optional[str]] :returns: Tuple containing a list of up to 100 video watch ids and a continuation token, if more videos are available """ initial_data = json.loads(raw_json) try: # this is the json tree structure, if the json was extracted from # html section_contents = initial_data["contents"][ "twoColumnBrowseResultsRenderer"][ "tabs"][0]["tabRenderer"]["content"][ "sectionListRenderer"]["contents"] try: # Playlist without submenus important_content = section_contents[ 0]["itemSectionRenderer"][ "contents"][0]["playlistVideoListRenderer"] except (KeyError, IndexError, TypeError): # Playlist with submenus important_content = section_contents[ 1]["itemSectionRenderer"][ "contents"][0]["playlistVideoListRenderer"] videos = important_content["contents"] except (KeyError, IndexError, TypeError): try: # this is the json tree structure, if the json was directly sent # by the server in a continuation response # no longer a list and no longer has the "response" key important_content = initial_data['onResponseReceivedActions'][0][ 'appendContinuationItemsAction']['continuationItems'] videos = important_content except (KeyError, IndexError, TypeError) as p: logger.info(p) return [], None try: continuation = videos[-1]['continuationItemRenderer'][ 'continuationEndpoint' ]['continuationCommand']['token'] videos = videos[:-1] except (KeyError, IndexError): # if there is an error, no continuation is available continuation = None # remove duplicates return ( uniqueify( list( # only extract the video ids from the video data map( lambda x: ( f"/watch?v=" f"{x['playlistVideoRenderer']['videoId']}" ), videos ) ), ), continuation, ) def trimmed(self, video_id: str) -> Iterable[str]: """Retrieve a list of YouTube video URLs trimmed at the given video ID i.e. if the playlist has video IDs 1,2,3,4 calling trimmed(3) returns [1,2] :type video_id: str video ID to trim the returned list of playlist URLs at :rtype: List[str] :returns: List of video URLs from the playlist trimmed at the given ID """ for page in self._paginate(until_watch_id=video_id): yield from (self._video_url(watch_path) for watch_path in page) def url_generator(self): """Generator that yields video URLs. :Yields: Video URLs """ for page in self._paginate(): for video in page: yield self._video_url(video) @property # type: ignore @cache def video_urls(self) -> DeferredGeneratorList: """Complete links of all the videos in playlist :rtype: List[str] :returns: List of video URLs """ return DeferredGeneratorList(self.url_generator()) def videos_generator(self): for url in self.video_urls: yield YouTube(url) @property def videos(self) -> Iterable[YouTube]: """Yields YouTube objects of videos in this playlist :rtype: List[YouTube] :returns: List of YouTube """ return DeferredGeneratorList(self.videos_generator()) def __getitem__(self, i: Union[slice, int]) -> Union[str, List[str]]: return self.video_urls[i] def __len__(self) -> int: return len(self.video_urls) def __repr__(self) -> str: return f"{repr(self.video_urls)}" @property @cache def last_updated(self) -> Optional[date]: """Extract the date that the playlist was last updated. For some playlists, this will be a specific date, which is returned as a datetime object. For other playlists, this is an estimate such as "1 week ago". Due to the fact that this value is returned as a string, pytube does a best-effort parsing where possible, and returns the raw string where it is not possible. :return: Date of last playlist update where possible, else the string provided :rtype: datetime.date """ last_updated_text = self.sidebar_info[0]['playlistSidebarPrimaryInfoRenderer'][ 'stats'][2]['runs'][1]['text'] try: date_components = last_updated_text.split() month = date_components[0] day = date_components[1].strip(',') year = date_components[2] return datetime.strptime( f"{month} {day:0>2} {year}", "%b %d %Y" ).date() except (IndexError, KeyError): return last_updated_text @property @cache def title(self) -> Optional[str]: """Extract playlist title :return: playlist title (name) :rtype: Optional[str] """ return self.sidebar_info[0]['playlistSidebarPrimaryInfoRenderer'][ 'title']['runs'][0]['text'] @property def description(self) -> str: return self.sidebar_info[0]['playlistSidebarPrimaryInfoRenderer'][ 'description']['simpleText'] @property def length(self): """Extract the number of videos in the playlist. :return: Playlist video count :rtype: int """ count_text = self.sidebar_info[0]['playlistSidebarPrimaryInfoRenderer'][ 'stats'][0]['runs'][0]['text'] count_text = count_text.replace(',','') return int(count_text) @property def views(self): """Extract view count for playlist. :return: Playlist view count :rtype: int """ # "1,234,567 views" views_text = self.sidebar_info[0]['playlistSidebarPrimaryInfoRenderer'][ 'stats'][1]['simpleText'] # "1,234,567" count_text = views_text.split()[0] # "1234567" count_text = count_text.replace(',', '') return int(count_text) @property def owner(self): """Extract the owner of the playlist. :return: Playlist owner name. :rtype: str """ return self.sidebar_info[1]['playlistSidebarSecondaryInfoRenderer'][ 'videoOwner']['videoOwnerRenderer']['title']['runs'][0]['text'] @property def owner_id(self): """Extract the channel_id of the owner of the playlist. :return: Playlist owner's channel ID. :rtype: str """ return self.sidebar_info[1]['playlistSidebarSecondaryInfoRenderer'][ 'videoOwner']['videoOwnerRenderer']['title']['runs'][0][ 'navigationEndpoint']['browseEndpoint']['browseId'] @property def owner_url(self): """Create the channel url of the owner of the playlist. :return: Playlist owner's channel url. :rtype: str """ return f'https://www.youtube.com/channel/{self.owner_id}' @staticmethod def _video_url(watch_path: str): return f"https://www.youtube.com{watch_path}"
unlicense
mozilla-iam/cis
python-modules/cis_profile/tests/test_fake_profile.py
1
2844
import pytest from boto3.dynamodb.types import TypeDeserializer from cis_profile import fake_profile from cis_profile import profile from cis_profile import exceptions class TestFakeProfile(object): def test_fake_user(self): u = fake_profile.FakeUser() print(u.user_id.value) j = u.as_json() d = u.as_dict() assert j is not None assert d is not None assert u.user_id.value is not None u.validate() u.verify_all_publishers(u) def test_same_fake_user(self): a = fake_profile.FakeUser(seed=1337) b = fake_profile.FakeUser(seed=1337) c = fake_profile.FakeUser(seed=23) assert a.uuid.value == b.uuid.value assert a.uuid.value != c.uuid.value def test_batch_create(self): profiles = fake_profile.batch_create_fake_profiles(seed=1337, number=3) assert len(profiles) == 3 for i, p in enumerate(profiles, 1): assert p is not None assert p["access_information"]["hris"]["values"]["employee_id"] == i def test_with_and_without_uuid(self): c_with_uuid = fake_profile.FakeProfileConfig().uuid_username() c_without_uuid = fake_profile.FakeProfileConfig() a = fake_profile.FakeUser(seed=23, config=c_with_uuid) assert a.uuid.value is not None b = fake_profile.FakeUser(seed=23, config=c_without_uuid) assert b.uuid.value is None def test_null_create_profile(self): empty_profile = profile.User() create_profile = fake_profile.FakeUser( seed=1337, config=fake_profile.FakeProfileConfig().default().minimal().create().no_display() ) update_profile = fake_profile.FakeUser( seed=1337, config=fake_profile.FakeProfileConfig().minimal().default().no_display() ) update_profile.uuid = create_profile.uuid update_profile.user_id = create_profile.user_id update_profile.primary_username.value = "test" # invalid because default publisher is cis update_profile.primary_username.signature.publisher.name = "mozilliansorg" with pytest.raises(exceptions.PublisherVerificationFailure): update_profile.verify_all_publishers(empty_profile) assert create_profile.verify_all_publishers(empty_profile) is True assert update_profile.verify_all_publishers(create_profile) is True def test_fake_dynamo_flat_dict_output(self): # This profile must have ldap groups and staff data p = fake_profile.FakeUser(seed=44) ddb = p.as_dynamo_flat_dict() deserializer = TypeDeserializer() res = {k: deserializer.deserialize(v) for k, v in ddb.items()} assert len(res["access_information"]["ldap"]) > 0 assert len(res["staff_information"]["office_location"]) > 0
mpl-2.0
mozilla-iam/cis
python-modules/cis_crypto/cis_crypto/cli.py
1
3035
#!/usr/bin/env python3 import argparse import jose import logging import sys from cis_crypto import common from cis_crypto import operation class cli: def __init__(self): self.config = None self.prog = sys.argv[0].split("/")[-1] def parse_args(self, args): parser = argparse.ArgumentParser( description=""" Command line wrapper for mozilla-iam sign verify/operations of JSON and YAML using JWKS. """ ) subparsers = parser.add_subparsers(dest="cryptographic-operation") subparsers.required = True sign_operation_parser = subparsers.add_parser( "sign", help="Use a jwks key to generate a signature for a file. (Assumes a json or yaml file)" ) sign_operation_parser.add_argument( "--file", help="The path to the file you would like to sign. (Assumes a json or yaml file)" ) sign_operation_parser.set_defaults(func="sign_operation") verify_operation_parser = subparsers.add_parser( "verify", help="Verify a signture with a known file. (Assumes a json file)" ) verify_operation_parser.add_argument("--file", help="The path to the file you would like to sign.") verify_operation_parser.set_defaults(func="verify_operation") return parser.parse_args(args) def run(self): logger = logging.getLogger(__name__) self.config = self.parse_args(sys.argv[1:]) if self.config.func == "sign_operation": logger.info("Attempting to sign file: {}".format(self.config.file)) file_content = common.load_file(self.config.file) signing_object = operation.Sign() signing_object.load(file_content) jws = signing_object.jws() common.write_file(jws, "{}.jws".format(self.config.file)) logger.info("File signed. Your signed file is now: {}.jws".format(self.config.file)) logger.info("To verify this file use cis_crypto verify --file {}.jws".format(self.config.file)) elif self.config.func == "verify_operation": logger.info("Attempting verification of signature for file: {}".format(self.config.file)) everett_config = common.get_config() logger.info( "Attempting fetch of .well-known data from: {}".format( everett_config("public_key_name", namespace="cis", default="access-file-key.pub.pem") ) ) file_content = common.load_file(self.config.file) verify_object = operation.Verify() verify_object.load(file_content) try: jws = verify_object.jws() # This will raise if the signature is invalid. logger.info("Signature verified for file: {}".format(self.config.file)) except jose.exceptions.JWSError: logger.error("The signature could not be verified.") sys.exit() sys.exit()
mpl-2.0
mozilla-iam/cis
python-modules/cis_notifications/cis_notifications/event.py
1
5547
import logging import time import requests from cis_notifications import common from cis_notifications import secret logger = logging.getLogger(__name__) def expired(ts, leeway=0): return ts < time.time() + leeway class Event(object): """Handle events from lambda and generate hooks out to publishers.""" def __init__(self, event): """[summary] Arguments: object {[type]} -- [an instance of the event class.] event {[type]} -- [the event as ingested from the kinesis stream.] subscriptions {[type]} -- [list of urls to post notifications to.] """ self.config = common.get_config() self.event = event self.secret_manager = secret.Manager() self.access_token = None def to_notification(self): """[summary] Transform the instance of the event from the stream into a notification payload. [return] JSON data structure to send using requests. """ logger.debug("An event was received", extra={"event": self.event}) updated_record = self.event.get("dynamodb") operation = "foxy" # Just a place holder in case we have an unhandled event. if self.event.get("eventName") == "INSERT": operation = "create" if self.event.get("eventName") == "MODIFY": operation = "update" if self.event.get("eventName") == "REMOVE": operation = "delete" if updated_record is not None: # Provided the event is the structure that notification = { "operation": operation, "id": updated_record["Keys"]["id"]["S"], "time": updated_record["ApproximateCreationDateTime"], } logger.debug("Notification generated.", extra={"notification": notification}) return notification else: logger.debug("No notification generated.") return {} def send(self, notification): """[summary] Get the list of notification endpoints from the object constructor and send a POST with the json payload. Arguments: object {[type]} -- [an instance of the event class.] object {[notification]} -- [A json payload that you would like to send to the RP.] [return] Dictionary of status codes by publisher. """ # Not in-memory access token? if not self.access_token: # Load whatever is in our secrets self.access_token_dict = self.secret_manager.secretmgr("az_access_token") # Check if what we had in secrets is still valid! # This includes 10s leeway for clock sync issues and 15min (900s) for max-lambda function time. # Since tokens are normally valid for 86400s (1 day) that should accomodate for all cases. If these were to # be less than 15min for any reason, it would simply bypass the cache if expired(float(self.access_token_dict.get("exp", 0.0)), leeway=910): logger.info("Access token has expired, refreshing") authzero = self._get_authzero_client() self.access_token_dict = authzero.exchange_for_access_token() # Auth0 gives us the difference (expires_in) not a time stamp, so we need to calculate when the token # expires. self.access_token_dict["exp"] = time.time() + float(self.access_token_dict.get("expires_in", 60.0)) self.secret_manager.secretmgr_store("az_access_token", self.access_token_dict) else: logger.info("Re-using cached access token") self.access_token = self.access_token_dict["access_token"] if notification != {}: rp_urls = self.config( "rp_urls", namespace="cis", default="https://dinopark.k8s.dev.sso.allizom.org/events/update" ) results = {} for url in rp_urls.split(","): result = self._notify_via_post(url, notification, self.access_token) results[url] = result return results def _get_authzero_client(self): authzero = secret.AuthZero( client_id=self.secret_manager.secret("client_id"), client_secret=self.secret_manager.secret("client_secret"), api_identifier=self.config("api_identifier", namespace="cis", default="hook.dev.sso.allizom.org"), authzero_tenant=self.config("authzero_tenant", namespace="cis", default="auth.mozilla.auth0.com"), ) return authzero def _notify_via_post(self, url, json_payload, access_token): """[summary] Notify a single publisher of the user_id that was updated and return only the status code. Arguments: url {[type]} -- [the url of the publisher you woud like to notify.] json_payload {[type]} -- [the event to send to the publisher.] """ try: response = requests.post( url, json=json_payload, headers={"authorization": "Bearer {}".format(access_token)} ) return response.status_code except requests.exceptions.RequestException: return "Unknown" except requests.exceptions.HTTPError: return "HTTPError" except requests.exceptions.ConnectionError: return "ConnectionError" except requests.exceptions.Timeout: return "Timeout"
mpl-2.0
mozilla-iam/cis
python-modules/cis_logger/setup.py
1
1188
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages with open("README.md", "r") as fh: long_description = fh.read() requirements = ["python-json-logger", "everett", "everett[ini]"] test_requirements = ["pytest", "pytest-watch", "pytest-cov", "flake8", "flask", "flask_graphql", "flask_restful"] setup_requirements = ["pytest-runner", "setuptools>=40.5.0"] extras = {"test": test_requirements} setup( name="cis_logger", version="0.0.1", author="Andrew Krug", author_email="akrug@mozilla.com", description="Mozilla IAM logger wrapper.", long_description=long_description, url="https://github.com/mozilla-iam/cis", classifiers=[ "Programming Language :: Python :: 3", "License :: OSI Approved :: Mozilla Public License 2.0 (MPL 2.0)", "Operating System :: OS Independent", ], install_requires=requirements, license="Mozilla Public License 2.0", include_package_data=True, packages=find_packages(include=["cis_logger"]), setup_requires=setup_requirements, tests_require=test_requirements, extras_require=extras, test_suite="tests", zip_safe=True, )
mpl-2.0
mozilla-iam/cis
python-modules/cis_crypto/cis_crypto/secret.py
1
5055
"""Class for following a default provider chain in the fetching of key material for sign/verify operations.""" import boto3 import json import os import logging import time from cis_crypto import common from jose import jwk from botocore.exceptions import ClientError logger = logging.getLogger(__name__) class Manager(object): """Top level manager object. Will instantiate the appropriate provider based on configuration.""" def __init__(self, provider_type): self.provider_type = provider_type def get_key(self, key_name): provider = self._load_provider() return provider.key(key_name) def _load_provider(self): logger.debug("Using secret manager provider type: {}".format(self.provider_type)) if self.provider_type.lower() == "file": return FileProvider() elif self.provider_type.lower() == "aws-ssm": return AWSParameterstoreProvider() else: raise ValueError("The secret provider selected is not yet supported file|aws-ssm are currently available.") class FileProvider(object): """Support loading key material from disk.""" def key(self, key_name): """Takes key_name returns bytes""" config = common.get_config() key_dir = config( "secret_manager_file_path", namespace="cis", default=("{}/.mozilla-iam/keys/".format(os.path.expanduser("~"))), ) file_name = "{}".format(key_name) logger.debug("Secret manager file provider loading key file: {}/{}".format(key_dir, key_name)) fh = open((os.path.join(key_dir, file_name)), "rb") key_content = fh.read() key_construct = jwk.construct(key_content, "RS256") return key_construct class AWSParameterstoreProvider(object): """Support loading secure strings from AWS parameter store.""" def __init__(self): self.config = common.get_config() self.region_name = self.config("secret_manager_ssm_region", namespace="cis", default="us-west-2") self.boto_session = boto3.session.Session(region_name=self.region_name) self.ssm_client = self.boto_session.client("ssm") self._cache = {} def key(self, key_name): retries = 30 backoff = 1 result = None if "key_construct" in self._cache: logger.debug("Returning memory-cached version of the parameter key_construct") return self._cache["key_construct"] while result is None: try: ssm_namespace = self.config("secret_manager_ssm_path", namespace="cis", default="/iam") logger.debug("Secret manager SSM provider loading key: {}/{}".format(ssm_namespace, key_name)) ssm_response = self.ssm_client.get_parameter( Name="{}/{}".format(ssm_namespace, key_name), WithDecryption=True ) result = ssm_response.get("Parameter") except ClientError as e: retries = retries - 1 backoff = backoff + 1 time.sleep(backoff) logger.debug( "Backing-off: fetch secret ({}) due to: {} retries {} backoff {}".format( key_name, e, retries, backoff ) ) if retries <= 0: break if result is None: logger.error("Failed to fetch secret ({}) due to: retries {} backoff {}".format(key_name, retries, backoff)) try: key_dict = json.loads(result.get("Value")) key_construct = jwk.construct(key_dict, "RS256") except json.decoder.JSONDecodeError: key_construct = jwk.construct(result.get("Value"), "RS256") self._cache["key_construct"] = key_construct return key_construct def uuid_salt(self): retries = 30 backoff = 1 result = None if "uuid_salt" in self._cache: logger.debug("Returning memory-cached version of uuid_salt") return self._cache["uuid_salt"] while result is None: try: ssm_path = self.config("secret_manager_ssm_uuid_salt", namespace="cis", default="/iam") logger.debug("Secret manager SSM provider loading uuid_salt: {}".format(ssm_path)) ssm_response = self.ssm_client.get_parameter(Name=ssm_path, WithDecryption=True) result = ssm_response.get("Parameter").get("Value") except ClientError as e: retries = retries - 1 backoff = backoff + 1 logger.debug("Backing-off: fetch secret due to: {} retries {} backoff {}".format(e, retries, backoff)) time.sleep(backoff) if retries <= 0: break if result is None: logger.error("Failed to fetch uuid_salt due to: retries {} backoff {}".format(retries, backoff)) self._cache["uuid_salt"] = result return result
mpl-2.0
mozilla-iam/cis
python-modules/cis_profile_retrieval_service/cis_profile_retrieval_service/schema.py
1
1800
import json import graphene import cis_profile.graphene from cis_identity_vault.models import user from cis_profile_retrieval_service.common import get_table_resource def is_json(payload): """Check if a payload is valid JSON.""" try: json.loads(payload) except (TypeError, ValueError): return False else: return True class Query(graphene.ObjectType): """GraphQL Query class for the V2 Profiles.""" profiles = graphene.List(cis_profile.graphene.Profile, primaryEmail=graphene.String(required=False)) profile = graphene.Field(cis_profile.graphene.Profile, userId=graphene.String(required=True)) def resolve_profiles(self, info, **kwargs): """GraphQL resolver for the profiles attribute.""" table = get_table_resource() vault = user.Profile(table) profiles = [] if kwargs.get("primaryEmail"): search = vault.find_by_email(kwargs.get("primaryEmail")) if len(search.get("Items")) > 0: for profile in search.get("Items"): profiles.append(json.loads()) else: for vault_profile in vault.all: profiles.append(json.loads(vault_profile.get("profile"))) def resolve_profile(self, info, **kwargs): """GraphQL resolver for a single profile.""" table = get_table_resource() vault = user.Profile(table) if kwargs.get("userId"): search = vault.find_by_id(kwargs.get("userId")) if len(search.get("Items")) > 0: resp = search["Items"][0]["profile"] else: resp = json.dumps({}) return resp class AuthorizationMiddleware: def resolve(self, next, root, info, **kwargs): return next(root, info, **kwargs)
mpl-2.0
mozilla-iam/cis
python-modules/cis_profile/tests/test_well_known.py
1
1091
from cis_profile.common import WellKnown from cis_profile.profile import User import os class Test_WellKnown(object): def test_wellknown_file_force(self): wk = WellKnown(always_use_local_file=True) data = wk.get_well_known() assert isinstance(data, dict) assert isinstance(data.get("api"), dict) def test_wellknown_retrieve(self): wk = WellKnown() data = wk.get_well_known() assert isinstance(data, dict) assert isinstance(data.get("api"), dict) def test_schema_retrieve(self): wk = WellKnown() data = wk.get_schema() assert isinstance(data, dict) def test_rules_retrieve(self): wk = WellKnown() data = wk.get_publisher_rules() assert isinstance(data, dict) assert isinstance(data.get("create"), dict) def test_profile_env(self): os.environ["CIS_DISCOVERY_URL"] = "https://auth.allizom.org/.well-known/mozilla-iam" u = User() assert u._User__well_known.discovery_url == "https://auth.allizom.org/.well-known/mozilla-iam"
mpl-2.0
mozilla-iam/cis
python-modules/cis_crypto/setup.py
1
1388
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages with open("README.md", "r") as fh: long_description = fh.read() requirements = [ "python-jose[cryptography]", "cryptography", "everett", "everett[ini]", "configobj", "boto3", "boto", "botocore", "requests", "pyaml", ] setup_requirements = ["pytest-runner", "setuptools>=40.5.0"] test_requirements = ["pytest", "pytest-watch", "pytest-cov", "pytest-mock", "moto", "mock", "flake8", "cis_profile"] extras = {"test": test_requirements} setup( name="cis_crypto", version="0.0.1", author="Andrew Krug", author_email="akrug@mozilla.com", description="Per attribute signature system for jwks sign-verify in mozilla-iam.", long_description=long_description, url="https://github.com/mozilla-iam/cis", classifiers=[ "Programming Language :: Python :: 3", "License :: OSI Approved :: Mozilla Public License", "Operating System :: OS Independent", ], install_requires=requirements, license="Mozilla Public License 2.0", include_package_data=True, packages=find_packages(include=["cis_crypto", "bin"]), scripts=["bin/cis_crypto"], setup_requires=setup_requirements, test_suite="tests", tests_require=test_requirements, extras_require=extras, zip_safe=False, )
mpl-2.0
mozilla-iam/cis
python-modules/cis_change_service/tests/test_api.py
1
18809
import json import logging import mock import os import random import subprocess import string import cis_profile from cis_profile import common from cis_profile import FakeUser from cis_profile.fake_profile import FakeProfileConfig from cis_profile import profile from datetime import datetime from datetime import timedelta from tests.fake_auth0 import FakeBearer from tests.fake_auth0 import json_form_of_pk logging.basicConfig(level=logging.INFO, format="%(asctime)s:%(levelname)s:%(name)s:%(message)s") logging.getLogger("boto").setLevel(logging.CRITICAL) logging.getLogger("boto3").setLevel(logging.CRITICAL) logging.getLogger("botocore").setLevel(logging.CRITICAL) logging.getLogger("cis_profile.profile").setLevel(logging.DEBUG) logger = logging.getLogger(__name__) def get_complex_structures(): return ["staff_information", "access_information", "identities", "schema"] def ensure_appropriate_publishers_and_sign(fake_profile, publisher_rules, condition): os.environ["CIS_SECRET_MANAGER"] = "file" os.environ["CIS_SIGNING_KEY_NAME"] = "signing_key_name=fake-publisher-key_0.priv.pem" temp_profile = fake_profile complex_structures = get_complex_structures() for attr in publisher_rules[condition]: if attr == "primary_username" and temp_profile[attr]["value"] == "None": temp_profile[attr]["value"] = "".join( [random.choice(string.ascii_letters + string.digits) for n in range(32)] ) if attr not in complex_structures: successful_random_publisher = random.choice(publisher_rules[condition][attr]) temp_profile[attr]["signature"]["publisher"]["name"] = successful_random_publisher u = profile.User(user_structure_json=temp_profile) # Don't sign NULL attributes or invalid publishers if u._attribute_value_set(temp_profile[attr], strict=True) and ( temp_profile[attr]["signature"]["publisher"]["name"] == successful_random_publisher ): u.sign_attribute(attr, successful_random_publisher) temp_profile = u.as_dict() else: if attr != "schema" and attr in complex_structures: for k in temp_profile[attr]: try: successful_random_publisher = random.choice(publisher_rules[condition][attr]) except KeyError: successful_random_publisher = random.choice(publisher_rules[condition][attr][k]) temp_profile[attr][k]["signature"]["publisher"]["name"] = successful_random_publisher u = profile.User(user_structure_json=temp_profile) attribute = "{}.{}".format(attr, k) # Don't sign NULL attributes or invalid publishers if u._attribute_value_set(temp_profile[attr][k], strict=True) and ( temp_profile[attr][k]["signature"]["publisher"]["name"] == successful_random_publisher ): u.sign_attribute(attribute, successful_random_publisher) temp_profile = u.as_dict() return profile.User(user_structure_json=temp_profile) class TestAPI(object): def setup(self): self.dynalite_port = str(random.randint(32200, 32300)) os.environ["CIS_CONFIG_INI"] = "tests/mozilla-cis.ini" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-west-2" os.environ["AWS_ACCESS_KEY_ID"] = "foo" os.environ["AWS_SECRET_ACCESS_KEY"] = "bar" from cis_identity_vault import vault from cis_change_service.common import get_config self.patcher_salt = mock.patch("cis_crypto.secret.AWSParameterstoreProvider.uuid_salt") self.mock_salt = self.patcher_salt.start() config = get_config() os.environ["CIS_DYNALITE_PORT"] = str(random.randint(32300, 32400)) self.dynalite_port = config("dynalite_port", namespace="cis") self.dynaliteprocess = subprocess.Popen( [ "/usr/sbin/java", "-Djava.library.path=/opt/dynamodb_local/DynamoDBLocal_lib", "-jar", "/opt/dynamodb_local/DynamoDBLocal.jar", "-inMemory", "-port", self.dynalite_port, ], preexec_fn=os.setsid, ) v = vault.IdentityVault() v.connect() v.create() user_profile = FakeUser(config=FakeProfileConfig().minimal()) self.user_profile = user_profile.as_json() from cis_change_service import api api.app.testing = True self.app = api.app.test_client() self.publisher_rules = common.WellKnown().get_publisher_rules() self.complex_structures = get_complex_structures() def test_index_exists(self): result = self.app.get("/v2", follow_redirects=True) assert result.status_code == 200 @mock.patch("cis_change_service.idp.get_jwks") def test_stream_bypass_publishing_mode_it_should_succeed(self, fake_jwks): os.environ["CIS_STREAM_BYPASS"] = "true" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-west-2" from cis_change_service import api f = FakeBearer() fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_without_scope() api.app.testing = True self.app = api.app.test_client() result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(self.user_profile), content_type="application/json", follow_redirects=True, ) json.loads(result.get_data()) assert result.status_code == 200 @mock.patch("cis_change_service.idp.get_jwks") def test_change_endpoint_fails_with_invalid_token_and_jwt_validation_false(self, fake_jwks): os.environ["CIS_CONFIG_INI"] = "tests/mozilla-cis.ini" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-west-2" from cis_change_service import api os.environ["CIS_JWT_VALIDATION"] = "false" f = FakeBearer() bad_claims = { "iss": "https://auth-dev.mozilla.auth0.com/", "sub": "mc1l0G4sJI2eQfdWxqgVNcRAD9EAgHib@clients", "aud": "https://hacks", "iat": (datetime.utcnow() - timedelta(seconds=3100)).strftime("%s"), "exp": (datetime.utcnow() - timedelta(seconds=3100)).strftime("%s"), "scope": "read:allthething", "gty": "client-credentials", } fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_with_scope("read:profile", bad_claims) api.app.testing = True self.app = api.app.test_client() result = self.app.get( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(self.user_profile), content_type="application/json", follow_redirects=True, ) assert result.status_code == 200 @mock.patch("cis_change_service.idp.get_jwks") def test_wrong_publisher(self, fake_jwks): """ This verifies a wrong-publisher can't update it creates a valid user, then wrongly modify an attribute its not allowed to """ os.environ["CIS_CONFIG_INI"] = "tests/mozilla-cis-verify.ini" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-east-1" os.environ["AWS_ACCESS_KEY_ID"] = "foo" os.environ["AWS_SECRET_ACCESS_KEY"] = "bar" os.environ["DEFAULT_AWS_REGION"] = "us-east-1" os.environ["CIS_VERIFY_SIGNATURES"] = "true" os.environ["CIS_VERIFY_PUBLISHERS"] = "true" from cis_change_service import api fake_new_user = FakeUser(config=FakeProfileConfig().minimal().no_display()) # Create a brand new user patched_user_profile = ensure_appropriate_publishers_and_sign( fake_new_user.as_dict(), self.publisher_rules, "create" ) # Ensure a first_name is set as we'll use that for testing patched_user_profile.first_name.value = "test" patched_user_profile.first_name.signature.publisher.name = "ldap" patched_user_profile.first_name.metadata.display = "public" patched_user_profile.sign_attribute("first_name", "ldap") f = FakeBearer() fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_without_scope() api.app.testing = True self.app = api.app.test_client() result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(patched_user_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 200 assert response["condition"] == "create" # sign first_name again but with wrong publisher (but same value as before) new_user = cis_profile.User(user_id=patched_user_profile.user_id.value) new_user.first_name = patched_user_profile.first_name new_user.sign_attribute("first_name", "access_provider") result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(new_user.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert response["status_code"] == 202 # sign first_name again but with wrong publisher and different display (but same value as before) new_user = cis_profile.User(user_id=patched_user_profile.user_id.value) new_user.first_name = patched_user_profile.first_name new_user.first_name.metadata.display = "staff" new_user.sign_attribute("first_name", "access_provider") result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(new_user.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert response["code"] == "invalid_publisher" # sign first_name again but with wrong publisher and wrong value (it should fail) new_user.first_name.value = "new-test" new_user.sign_attribute("first_name", "access_provider") result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(new_user.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code != 200 @mock.patch("cis_change_service.idp.get_jwks") def test_partial_update_with_dynamo_down_it_should_fail(self, fake_jwks): os.environ["CIS_CONFIG_INI"] = "tests/mozilla-cis.ini" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-east-1" os.environ["AWS_ACCESS_KEY_ID"] = "foo" os.environ["AWS_SECRET_ACCESS_KEY"] = "bar" os.environ["DEFAULT_AWS_REGION"] = "us-east-1" from cis_change_service import api fake_new_user = FakeUser(config=FakeProfileConfig().minimal()) # Create a brand new user patched_user_profile = ensure_appropriate_publishers_and_sign( fake_new_user.as_dict(), self.publisher_rules, "create" ) f = FakeBearer() fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_without_scope() api.app.testing = True self.app = api.app.test_client() result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(patched_user_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 200 assert response["condition"] == "create" logger.info("A stub user has been created and verified to exist.") os.environ["CIS_DYNALITE_PORT"] = "31337" # bad port logger.info("Attempting failing partial update.") null_profile = profile.User(user_structure_json=None) null_profile.alternative_name.value = "iamanewpreferredlastname" null_profile.sign_attribute("alternative_name", "mozilliansorg") null_profile.user_id.value = "ad|wrong|LDAP" null_profile.active.value = True null_profile.sign_attribute("active", "access_provider") try: result = self.app.post( "/v2/user?user_id={}".format("mismatching_user_id"), headers={"Authorization": "Bearer " + token}, data=json.dumps(null_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 500 except Exception as e: logger.info(f"Exception was correctly raised because DynamoDB was unavailable: {e}") else: logger.warning("Exception was not raised, test has failed") assert False @mock.patch("cis_change_service.idp.get_jwks") def test_partial_update_it_should_fail(self, fake_jwks): os.environ["CIS_CONFIG_INI"] = "tests/mozilla-cis.ini" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_ENVIRONMENT"] = "local" os.environ["CIS_DYNALITE_PORT"] = self.dynalite_port os.environ["CIS_REGION_NAME"] = "us-east-1" os.environ["AWS_ACCESS_KEY_ID"] = "foo" os.environ["AWS_SECRET_ACCESS_KEY"] = "bar" os.environ["DEFAULT_AWS_REGION"] = "us-east-1" from cis_change_service import api fake_new_user = FakeUser(config=FakeProfileConfig().minimal()) # Create a brand new user patched_user_profile = ensure_appropriate_publishers_and_sign( fake_new_user.as_dict(), self.publisher_rules, "create" ) f = FakeBearer() fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_without_scope() api.app.testing = True self.app = api.app.test_client() result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(patched_user_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 200 assert response["condition"] == "create" logger.info("A stub user has been created and verified to exist.") logger.info("Attempting failing partial update.") null_profile = profile.User(user_structure_json=None) null_profile.alternative_name.value = "iamanewpreferredlastname" null_profile.sign_attribute("alternative_name", "mozilliansorg") null_profile.user_id.value = "ad|wrong|LDAP" null_profile.active.value = True null_profile.sign_attribute("active", "access_provider") result = self.app.post( "/v2/user?user_id={}".format("mismatching_user_id"), headers={"Authorization": "Bearer " + token}, data=json.dumps(null_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 400 @mock.patch("cis_change_service.idp.get_jwks") def test_partial_update_it_should_succeed(self, fake_jwks): os.environ["CIS_STREAM_BYPASS"] = "true" os.environ["AWS_XRAY_SDK_ENABLED"] = "false" os.environ["CIS_VERIFY_PUBLISHERS"] = "true" from cis_change_service import api fake_new_user = FakeUser(config=FakeProfileConfig().minimal()) # Create a brand new user patched_user_profile = ensure_appropriate_publishers_and_sign( fake_new_user.as_dict(), self.publisher_rules, "create" ) f = FakeBearer() fake_jwks.return_value = json_form_of_pk token = f.generate_bearer_without_scope() api.app.testing = True self.app = api.app.test_client() result = self.app.post( "/v2/user", headers={"Authorization": "Bearer " + token}, data=json.dumps(patched_user_profile.as_json()), content_type="application/json", follow_redirects=True, ) response = json.loads(result.get_data()) assert result.status_code == 200 assert response["condition"] == "create" logger.info("A stub user has been created and verified to exist.") logger.info("Attempting partial update.") # Now let's try a partial update :) null_profile = profile.User(user_structure_json=None) null_profile.active.value = True null_profile.sign_attribute("active", "access_provider") null_profile.last_name.value = "iamanewpreferredlastname" null_profile.sign_attribute("last_name", "mozilliansorg") result = self.app.post( "/v2/user?user_id={}".format(patched_user_profile.user_id.value), headers={"Authorization": "Bearer " + token}, data=json.dumps(null_profile.as_json()), content_type="application/json", follow_redirects=True, ) logger.info(result.get_data()) response = json.loads(result.get_data()) assert result.status_code == 200 assert response["condition"] == "update" def teardown(self): os.killpg(os.getpgid(self.dynaliteprocess.pid), 15) self.patcher_salt.stop()
mpl-2.0
ibm-watson-iot/iot-python
test/test_api_registry_devices_ext.py
2
3189
# ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import uuid import time from datetime import datetime import testUtils import wiotp.sdk.device from wiotp.sdk.api.registry.devices import DeviceUid, DeviceInfo, DeviceCreateRequest, DeviceLocation, LogEntry from wiotp.sdk.exceptions import ApiException class TestRegistryDevices(testUtils.AbstractTest): def testDeviceLocationGetAndUpdate(self, deviceType, device): assert device.deviceId in deviceType.devices locationBefore = device.getLocation() assert locationBefore is None device.setLocation({"latitude": 50, "longitude": 60}) locationAfter = device.getLocation() assert locationAfter.updatedDateTime is not None assert locationAfter.measuredDateTime is not None assert isinstance(locationAfter.updatedDateTime, datetime) assert isinstance(locationAfter.measuredDateTime, datetime) assert locationAfter.latitude == 50 assert locationAfter.longitude == 60 device.setLocation(DeviceLocation(latitude=80, longitude=75)) locationAfter = device.getLocation() assert locationAfter.latitude == 80 assert locationAfter.longitude == 75 def testDeviceLocationInvalid(self, deviceType, device): assert device.deviceId in deviceType.devices locationBefore = device.getLocation() assert locationBefore is None try: device.setLocation(DeviceLocation(latitude=100, longitude=120)) except ApiException as e: assert e.id == "CUDHT0300I" assert len(e.violations) == 1 def testDeviceMgmt(self, deviceType, device): assert device.deviceId in deviceType.devices mgmtInfo = device.getMgmt() assert mgmtInfo is None def testDeviceConnectionLogs(self, deviceType, device, authToken): assert device.deviceId in deviceType.devices options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": authToken}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() time.sleep(10) deviceClient.disconnect() deviceClient.connect() time.sleep(10) deviceClient.disconnect() # Allow 30 seconds for the logs to make it through time.sleep(30) connLogs = device.getConnectionLogs() # There may be more than 2 entries due to previous connection attempts if we re-used a device ID. But there should be at least two! assert len(connLogs) >= 2 for entry in connLogs: assert isinstance(entry, LogEntry) assert isinstance(entry.timestamp, datetime)
epl-1.0
ibm-watson-iot/iot-python
src/wiotp/sdk/device/client.py
2
4102
# ***************************************************************************** # Copyright (c) 2014, 2018 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** from datetime import datetime import json import logging import threading import paho.mqtt.client as paho import pytz from wiotp.sdk import ( AbstractClient, ConfigurationException, ConnectionException, MissingMessageEncoderException, InvalidEventException, ) from wiotp.sdk.device.command import Command from wiotp.sdk.device.config import DeviceClientConfig class DeviceClient(AbstractClient): """ Extends #wiotp.common.AbstractClient to implement a device client supporting messaging over MQTT # Parameters options (dict): Configuration options for the client logHandlers (list<logging.Handler>): Log handlers to configure. Defaults to `None`, which will result in a default log handler being created. """ _COMMAND_TOPIC = "iot-2/cmd/+/fmt/+" def __init__(self, config, logHandlers=None): self._config = DeviceClientConfig(**config) AbstractClient.__init__( self, domain=self._config.domain, organization=self._config.orgId, clientId=self._config.clientId, username=self._config.username, password=self._config.password, port=self._config.port, transport=self._config.transport, cleanStart=self._config.cleanStart, sessionExpiry=self._config.sessionExpiry, keepAlive=self._config.keepAlive, caFile=self._config.caFile, logLevel=self._config.logLevel, logHandlers=logHandlers, ) # Add handler for commands if not connected to QuickStart if not self._config.isQuickstart(): self.client.message_callback_add("iot-2/cmd/+/fmt/+", self._onCommand) # Initialize user supplied callback self.commandCallback = None # Register startup subscription list (only for non-Quickstart) if not self._config.isQuickstart(): self._subscriptions[self._COMMAND_TOPIC] = 1 def publishEvent(self, eventId, msgFormat, data, qos=0, onPublish=None): """ Publish an event to Watson IoT Platform. # Parameters eventId (string): Name of this event msgFormat (string): Format of the data for this event data (dict): Data for this event qos (int): MQTT quality of service level to use (`0`, `1`, or `2`) onPublish(function): A function that will be called when receipt of the publication is confirmed. # Callback and QoS The use of the optional #onPublish function has different implications depending on the level of qos used to publish the event: - qos 0: the client has asynchronously begun to send the event - qos 1 and 2: the client has confirmation of delivery from the platform """ topic = "iot-2/evt/{eventId}/fmt/{msgFormat}".format(eventId=eventId, msgFormat=msgFormat) return self._publishEvent(topic, eventId, msgFormat, data, qos, onPublish) def _onCommand(self, client, userdata, pahoMessage): """ Internal callback for device command messages, parses source device from topic string and passes the information on to the registered device command callback """ try: command = Command(pahoMessage, self._messageCodecs) except InvalidEventException as e: self.logger.critical(str(e)) else: self.logger.debug("Received command '%s'" % (command.commandId)) if self.commandCallback: self.commandCallback(command)
epl-1.0
ibm-watson-iot/iot-python
test/test_api_registry_devicetypes.py
2
6161
# ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import uuid import pytest import testUtils from wiotp.sdk.api.registry.devices import DeviceInfo from wiotp.sdk.exceptions import ApiException class TestRegistryDevicetypes(testUtils.AbstractTest): # ========================================================================= # Device Type tests # ========================================================================= def testDeviceTypeExistsCheck(self, deviceType): if deviceType.id in self.appClient.registry.devicetypes: pass else: raise Exception() if "doesntexist" not in self.appClient.registry.devicetypes: pass else: raise Exception() def testGetDeviceType(self, deviceType): retrievedDeviceType = self.appClient.registry.devicetypes[deviceType.id] assert retrievedDeviceType.id == deviceType.id assert retrievedDeviceType.classId == "Device" def testGetDeviceTypeThatDoesntExist(self): with pytest.raises(Exception): self.appClient.registry.devicetypes["doesntexist"] def testUnsupportedCreateUpdate(self): with pytest.raises(Exception): self.appClient.registry.devicetypes["d:hldtxx:vm:iot-test-06"] = {"foo", "bar"} def testListDeviceTypes(self, deviceType): count = 0 for type in self.appClient.registry.devicetypes: count += 1 if count > 10: break # DeviceTypeDescription test def testCreateDeviceType(self): typeId = str(uuid.uuid4()) myDeviceType = self.appClient.registry.devicetypes.create({"id": typeId, "description": "This is a test"}) myDeviceTypeRetrieved = self.appClient.registry.devicetypes[typeId] assert myDeviceTypeRetrieved.id == typeId assert myDeviceTypeRetrieved.description == "This is a test" del self.appClient.registry.devicetypes[typeId] def testCreateDeviceTypeNone(self): typeId = str(uuid.uuid4()) myDeviceType = self.appClient.registry.devicetypes.create({"id": typeId, "description": None}) myDeviceTypeRetrieved = self.appClient.registry.devicetypes[typeId] assert myDeviceTypeRetrieved.id == typeId assert myDeviceTypeRetrieved.description == None del self.appClient.registry.devicetypes[typeId] # Metadata test def testCreateDeviceMetadata(self): typeId = str(uuid.uuid4()) myDeviceType = self.appClient.registry.devicetypes.create( {"id": typeId, "description": "This is still a test", "metadata": {"test": "test"}} ) myDeviceTypeRetrieved = self.appClient.registry.devicetypes[typeId] assert myDeviceTypeRetrieved.id == typeId assert myDeviceTypeRetrieved.description == "This is still a test" assert myDeviceTypeRetrieved.metadata == {"test": "test"} del self.appClient.registry.devicetypes[typeId] def testCreateDeviceMetadataNone(self): typeId = str(uuid.uuid4()) myDeviceType = self.appClient.registry.devicetypes.create( {"id": typeId, "description": "This is still a test", "metadata": None} ) myDeviceTypeRetrieved = self.appClient.registry.devicetypes[typeId] assert myDeviceTypeRetrieved.id == typeId assert myDeviceTypeRetrieved.description == "This is still a test" assert myDeviceTypeRetrieved.metadata == None del self.appClient.registry.devicetypes[typeId] def testUpdateDeviceType(self, deviceType): self.appClient.registry.devicetypes.update(deviceType.id, description="This is still a test") updatedDeviceType = self.appClient.registry.devicetypes[deviceType.id] assert updatedDeviceType.description == "This is still a test" def testUpdateDeviceInfo(self, deviceType): self.appClient.registry.devicetypes.update(deviceType.id, deviceInfo=DeviceInfo(serialNumber="111")) updatedDeviceType = self.appClient.registry.devicetypes[deviceType.id] assert updatedDeviceType.deviceInfo.serialNumber == "111" # ========================================================================= # Device under DeviceType tests # ========================================================================= def testDeviceExistsCheck(self, deviceType, device): if device.deviceId in deviceType.devices: pass else: raise Exception() if "wheredidyago" not in deviceType.devices: pass else: raise Exception() def testGetDeviceFromDeviceType(self, deviceType, device): myDevice = self.appClient.registry.devicetypes[deviceType.id].devices[device.deviceId] def testListDevicesFromDeviceType(self, deviceType, device): # Get a device, and cache the response in a local object count = 0 for device in deviceType.devices: count += 1 if count > 10: break def testCreateDeviceType(self): with pytest.raises(ApiException): typeId = 1 r = self.appClient.registry.devicetypes.create(typeId) def testUpdateDeviceType(self): with pytest.raises(ApiException): data = None r = self.appClient.registry.devicetypes.update(data) def testDeleteTypeId(self, device, deviceType): typeId = str(uuid.uuid4()) self.appClient.registry.devicetypes.create( {"id": typeId, "description": "This is still a test", "metadata": {"test": "test"}} ) self.appClient.registry.devicetypes.delete(typeId) assert typeId not in deviceType.devices
epl-1.0
ibm-watson-iot/iot-python
test/test_api_state_logical_interfaces.py
2
7820
# ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** # import uuid from datetime import datetime import testUtils import time import pytest from wiotp.sdk.exceptions import ApiException import string import json @testUtils.oneJobOnlyTest class TestLogicalInterfaces(testUtils.AbstractTest): testSchemaName = "python-api-test-li-schema" testLISchema = { "$schema": "http://json-schema.org/draft-04/schema#", "type": "object", "title": "Environment Sensor Schema", "properties": { "temperature": { "description": "temperature in degrees Celsius", "type": "number", "minimum": -237.15, "default": 0.0, }, "humidity": {"description": "relative humidity (%)", "type": "number", "minimum": 0.0, "default": 0.0}, "publishTimestamp": {"description": "publishTimestamp", "type": "number", "minimum": 0.0, "default": 0.0}, }, "required": ["temperature", "humidity", "publishTimestamp"], } testLogicalInterfaceName = "python-api-test-logicalInterface" updatedLogicalInterfaceName = "python-api-test-logicalInterface-updated" # ========================================================================= # Set up services # ========================================================================= def testCleanup(self): for li in self.appClient.state.draft.logicalInterfaces: if li.name in ( TestLogicalInterfaces.testLogicalInterfaceName, TestLogicalInterfaces.updatedLogicalInterfaceName, ): # print("Deleting old test schema instance: %s" % (a)) del self.appClient.state.draft.logicalInterfaces[li.id] for s in self.appClient.state.draft.schemas: if s.name == TestLogicalInterfaces.testSchemaName: del self.appClient.state.draft.schemas[s.id] def checkLI(self, logicalInterface, name, description, schemaId, version, alias): assert logicalInterface.name == name assert logicalInterface.description == description assert logicalInterface.schemaId == schemaId assert logicalInterface.version == version assert logicalInterface.alias == alias assert isinstance(logicalInterface.created, datetime) assert isinstance(logicalInterface.createdBy, str) assert isinstance(logicalInterface.updated, datetime) assert isinstance(logicalInterface.updatedBy, str) def doesSchemaNameExist(self, name): for a in self.appClient.state.draft.schemas.find({"name": name}): if a.name == name: return True return False def doesLINameExist(self, name): for li in self.appClient.state.draft.logicalInterfaces.find({"name": name}): if li.name == name: return True return False def createSchema(self, name, schemaFileName, schemaContents, description): jsonSchemaContents = json.dumps(schemaContents) createdSchema = self.appClient.state.draft.schemas.create(name, schemaFileName, jsonSchemaContents, description) return createdSchema def createAndCheckLI(self, name, description, schemaId, version, alias): createdLI = self.appClient.state.draft.logicalInterfaces.create( {"name": name, "description": description, "schemaId": schemaId, "version": version, "alias": alias} ) self.checkLI(createdLI, name, description, schemaId, version, alias) # now actively refetch the LI to check it is stored fetchedLI = self.appClient.state.draft.logicalInterfaces.__getitem__(createdLI.id) assert createdLI == fetchedLI return createdLI def testLogicalInterfaceCRUD(self): test_schema_name = TestLogicalInterfaces.testSchemaName assert self.doesSchemaNameExist(test_schema_name) == False testLIName = TestLogicalInterfaces.testLogicalInterfaceName assert self.doesLINameExist(testLIName) == False # Create a schema createdSchema = self.createSchema( test_schema_name, "liSchema.json", TestLogicalInterfaces.testLISchema, "Test schema description" ) # Create a Logical Interface createdLI = self.createAndCheckLI( testLIName, "Test Logical Interface description", createdSchema.id, "draft", "alias" ) # Can we search for it assert self.doesLINameExist(testLIName) == True # Update the LI updated_li_name = TestLogicalInterfaces.updatedLogicalInterfaceName updatedLI = self.appClient.state.draft.logicalInterfaces.update( createdLI.id, { "id": createdLI.id, "name": updated_li_name, "description": "Test LI updated description", "schemaId": createdSchema.id, "version": "draft", "alias": "test", }, ) self.checkLI(updatedLI, updated_li_name, "Test LI updated description", createdSchema.id, "draft", "test") # Delete the LI del self.appClient.state.draft.logicalInterfaces[createdLI.id] # It should be gone assert self.doesLINameExist(testLIName) == False # Delete the schema del self.appClient.state.draft.schemas[createdSchema.id] # It should be gone assert self.doesSchemaNameExist(test_schema_name) == False def testLogicalInterfaceActivation(self): test_schema_name = TestLogicalInterfaces.testSchemaName assert self.doesSchemaNameExist(test_schema_name) == False testLIName = TestLogicalInterfaces.testLogicalInterfaceName assert self.doesLINameExist(testLIName) == False # Create a schema createdSchema = self.createSchema( test_schema_name, "liSchema.json", TestLogicalInterfaces.testLISchema, "Test schema description" ) # Create a Logical Interface createdLI = self.createAndCheckLI( testLIName, "Test Logical Interface description", createdSchema.id, "draft", "alias" ) # Can we search for it assert self.doesLINameExist(testLIName) == True # Validate and Activate the LI createdLI.validate() print("LI Differences: %s " % createdLI.validate()) # Activating the Li should fail as it is not yet associated with a Device or Thing Type. try: createdLI.activate() # Hmm, the activate should raise an exception assert False except: assert True # The expected exception was raised # This should fail as there are currently no differences with the LI try: createdLI.differences() # Should raise an exception assert False except: assert True # The expected exception was raised # Delete the LI del self.appClient.state.draft.logicalInterfaces[createdLI.id] # It should be gone assert self.doesLINameExist(testLIName) == False # Delete the schema del self.appClient.state.draft.schemas[createdSchema.id] # It should be gone assert self.doesSchemaNameExist(test_schema_name) == False
epl-1.0
ibm-watson-iot/iot-python
test/test_codecs_utf8.py
2
1309
# ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import os import testUtils from wiotp.sdk import InvalidEventException, Utf8Codec class NonJsonDummyPahoMessage(object): def __init__(self, object): self.payload = bytearray() try: self.payload.extend(object) except: # python 3 self.payload.extend(map(ord, object)) class TestDevice(testUtils.AbstractTest): def testFileObject(self): cwd = os.getcwd() fileContent = None with open("%s/README.md" % cwd) as fileIn: fileContent = fileIn.read() assert fileContent is not None encodedPayload = Utf8Codec.encode(fileContent, None) message = Utf8Codec.decode(NonJsonDummyPahoMessage(encodedPayload)) assert message.data.__class__.__name__ in ["str", "unicode"] assert message.data == fileContent
epl-1.0
ibm-watson-iot/iot-python
src/wiotp/sdk/application/client.py
2
21444
# ***************************************************************************** # Copyright (c) 2014, 2018 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import os import re import json import iso8601 import uuid from datetime import datetime from wiotp.sdk import ConnectionException, MissingMessageEncoderException, AbstractClient, InvalidEventException from wiotp.sdk.application.messages import Status, Command, Event, State, Error, ThingError, DeviceState from wiotp.sdk.application.config import ApplicationClientConfig from wiotp.sdk.api import ApiClient, Registry, Usage, ServiceStatus, DSC, LEC, Mgmt, ServiceBindings, Actions, StateMgr import paho.mqtt.client as paho import requests class ApplicationClient(AbstractClient): """ Extends #wiotp.AbstractClient to implement an application client supporting messaging over MQTT # Parameters options (dict): Configuration options for the client logHandlers (list<logging.Handler>): Log handlers to configure. Defaults to `None`, which will result in a default log handler being created. # Configuration Options The options parameter expects a Python dictionary containing the following keys: - `auth-key` The API key to to securely connect your application to Watson IoT Platform. - `auth-token` An authentication token to securely connect your application to Watson IoT Platform. - `clean-session` A boolean value indicating whether to use MQTT clean session. """ def __init__(self, config, logHandlers=None): self._config = ApplicationClientConfig(**config) # Call parent constructor AbstractClient.__init__( self, domain=self._config.domain, organization=self._config.orgId, clientId=self._config.clientId, username=self._config.username, password=self._config.password, logHandlers=logHandlers, cleanStart=self._config.cleanStart, port=self._config.port, transport=self._config.transport, caFile=self._config.caFile, logLevel=self._config.logLevel, sessionExpiry=self._config.sessionExpiry, keepAlive=self._config.keepAlive, ) # Add handlers for events and status self.client.message_callback_add("iot-2/type/+/id/+/evt/+/fmt/+", self._onDeviceEvent) self.client.message_callback_add("iot-2/type/+/id/+/mon", self._onDeviceStatus) self.client.message_callback_add("iot-2/app/+/mon", self._onAppStatus) self.client.message_callback_add("iot-2/type/+/id/+/intf/+/evt/state", self._onDeviceState) self.client.message_callback_add("iot-2/thing/type/+/id/+/intf/+/evt/state", self._onThingState) self.client.message_callback_add("iot-2/type/+/id/+/err/data", self._onErrorTopic) self.client.message_callback_add("iot-2/thing/type/+/id/+/err/data", self._onThingError) # Add handler for commands if not connected to QuickStart if not self._config.isQuickstart(): self.client.message_callback_add("iot-2/type/+/id/+/cmd/+/fmt/+", self._onDeviceCommand) # Attach fallback handler self.client.on_message = self._onUnsupportedMessage # Initialize user supplied callbacks self.deviceEventCallback = None self.deviceCommandCallback = None self.deviceStateCallback = None self.deviceStatusCallback = None self.thingStateCallback = None self.errorTopicCallback = None self.appStatusCallback = None # Create an api client if not connected in QuickStart mode if not self._config.isQuickstart(): apiClient = ApiClient(self._config, self.logger) self.registry = Registry(apiClient) self.usage = Usage(apiClient) self.dsc = DSC(apiClient) self.lec = LEC(apiClient) self.mgmt = Mgmt(apiClient) self.serviceBindings = ServiceBindings(apiClient) self.actions = Actions(apiClient) self.state = StateMgr(apiClient) # We directly expose the get() method via self.serviceStatus() self._serviceStatus = ServiceStatus(apiClient) def serviceStatus(self): if not self._config.isQuickstart(): return self._serviceStatus.get() else: return None def subscribeToDeviceEvents(self, typeId="+", deviceId="+", eventId="+", msgFormat="+", qos=0): """ Subscribe to device event messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all device types (MQTT `+` wildcard) deviceId (string): deviceId for the subscription, optional. Defaults to all devices (MQTT `+` wildcard) eventId (string): eventId for the subscription, optional. Defaults to all events (MQTT `+` wildcard) msgFormat (string): msgFormat for the subscription, optional. Defaults to all formats (MQTT `+` wildcard) qos (int): MQTT quality of service level to use (`0`, `1`, or `2`) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart() and deviceId == "+": self.logger.warning( "QuickStart applications do not support wildcard subscription to events from all devices" ) return 0 topic = "iot-2/type/%s/id/%s/evt/%s/fmt/%s" % (typeId, deviceId, eventId, msgFormat) return self._subscribe(topic, qos) def subscribeToDeviceStatus(self, typeId="+", deviceId="+"): """ Subscribe to device status messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all device types (MQTT `+` wildcard) deviceId (string): deviceId for the subscription, optional. Defaults to all devices (MQTT `+` wildcard) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart() and deviceId == "+": self.logger.warning("QuickStart applications do not support wildcard subscription to device status") return 0 topic = "iot-2/type/%s/id/%s/mon" % (typeId, deviceId) return self._subscribe(topic, 0) def subscribeToErrorTopic(self, typeId="+", Id="+"): """ Subscribe to device error messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all device types (MQTT `+` wildcard) Id (string): deviceId for the subscription, optional. Defaults to all devices (MQTT `+` wildcard) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart() and Id == "+": self.logger.warning("QuickStart applications do not support wildcard subscription to error topics") return 0 topic = "iot-2/type/%s/id/%s/err/data" % (typeId, Id) return self._subscribe(topic, 0) def subscribeToThingErrors(self, typeId="+", Id="+"): """ Subscribe to thingerror messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all thing types (MQTT `+` wildcard) Id (string): thingId for the subscription, optional. Defaults to all things (MQTT `+` wildcard) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart() and Id == "+": self.logger.warning("QuickStart applications do not support wildcard subscription to error topics") return 0 topic = "iot-2/thing/type/%s/id/%s/err/data" % (typeId, Id) return self._subscribe(topic, 0) def subscribeToThingState(self, typeId="+", thingId="+", logicalInterfaceId="+"): """ Subscribe to thing state messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all thing types (MQTT `+` wildcard) thingId (string): thingId for the subscription, optional. Defaults to all things (MQTT `+` wildcard) logicalInterfaceId (string): logicalInterfaceId for the subscription, optional. Defaults to all LIs (MQTT `+` wildcard) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart(): self.logger.warning("QuickStart applications do not support thing state") return 0 topic = "iot-2/thing/type/%s/id/%s/intf/%s/evt/state" % (typeId, thingId, logicalInterfaceId) return self._subscribe(topic, 0) def subscribeToDeviceState(self, typeId="+", deviceId="+", logicalInterfaceId="+"): """ Subscribe to device state messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all thing types (MQTT `+` wildcard) deviceId (string): thingId for the subscription, optional. Defaults to all devices (MQTT `+` wildcard) logicalInterfaceId (string): logicalInterfaceId for the subscription, optional. Defaults to all LIs (MQTT `+` wildcard) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart(): self.logger.warning("QuickStart applications do not support device state") return 0 topic = "iot-2/type/%s/id/%s/intf/%s/evt/state" % (typeId, deviceId, logicalInterfaceId) return self._subscribe(topic, 0) def subscribeToDeviceCommands(self, typeId="+", deviceId="+", commandId="+", msgFormat="+"): """ Subscribe to device command messages # Parameters typeId (string): typeId for the subscription, optional. Defaults to all device types (MQTT `+` wildcard) deviceId (string): deviceId for the subscription, optional. Defaults to all devices (MQTT `+` wildcard) commandId (string): commandId for the subscription, optional. Defaults to all commands (MQTT `+` wildcard) msgFormat (string): msgFormat for the subscription, optional. Defaults to all formats (MQTT `+` wildcard) qos (int): MQTT quality of service level to use (`0`, `1`, or `2`) # Returns int: If the subscription was successful then the return Message ID (mid) for the subscribe request will be returned. The mid value can be used to track the subscribe request by checking against the mid argument if you register a subscriptionCallback method. If the subscription fails then the return value will be `0` """ if self._config.isQuickstart(): self.logger.warning("QuickStart applications do not support commands") return 0 topic = "iot-2/type/%s/id/%s/cmd/%s/fmt/%s" % (typeId, deviceId, commandId, msgFormat) return self._subscribe(topic, 0) def publishEvent(self, typeId, deviceId, eventId, msgFormat, data, qos=0, onPublish=None): topic = "iot-2/type/%s/id/%s/evt/%s/fmt/%s" % (typeId, deviceId, eventId, msgFormat) return self._publishEvent(topic, eventId, msgFormat, data, qos, onPublish) def publishCommand(self, typeId, deviceId, commandId, msgFormat, data=None, qos=0, onPublish=None): """ Publish a command to a device # Parameters typeId (string) : The type of the device this command is to be published to deviceId (string): The id of the device this command is to be published to command (string) : The name of the command msgFormat (string) : The format of the command payload data (dict) : The command data qos (int) : The equivalent MQTT semantics of quality of service using the same constants (optional, defaults to `0`) onPublish (function) : A function that will be called when receipt of the publication is confirmed. This has different implications depending on the qos: - qos 0 : the client has asynchronously begun to send the event - qos 1 and 2 : the client has confirmation of delivery from WIoTP """ if self._config.isQuickstart(): self.logger.warning("QuickStart applications do not support sending commands") return False if not self.connectEvent.wait(timeout=10): return False else: topic = "iot-2/type/%s/id/%s/cmd/%s/fmt/%s" % (typeId, deviceId, commandId, msgFormat) # Raise an exception if there is no codec for this msgFormat if self.getMessageCodec(msgFormat) is None: raise MissingMessageEncoderException(msgFormat) payload = self.getMessageCodec(msgFormat).encode(data, datetime.now()) result = self.client.publish(topic, payload=payload, qos=qos, retain=False) if result[0] == paho.MQTT_ERR_SUCCESS: # Because we are dealing with aync pub/sub model and callbacks it is possible that # the _onPublish() callback for this mid is called before we obtain the lock to place # the mid into the _onPublishCallbacks list. # # _onPublish knows how to handle a scenario where the mid is not present (no nothing) # in this scenario we will need to invoke the callback directly here, because at the time # the callback was invoked the mid was not yet in the list. with self._messagesLock: if result[1] in self._onPublishCallbacks: # paho callback beat this thread so call callback inline now del self._onPublishCallbacks[result[1]] if onPublish is not None: onPublish() else: # this thread beat paho callback so set up for call later self._onPublishCallbacks[result[1]] = onPublish return True else: return False def _onUnsupportedMessage(self, client, userdata, message): """ Internal callback for messages that have not been handled by any of the specific internal callbacks, these messages are not passed on to any user provided callback """ self.logger.warning( "Received messaging on unsupported topic '%s' on topic '%s'" % (message.payload, message.topic) ) def _onDeviceEvent(self, client, userdata, pahoMessage): """ Internal callback for device event messages, parses source device from topic string and passes the information on to the registerd device event callback """ try: event = Event(pahoMessage, self._messageCodecs) self.logger.debug("Received event '%s' from %s:%s" % (event.eventId, event.typeId, event.deviceId)) if self.deviceEventCallback: self.deviceEventCallback(event) except InvalidEventException as e: self.logger.critical(str(e)) def _onThingState(self, client, userdata, pahoMessage): """ Internal callback for thing state messages, parses source thing from topic string and passes the information on to the registerd thing state callback """ try: state = State(pahoMessage) self.logger.debug("Received state from %s:%s" % (state.typeId, state.thingId)) if self.thingStateCallback: self.thingStateCallback(state) except InvalidEventException as e: self.logger.critical(str(e)) def _onDeviceState(self, client, userdata, pahoMessage): """ Internal callback for thing state messages, parses source thing from topic string and passes the information on to the registerd thing state callback """ try: state = DeviceState(pahoMessage) self.logger.debug("Received state from %s:%s" % (state.typeId, state.deviceId)) if self.deviceStateCallback: self.deviceStateCallback(state) except InvalidEventException as e: self.logger.critical(str(e)) def _onErrorTopic(self, client, userdata, pahoMessage): """ Internal callback for error messages, parses source thing/device from topic string and passes the information on to the registerd error callback """ try: error = Error(pahoMessage) self.logger.debug("Received error from device %s:%s" % (error.typeId, error.id)) if self.errorTopicCallback: self.errorTopicCallback(error) except InvalidEventException as e: self.logger.critical(str(e)) def _onThingError(self, client, userdata, pahoMessage): """ Internal callback for error messages, parses source thing/device from topic string and passes the information on to the registerd error callback """ try: error = ThingError(pahoMessage) self.logger.debug("Received error from thing %s:%s" % (error.typeId, error.id)) if self.errorTopicCallback: self.errorTopicCallback(error) except InvalidEventException as e: self.logger.critical(str(e)) def _onDeviceCommand(self, client, userdata, pahoMessage): """ Internal callback for device command messages, parses source device from topic string and passes the information on to the registerd device command callback """ try: command = Command(pahoMessage, self._messageCodecs) self.logger.debug( "Received command '%s' from %s:%s" % (command.commandId, command.typeId, command.deviceId) ) if self.deviceCommandCallback: self.deviceCommandCallback(command) except InvalidEventException as e: self.logger.critical(str(e)) def _onDeviceStatus(self, client, userdata, pahoMessage): """ Internal callback for device status messages, parses source device from topic string and passes the information on to the registerd device status callback """ try: status = Status(pahoMessage) self.logger.debug("Received %s action from %s" % (status.action, status.clientId)) if self.deviceStatusCallback: self.deviceStatusCallback(status) except InvalidEventException as e: self.logger.critical(str(e)) def _onAppStatus(self, client, userdata, pahoMessage): """ Internal callback for application command messages, parses source application from topic string and passes the information on to the registerd applicaion status callback """ try: status = Status(pahoMessage) self.logger.debug("Received %s action from %s" % (status.action, status.clientId)) if self.appStatusCallback: self.appStatusCallback(status) except InvalidEventException as e: self.logger.critical(str(e))
epl-1.0
ibm-watson-iot/iot-python
src/wiotp/sdk/gateway/client.py
2
6109
# ***************************************************************************** # Copyright (c) 2016, 2018 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import json import re import pytz import uuid import threading import requests import logging import paho.mqtt.client as paho from datetime import datetime from wiotp.sdk import ( AbstractClient, InvalidEventException, UnsupportedAuthenticationMethod, ConfigurationException, ConnectionException, MissingMessageEncoderException, MissingMessageDecoderException, ) from wiotp.sdk.device import DeviceClient from wiotp.sdk.device.command import Command from wiotp.sdk.gateway.config import GatewayClientConfig from wiotp.sdk.gateway.messages import Notification class GatewayClient(DeviceClient): def __init__(self, config, logHandlers=None): """ Override the constructor """ self._config = GatewayClientConfig(**config) AbstractClient.__init__( self, domain=self._config.domain, organization=self._config.orgId, clientId=self._config.clientId, username=self._config.username, password=self._config.password, port=self._config.port, transport=self._config.transport, cleanStart=self._config.cleanStart, sessionExpiry=self._config.sessionExpiry, keepAlive=self._config.keepAlive, caFile=self._config.caFile, logLevel=self._config.logLevel, logHandlers=logHandlers, ) self.COMMAND_TOPIC = "iot-2/type/" + self._config.typeId + "/id/" + self._config.deviceId + "/cmd/+/fmt/+" gatewayCommandTopic = "iot-2/type/" + self._config.typeId + "/id/" + self._config.deviceId + "/cmd/+/fmt/json" deviceCommandTopic = "iot-2/type/+/id/+/cmd/+/fmt/+" messageNotificationTopic = "iot-2/type/" + self._config.typeId + "/id/" + self._config.deviceId + "/notify" self.client.message_callback_add(gatewayCommandTopic, self._onCommand) self.client.message_callback_add(deviceCommandTopic, self._onDeviceCommand) self.client.message_callback_add(messageNotificationTopic, self._onMessageNotification) # Initialize user supplied callback self.commandCallback = None self.deviceCommandCallback = None self.notificationCallback = None self.client.on_connect = self._onConnect self.client.on_disconnect = self._onDisconnect def publishDeviceEvent(self, typeId, deviceId, eventId, msgFormat, data, qos=0, onPublish=None): topic = "iot-2/type/" + typeId + "/id/" + deviceId + "/evt/" + eventId + "/fmt/" + msgFormat return self._publishEvent(topic, eventId, msgFormat, data, qos, onPublish) def publishEvent(self, eventId, msgFormat, data, qos=0, onPublish=None): topic = ( "iot-2/type/" + self._config.typeId + "/id/" + self._config.deviceId + "/evt/" + eventId + "/fmt/" + msgFormat ) return self._publishEvent(topic, eventId, msgFormat, data, qos, onPublish) def subscribeToDeviceCommands(self, typeId, deviceId, commandId="+", msgFormat="json", qos=1): topic = "iot-2/type/" + typeId + "/id/" + deviceId + "/cmd/" + commandId + "/fmt/" + msgFormat return self._subscribe(topic, qos=1) def subscribeToCommands(self, commandId="+", msgFormat="json", qos=1): typeId = self._config.typeId deviceId = self._config.deviceId topic = "iot-2/type/" + typeId + "/id/" + deviceId + "/cmd/" + commandId + "/fmt/" + msgFormat return self._subscribe(topic, qos=1) def subscribeToNotifications(self): typeId = self._config.typeId deviceId = self._config.deviceId topic = "iot-2/type/" + typeId + "/id/" + deviceId + "/notify" return self._subscribe(topic, qos=0) def _onCommand(self, client, userdata, pahoMessage): """ Internal callback for device command messages, parses source device from topic string and passes the information on to the registered device command callback """ try: command = Command(pahoMessage, self._messageCodecs) except InvalidEventException as e: self.logger.critical(str(e)) else: self.logger.debug("Received device command '%s'" % (command.commandId)) if self.commandCallback: self.commandCallback(command) def _onDeviceCommand(self, client, userdata, pahoMessage): """ Internal callback for gateway command messages, parses source device from topic string and passes the information on to the registered device command callback """ try: command = Command(pahoMessage, self._messageCodecs) except InvalidEventException as e: self.logger.critical(str(e)) else: self.logger.debug("Received gateway command '%s'" % (command.commandId)) if self.deviceCommandCallback: self.deviceCommandCallback(command) def _onMessageNotification(self, client, userdata, pahoMessage): """ Internal callback for gateway notification messages, parses source device from topic string and passes the information on to the registered device command callback """ try: note = Notification(pahoMessage, self._messageCodecs) except InvalidEventException as e: self.logger.critical(str(e)) else: self.logger.debug("Received Notification") if self.notificationCallback: self.notificationCallback(note)
epl-1.0
ibm-watson-iot/iot-python
samples/deviceFactory/deviceStatus.py
2
3130
#!/usr/bin/env python # ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import argparse import sys import os import yaml import wiotp.sdk.application def loadConfigFile(source): data = {} with open(source, "r") as sourceFile: data = yaml.full_load(sourceFile) return data if __name__ == "__main__": # Initialize the properties we need parser = argparse.ArgumentParser( description="IBM Watson IoT Platform Device Status. For more information see https://github.com/ibm-watson-iot/iot-python/samples/deviceFactory" ) parser.add_argument( "-t", "--typeId", required=False, default="iotpsutil", help="Set the typeId for the device batch. Defaults to iotpsutil", ) parser.add_argument( "-b", "--batchId", required=True, help="DeviceIDs will be prefixed by the batch number, e.g. batchID-0001, batchID-0002", ) parser.add_argument( "-n", "--numberOfDevices", required=True, type=int, help="How many device configuration files should be produced by the factory. Max value is 1000", ) args, unknown = parser.parse_known_args() options = wiotp.sdk.application.parseEnvVars() client = wiotp.sdk.application.ApplicationClient(options) # Terminal colour mods red = "%c[31m" % chr(27) green = "%c[32m" % chr(27) off = "%c[0m" % chr(27) statuses = client.registry.connectionStatus.find(typeId=args.typeId) output = {} for status in statuses: # print(status) clientId = status["id"] deviceId = clientId.split(":")[3] if not deviceId.startswith(args.batchId): continue (batchId, batchNum) = clientId.split("-") if status["connectionStatus"] == "disconnected": output[batchNum] = "%s%s%s" % (red, batchNum, off) elif status["connectionStatus"] == "connected": output[batchNum] = "%s%s%s" % (green, batchNum, off) else: output[batchNum] = "%s" % (batchNum) print("=================================================") print("Device Connection State Report") print("") print("%s:%s-x" % (args.typeId, args.batchId)) print("") print("%sconnected%s / %sdisconnected%s / unknown" % (green, off, red, off)) print("=================================================") outStr = "" for i in range(1, args.numberOfDevices + 1): batchNum = "%04d" % (i) if batchNum in output: outStr += output[batchNum] + " " else: outStr += batchNum + " " if batchNum[3] == "0": outStr += "\n" print(outStr)
epl-1.0
ibm-watson-iot/iot-python
src/wiotp/sdk/device/managedClient.py
2
27070
# ***************************************************************************** # Copyright (c) 2014, 2018 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** from datetime import datetime import json import logging import threading import pytz import uuid from wiotp.sdk import ConnectionException, ConfigurationException from wiotp.sdk.device.client import DeviceClient from wiotp.sdk.device.deviceInfo import DeviceInfo from wiotp.sdk.device.deviceFirmware import DeviceFirmware class ManagedDeviceClient(DeviceClient): # Publish MQTT topics MANAGE_TOPIC = "iotdevice-1/mgmt/manage" UNMANAGE_TOPIC = "iotdevice-1/mgmt/unmanage" UPDATE_LOCATION_TOPIC = "iotdevice-1/device/update/location" ADD_ERROR_CODE_TOPIC = "iotdevice-1/add/diag/errorCodes" CLEAR_ERROR_CODES_TOPIC = "iotdevice-1/clear/diag/errorCodes" NOTIFY_TOPIC = "iotdevice-1/notify" RESPONSE_TOPIC = "iotdevice-1/response" ADD_LOG_TOPIC = "iotdevice-1/add/diag/log" CLEAR_LOG_TOPIC = "iotdevice-1/clear/diag/log" # Subscribe MQTT topics DM_RESPONSE_TOPIC = "iotdm-1/response" DM_OBSERVE_TOPIC = "iotdm-1/observe" DM_REBOOT_TOPIC = "iotdm-1/mgmt/initiate/device/reboot" DM_FACTORY_REESET = "iotdm-1/mgmt/initiate/device/factory_reset" DM_UPDATE_TOPIC = "iotdm-1/device/update" DM_CANCEL_OBSERVE_TOPIC = "iotdm-1/cancel" DM_FIRMWARE_DOWNLOAD_TOPIC = "iotdm-1/mgmt/initiate/firmware/download" DM_FIRMWARE_UPDATE_TOPIC = "iotdm-1/mgmt/initiate/firmware/update" DME_ACTION_TOPIC = "iotdm-1/mgmt/custom/#" # ResponceCode RESPONSECODE_FUNCTION_NOT_SUPPORTED = 501 RESPONSECODE_ACCEPTED = 202 RESPONSECODE_INTERNAL_ERROR = 500 RESPONSECODE_BAD_REQUEST = 400 UPDATESTATE_IDLE = 0 UPDATESTATE_DOWNLOADING = 1 UPDATESTATE_DOWNLOADED = 2 UPDATESTATE_SUCCESS = 0 UPDATESTATE_IN_PROGRESS = 1 UPDATESTATE_OUT_OF_MEMORY = 2 UPDATESTATE_CONNECTION_LOST = 3 UPDATESTATE_VERIFICATION_FAILED = 4 UPDATESTATE_UNSUPPORTED_IMAGE = 5 UPDATESTATE_INVALID_URI = 6 def __init__(self, config, logHandlers=None, deviceInfo=None): if config["identity"]["orgId"] == "quickstart": raise ConfigurationException("QuickStart does not support device management") DeviceClient.__init__(self, config, logHandlers) # Initialize user supplied callback self.deviceActionCallback = None self.firmwereActionCallback = None self.dmeActionCallback = None messages_callbacks = ( ("iotdm-1/#", self.__onDeviceMgmtResponse), (ManagedDeviceClient.DM_REBOOT_TOPIC, self.__onRebootRequest), (ManagedDeviceClient.DM_FACTORY_REESET, self.__onFactoryResetRequest), (ManagedDeviceClient.DM_FIRMWARE_UPDATE_TOPIC, self.__onFirmwereUpdate), (ManagedDeviceClient.DM_OBSERVE_TOPIC, self.__onFirmwereObserve), (ManagedDeviceClient.DM_FIRMWARE_DOWNLOAD_TOPIC, self.__onFirmwereDownload), (ManagedDeviceClient.DM_UPDATE_TOPIC, self.__onUpdatedDevice), (ManagedDeviceClient.DM_CANCEL_OBSERVE_TOPIC, self.__onFirmwereCancel), (ManagedDeviceClient.DME_ACTION_TOPIC, self.__onDMEActionRequest), ) # Add handler for supported device management commands for message, callback in messages_callbacks: self.client.message_callback_add(message, callback) # Initialize user supplied callback self.client.on_subscribe = self._onSubscribe self.client.on_disconnect = self._onDisconnect self.readyForDeviceMgmt = threading.Event() # List of DM requests that have not received a response yet self._deviceMgmtRequestsPendingLock = threading.Lock() self._deviceMgmtRequestsPending = {} # List of DM notify hook self._deviceMgmtObservationsLock = threading.Lock() self._deviceMgmtObservations = [] # Initialize local device data model self.metadata = {} if deviceInfo is not None: self._deviceInfo = deviceInfo else: self._deviceInfo = DeviceInfo() self._location = None self._errorCode = None self.__firmwareUpdate = None self.manageTimer = None # Register startup subscription list self._subscriptions[self.DM_RESPONSE_TOPIC] = 1 self._subscriptions[self.DM_OBSERVE_TOPIC] = 1 self._subscriptions[self.DM_REBOOT_TOPIC] = 1 self._subscriptions[self.DM_FACTORY_REESET] = 1 self._subscriptions[self.DM_UPDATE_TOPIC] = 1 self._subscriptions[self.DM_FIRMWARE_UPDATE_TOPIC] = 1 self._subscriptions[self.DM_FIRMWARE_DOWNLOAD_TOPIC] = 1 self._subscriptions[self.DM_CANCEL_OBSERVE_TOPIC] = 1 self._subscriptions[self._COMMAND_TOPIC] = 1 self._subscriptions[self.DME_ACTION_TOPIC] = 1 def setProperty(self, name, value): if name not in [ "serialNumber", "manufacturer", "model", "deviceClass", "description", "fwVersion", "hwVersion", "descriptiveLocation", ]: raise Exception("Unsupported property name: %s" % name) self._deviceInfo[name] = value return self.notifyFieldChange("deviceInfo.%s" % name, value) def notifyFieldChange(self, field, value): with self._deviceMgmtObservationsLock: if field in self._deviceMgmtObservations: if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning( "Unable to notify service of field " "change because device is not ready " "for device management" ) return threading.Event().set() reqId = str(uuid.uuid4()) message = {"d": {"field": field, "value": value}, "reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.NOTIFY_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.NOTIFY_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent else: return threading.Event().set() def _onSubscribe(self, mqttc, userdata, mid, granted_qos): super(ManagedDeviceClient, self)._onSubscribe(mqttc, userdata, mid, granted_qos) # Once IoTF acknowledges the subscriptions we are able to process commands and responses from device management server self.manage() def manage( self, lifetime=3600, supportDeviceActions=True, supportFirmwareActions=True, supportDeviceMgmtExtActions=False, bundleIds=[], ): # TODO: throw an error, minimum lifetime this client will support is 1 hour, but for now set lifetime to infinite if it's invalid if lifetime < 3600: lifetime = 0 if not self.subscriptionsAcknowledged.wait(timeout=10): self.logger.warning( "Unable to send register for device " "management because device subscriptions " "are not in place" ) return threading.Event().set() reqId = str(uuid.uuid4()) message = { "d": { "lifetime": lifetime, "supports": {"deviceActions": supportDeviceActions, "firmwareActions": supportFirmwareActions}, "deviceInfo": self._deviceInfo.__dict__, "metadata": self.metadata, }, "reqId": reqId, } if supportDeviceMgmtExtActions and len(bundleIds) > 0: for bundleId in bundleIds: message["d"]["supports"][bundleId] = supportDeviceMgmtExtActions resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.MANAGE_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.MANAGE_TOPIC, "message": message, "event": resolvedEvent, } # Register the future call back to Watson IoT Platform 2 minutes before the device lifetime expiry if lifetime != 0: if self.manageTimer is not None: self.logger.debug("Cancelling existing manage timer") self.manageTimer.cancel() self.manageTimer = threading.Timer( lifetime - 120, self.manage, [lifetime, supportDeviceActions, supportFirmwareActions, supportDeviceMgmtExtActions, bundleIds], ) self.manageTimer.start() return resolvedEvent def unmanage(self): if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning( "Unable to set device to unmanaged because " "device is not ready for device management" ) return threading.Event().set() reqId = str(uuid.uuid4()) message = {"reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.UNMANAGE_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.UNMANAGE_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def setLocation(self, longitude, latitude, elevation=None, accuracy=None): # TODO: Add validation (e.g. ensure numeric values) if self._location is None: self._location = {} self._location["longitude"] = longitude self._location["latitude"] = latitude if elevation: self._location["elevation"] = elevation self._location["measuredDateTime"] = datetime.now(pytz.timezone("UTC")).isoformat() if accuracy: self._location["accuracy"] = accuracy elif "accuracy" in self._location: del self._location["accuracy"] if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning( "Unable to publish device location because " "device is not ready for device management" ) return threading.Event().set() reqId = str(uuid.uuid4()) message = {"d": self._location, "reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.UPDATE_LOCATION_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.UPDATE_LOCATION_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def setErrorCode(self, errorCode=0): if errorCode is None: errorCode = 0 self._errorCode = errorCode if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning("Unable to publish error code because " "device is not ready for device management") return threading.Event().set() reqId = str(uuid.uuid4()) message = {"d": {"errorCode": errorCode}, "reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.ADD_ERROR_CODE_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.ADD_ERROR_CODE_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def clearErrorCodes(self): self._errorCode = None if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning("Unable to clear error codes because " "device is not ready for device management") return threading.Event().set() reqId = str(uuid.uuid4()) message = {"reqId": reqId} resolvedEvent = threading.Event() self.client.publish( ManagedDeviceClient.CLEAR_ERROR_CODES_TOPIC, payload=json.dumps(message), qos=1, retain=False ) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.CLEAR_ERROR_CODES_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def addLog(self, msg="", data="", sensitivity=0): timestamp = datetime.now().isoformat() if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning("Unable to publish error code because " "device is not ready for device management") return threading.Event().set() reqId = str(uuid.uuid4()) message = {"d": {"message": msg, "timestamp": timestamp, "data": data, "severity": sensitivity}, "reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.ADD_LOG_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.ADD_LOG_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def clearLog(self): if not self.readyForDeviceMgmt.wait(timeout=10): self.logger.warning("Unable to clear log because device is not ready for device management") return threading.Event().set() reqId = str(uuid.uuid4()) message = {"reqId": reqId} resolvedEvent = threading.Event() self.client.publish(ManagedDeviceClient.CLEAR_LOG_TOPIC, payload=json.dumps(message), qos=1, retain=False) with self._deviceMgmtRequestsPendingLock: self._deviceMgmtRequestsPending[reqId] = { "topic": ManagedDeviceClient.CLEAR_LOG_TOPIC, "message": message, "event": resolvedEvent, } return resolvedEvent def __onDeviceMgmtResponse(self, client, userdata, pahoMessage): try: data = json.loads(pahoMessage.payload.decode("utf-8")) if "rc" not in data: return True rc = data["rc"] reqId = data["reqId"] except ValueError as e: raise Exception('Unable to parse JSON. payload="%s" error=%s' % (pahoMessage.payload, str(e))) else: request = None with self._deviceMgmtRequestsPendingLock: try: request = self._deviceMgmtRequestsPending.pop(reqId) except KeyError: self.logger.warning("Received unexpected response from " "device management: %s", reqId) else: self.logger.debug( "Remaining unprocessed device " "management requests: %s", len(self._deviceMgmtRequestsPending) ) if request is None: return False state = { ManagedDeviceClient.MANAGE_TOPIC: { # rc, json.dumps(request['message']) "msg_succ": "[%s] Manage action completed: %s", "msg_fail": "[%s] Manage action failed: %s", }, ManagedDeviceClient.UNMANAGE_TOPIC: { "msg_succ": "[%s] Unmanage action completed: %s", "msg_fail": "[%s] Unmanage action failed: %s", }, ManagedDeviceClient.UPDATE_LOCATION_TOPIC: { "msg_succ": "[%s] Location update action completed: %s", "msg_fail": "[%s] Location update action failed: %s", }, ManagedDeviceClient.ADD_ERROR_CODE_TOPIC: { "msg_succ": "[%s] Add error code action completed: %s", "msg_fail": "[%s] Add error code action failed: %s", }, ManagedDeviceClient.CLEAR_ERROR_CODES_TOPIC: { "msg_succ": "[%s] Clear error codes action completed: %s", "msg_fail": "[%s] Clear error codes action failed: %s", }, ManagedDeviceClient.ADD_LOG_TOPIC: { "msg_succ": "[%s] Add log action completed: %s", "msg_fail": "[%s] Add log action failed: %s", }, ManagedDeviceClient.CLEAR_LOG_TOPIC: { "msg_succ": "[%s] Clear log action completed: %s", "msg_fail": "[%s] Clear log action failed: %s", }, } try: msg_succ = state[request["topic"]]["msg_succ"] msg_fail = state[request["topic"]]["msg_fail"] except Exception as e: self.logger.warning("[%s] Unknown action response: %s", rc, json.dumps(request["message"])) else: dump_str = json.dumps(request["message"]) if rc == 200: self.logger.info(msg_succ, rc, dump_str) else: self.logger.critical(msg_fail, rc, dump_str) if request["topic"] == ManagedDeviceClient.MANAGE_TOPIC: self.readyForDeviceMgmt.set() elif request["topic"] == ManagedDeviceClient.UNMANAGE_TOPIC: self.readyForDeviceMgmt.clear() # Now clear the event, allowing anyone that was waiting on this to proceed request["event"].set() return True # Device Action Handlers def __onRebootRequest(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_REBOOT_TOPIC, paho_payload ) try: data = json.loads(paho_payload) reqId = data["reqId"] if self.deviceActionCallback: self.deviceActionCallback(reqId, "reboot") except ValueError as e: raise Exception('Unable to process Reboot request. payload="%s" error=%s' % (pahoMessage.payload, str(e))) def __onFactoryResetRequest(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_FACTORY_REESET, paho_payload ) try: data = json.loads(paho_payload) reqId = data["reqId"] if self.deviceActionCallback: self.deviceActionCallback(reqId, "reset") except ValueError as e: raise Exception( 'Unable to process Factory Reset request. payload="%s" error=%s' % (pahoMessage.payload, str(e)) ) def respondDeviceAction(self, reqId, responseCode=202, message=""): response = {"rc": responseCode, "message": message, "reqId": reqId} payload = json.dumps(response) self.logger.info("Publishing Device Action response with payload :%s", payload) self.client.publish("iotdevice-1/response", payload, qos=1, retain=False) # Firmware Handlers def __onFirmwereDownload(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_FIRMWARE_DOWNLOAD_TOPIC, paho_payload, ) data = json.loads(paho_payload) reqId = data["reqId"] rc = ManagedDeviceClient.RESPONSECODE_ACCEPTED msg = "" if self.__firmwareUpdate.state != ManagedDeviceClient.UPDATESTATE_IDLE: rc = ManagedDeviceClient.RESPONSECODE_BAD_REQUEST msg = "Cannot download as the device is not in idle state" thread = threading.Thread(target=self.respondDeviceAction, args=(reqId, rc, msg), name="respondDeviceAction") thread.start() if self.firmwereActionCallback: self.firmwereActionCallback("download", self.__firmwareUpdate) def __onFirmwereCancel(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_CANCEL_OBSERVE_TOPIC, paho_payload ) data = json.loads(paho_payload) reqId = data["reqId"] thread = threading.Thread(target=self.respondDeviceAction, args=(reqId, 200, ""), name="respondDeviceAction") thread.start() def __onFirmwereObserve(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_OBSERVE_TOPIC, paho_payload ) data = json.loads(paho_payload) reqId = data["reqId"] # TODO: Proprer validation for fields in payload thread = threading.Thread(target=self.respondDeviceAction, args=(reqId, 200, ""), name="respondDeviceAction") thread.start() def __onUpdatedDevice(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_UPDATE_TOPIC, paho_payload ) data = json.loads(paho_payload) if "reqId" in data: reqId = data["reqId"] d = data["d"] value = None for obj in d["fields"]: if "field" in obj: if obj["field"] == "mgmt.firmware": value = obj["value"] if value is not None: self.__firmwareUpdate = DeviceFirmware( value["version"], value["name"], value["uri"], value["verifier"], value["state"], value["updateStatus"], value["updatedDateTime"], ) thread = threading.Thread( target=self.respondDeviceAction, args=(reqId, 204, ""), name="respondDeviceAction" ) thread.start() else: d = data["d"] value = None for obj in d["fields"]: if "field" in obj: if obj["field"] == "metadata": value = obj["value"] if value is not None: self.metadata = value def setState(self, status): notify = {"d": {"fields": [{"field": "mgmt.firmware", "value": {"state": status}}]}} if self.__firmwareUpdate is not None: self.__firmwareUpdate.state = status self.logger.info("Publishing state Update with payload :%s", json.dumps(notify)) thread = threading.Thread( target=self.client.publish, args=("iotdevice-1/notify", json.dumps(notify), 1, False), name="client.publish" ) thread.start() def setUpdateStatus(self, status): notify = { "d": { "fields": [ { "field": "mgmt.firmware", "value": {"state": ManagedDeviceClient.UPDATESTATE_IDLE, "updateStatus": status}, } ] } } if self.__firmwareUpdate is not None: self.__firmwareUpdate.state = ManagedDeviceClient.UPDATESTATE_IDLE self.__firmwareUpdate.updateStatus = status self.logger.info("Publishing Update Status with payload :%s", json.dumps(notify)) thread = threading.Thread( target=self.client.publish, args=("iotdevice-1/notify", json.dumps(notify), 1, False), name="client.publish" ) thread.start() def __onFirmwereUpdate(self, client, userdata, pahoMessage): paho_payload = pahoMessage.payload.decode("utf-8") self.logger.info( "Message received on topic :%s with payload %s", ManagedDeviceClient.DM_FIRMWARE_UPDATE_TOPIC, paho_payload ) data = json.loads(paho_payload) reqId = data["reqId"] rc = ManagedDeviceClient.RESPONSECODE_ACCEPTED msg = "" if self.__firmwareUpdate.state != ManagedDeviceClient.UPDATESTATE_DOWNLOADED: rc = ManagedDeviceClient.RESPONSECODE_BAD_REQUEST msg = "Firmware is still not successfully downloaded." thread = threading.Thread(target=self.respondDeviceAction, args=(reqId, rc, msg), name="respondDeviceAction") thread.start() if self.firmwereActionCallback: self.firmwereActionCallback("update", self.__firmwareUpdate) def __onDMEActionRequest(self, client, userdata, pahoMessage): data = json.loads(pahoMessage.payload.decode("utf-8")) self.logger.info("Message received on topic :%s with payload %s", ManagedDeviceClient.DME_ACTION_TOPIC, data) reqId = data["reqId"] if self.dmeActionCallback: if self.dmeActionCallback(pahoMessage.topic, data, reqId): msg = "DME Action successfully completed from Callback" thread = threading.Thread( target=self.respondDeviceAction, args=(reqId, 200, msg), name="respondDeviceAction" ) thread.start() else: msg = "Unexpected device error" thread = threading.Thread( target=self.respondDeviceAction, args=(reqId, 500, msg), name="respondDeviceAction" ) thread.start() else: thread = threading.Thread( target=self.respondDeviceAction, args=(reqId, 501, "Operation not implemented"), name="respondDeviceAction", ) thread.start()
epl-1.0
ibm-watson-iot/iot-python
test/test_device.py
2
9913
# ***************************************************************************** # Copyright (c) 2016-2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import copy import uuid import testUtils import pytest import wiotp.sdk.device import wiotp.sdk.application from wiotp.sdk.exceptions import ApiException, ConnectionException, MissingMessageEncoderException class TestDevice(testUtils.AbstractTest): def testDeviceInfoInstance(self): deviceInfoObj = wiotp.sdk.device.DeviceInfo() deviceInfoObj.serialNumber = "0101" deviceInfoObj.manufacturer = "0202" deviceInfoObj.model = "0303" deviceInfoObj.deviceClass = "0404" deviceInfoObj.description = "0505" deviceInfoObj.fwVersion = "0606" deviceInfoObj.hwVersion = "0707" deviceInfoObj.descriptiveLocation = "0808" assert isinstance(deviceInfoObj, wiotp.sdk.device.DeviceInfo) print(deviceInfoObj) assert deviceInfoObj.serialNumber == "0101" assert deviceInfoObj.manufacturer == "0202" assert deviceInfoObj.model == "0303" assert deviceInfoObj.deviceClass == "0404" assert deviceInfoObj.description == "0505" assert deviceInfoObj.fwVersion == "0606" assert deviceInfoObj.hwVersion == "0707" assert deviceInfoObj.descriptiveLocation == "0808" def testDeviceFirmwareInstance(self): deviceFWObj = wiotp.sdk.device.DeviceFirmware() assert isinstance(deviceFWObj, wiotp.sdk.device.DeviceFirmware) print(deviceFWObj) def testNotAuthorizedConnect(self, device): client = wiotp.sdk.device.DeviceClient( { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": "MGhUixxxxxxxxxxxx"}, } ) with pytest.raises(ConnectionException) as e: client.connect() assert e.value.msg == "Not authorized" assert client.isConnected() == False def testMissingMessageEncoder(self, device): options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, } deviceClient = wiotp.sdk.device.DeviceClient(options) assert isinstance(deviceClient, wiotp.sdk.device.DeviceClient) with pytest.raises(MissingMessageEncoderException) as e: myData = {"name": "foo", "cpu": 60, "mem": 50} deviceClient.connect() deviceClient.publishEvent("missingMsgEncode", "jason", myData) def testPublishEvent(self, device): def devPublishCallback(): print("Device Publish Event done!!!") options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, } deviceClient = wiotp.sdk.device.DeviceClient(options) myData = {"name": "foo", "cpu": 60, "mem": 50} deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort1883(self, device): def devPublishCallback(): print("Device Publish Event done!!!") options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 1883}}, } myData = {"name": "foo", "cpu": 60, "mem": 50} deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() @pytest.mark.skip(reason="Currently, port 80 only works with websockets, not tcp :/") def testPublishEventPort80(self, device): def devPublishCallback(): print("Device Publish Event done!!!") options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 80}}, } myData = {"name": "foo", "cpu": 60, "mem": 50} deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort80ws(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 80, "transport": "websockets"}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort1883ws(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 1883, "transport": "websockets"}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort8883(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 8883}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort8883ws(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 1883, "transport": "websockets"}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort443(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 443}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testPublishEventPort443ws(self, device): def devPublishCallback(): print("Device Publish Event done!!!") myData = {"name": "foo", "cpu": 60, "mem": 50} options = { "identity": {"orgId": self.ORG_ID, "typeId": device.typeId, "deviceId": device.deviceId}, "auth": {"token": device.authToken}, "options": {"mqtt": {"port": 443, "transport": "websockets"}}, } deviceClient = wiotp.sdk.device.DeviceClient(options) deviceClient.connect() assert ( deviceClient.publishEvent("testPublishJsonEvent", "json", myData, onPublish=devPublishCallback, qos=2) == True ) deviceClient.disconnect() def testConfigPortInvalid(self): options = { "identity": {"orgId": self.ORG_ID, "typeId": "xxx", "deviceId": "xxx"}, "auth": {"token": "xxx"}, "options": {"mqtt": {"port": 100}}, } with pytest.raises(Exception) as e: deviceClient = wiotp.sdk.device.DeviceClient(options) assert str(e.value) == "Unsupported value for port override: 100. Supported values are 1883 & 8883."
epl-1.0
ibm-watson-iot/iot-python
src/wiotp/sdk/gateway/messages.py
2
2136
# ***************************************************************************** # Copyright (c) 2019 IBM Corporation and other Contributors. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Eclipse Public License v1.0 # which accompanies this distribution, and is available at # http://www.eclipse.org/legal/epl-v10.html # ***************************************************************************** import re from wiotp.sdk import MissingMessageDecoderException, InvalidEventException COMMAND_RE = re.compile("iot-2/type/(.+)/id/(.+)/cmd/(.+)/fmt/(.+)") class Command: def __init__(self, pahoMessage, messageEncoderModules): result = COMMAND_RE.match(pahoMessage.topic) if result: self.typeId = result.group(1) self.deviceId = result.group(2) self.commandId = result.group(3) self.format = result.group(4) if self.format in messageEncoderModules: message = messageEncoderModules[self.format].decode(pahoMessage) self.timestamp = message.timestamp self.data = message.data else: raise MissingMessageDecoderException(self.format) else: raise InvalidEventException("Received command on invalid topic: %s" % (pahoMessage.topic)) NOTIFY_RE = re.compile("iot-2/type/(.+)/id/(.+)/notify") class Notification: def __init__(self, pahoMessage, messageEncoderModules): result = NOTIFY_RE.match(pahoMessage.topic) if result: self.typeId = result.group(1) self.deviceId = result.group(2) self.format = "json" if self.format in messageEncoderModules: message = messageEncoderModules[self.format].decode(pahoMessage) self.timestamp = message.timestamp self.data = message.data else: raise MissingMessageDecoderException(self.format) else: raise InvalidEventException("Received notification on invalid topic: %s" % (pahoMessage.topic))
epl-1.0
mbj4668/pyang
pyang/transforms/edit.py
1
12718
"""Edit transform plugin This plugin currently has quite limited functionality. Only some specific top-level items can be edited, and only existing statements are edited. """ import copy import optparse import re import sys from pyang import error from pyang import plugin from pyang import statements plugin_name = 'edit' # noinspection PyUnusedLocal def check_date(option, opt, value): if not re.match(r'^\d{4}-\d{2}-\d{2}$', value): raise optparse.OptionValueError( 'option %s: invalid yyyy-mm-dd date: %s' % (opt, value)) return value class EditOption(optparse.Option): TYPES = optparse.Option.TYPES + ('date',) TYPE_CHECKER = copy.copy(optparse.Option.TYPE_CHECKER) TYPE_CHECKER['date'] = check_date def pyang_plugin_init(): plugin.register_plugin(EditPlugin()) class EditPlugin(plugin.PyangPlugin): def add_opts(self, optparser): optlist = [ # set YANG version (this does nothing if there's no yang-version # statement) EditOption("--edit-yang-version", dest="edit_yang_version", metavar="VERSION", help="Set YANG version to the supplied value"), # set namespace (this does nothing if there's no namespace # statement) EditOption("--edit-namespace", dest="edit_namespace", metavar="NAMESPACE", help="Set YANG namespace to the supplied value"), # set imported/included module/submodule revision dates EditOption("--edit-update-import-dates", dest="edit_update_import_dates", default=False, action="store_true", help="Set import/include revision-date " "statements to match imported/included " "modules/submodules"), EditOption("--edit-delete-import-dates", dest="edit_delete_import_dates", default=False, action="store_true", help="Delete import/include revision-date " "statements"), # set meta info (these do nothing if there's no corresponding # metadata statement) EditOption("--edit-organization", dest="edit_organization", metavar="ORGANIZATION", help="Set module/submodule organization " "to the supplied value"), EditOption("--edit-contact", dest="edit_contact", metavar="CONTACT", help="Set module/submodule contact " "to the supplied value"), EditOption("--edit-description", dest="edit_description", metavar="DESCRIPTION", help="Set module/submodule description " "to the supplied value"), # set revision info (these do nothing if there's no revision # statement) EditOption("--edit-delete-revisions-after", dest="edit_delete_revisions_after", type="date", metavar="PREVDATE", help="Delete any revisions after " "the supplied yyyy-mm-dd"), EditOption("--edit-revision-date", dest="edit_revision_date", type="date", metavar="DATE", help="Set most recent revision date " "to the supplied yyyy-mm-dd"), EditOption("--edit-revision-description", dest="edit_revision_description", metavar="DESCRIPTION", help="Set most recent revision description " "to the supplied value"), EditOption("--edit-revision-reference", dest="edit_revision_reference", metavar="REFERENCE", help="Set most recent revision reference " "to the supplied value") ] g = optparser.add_option_group("Edit transform specific options") g.add_options(optlist) def add_transform(self, xforms): xforms[plugin_name] = self def transform(self, ctx, modules): edit_tree(ctx, modules) def edit_tree(ctx, modules): def optval(key): dest = ('%s-%s' % (plugin_name, key)).replace('-', '_') return getattr(ctx.opts, dest, None) for module in modules: for keyword in ['yang-version', 'namespace']: arg = optval(keyword) if arg is not None: update_or_add_stmt(module, keyword, arg) substmts = [] revision_done = False for stmt in module.substmts: replstmts = None if stmt.keyword in ['import', 'include']: # XXX should check that these options aren't both set if ctx.opts.edit_update_import_dates: update_import_date(ctx, stmt) elif ctx.opts.edit_delete_import_dates: delete_import_date(ctx, stmt) elif stmt.keyword in ['organization', 'contact', 'description']: arg = optval(stmt.keyword) if arg is not None: set_meta_details(ctx, stmt, arg) elif stmt.keyword == 'revision' and not revision_done: allrevs = module.search('revision') lastrev = stmt == allrevs[-1] replstmts, revision_done = set_revision_details(ctx, stmt, lastrev) substmts += [stmt] if replstmts is None else replstmts # XXX should we tidy up any of the deleted statements? module.substmts = substmts def update_import_date(ctx, stmt): imprev = stmt.search_one('revision-date') imprevdate = imprev.arg if imprev else None impmod = ctx.get_module(stmt.arg, imprevdate) impmodrev = impmod.search_one('revision') if impmod else None impmodrevdate = impmodrev.arg if impmodrev else None if not imprev or impmodrevdate > imprevdate: update_or_add_stmt(stmt, 'revision-date', impmodrevdate) # noinspection PyUnusedLocal def delete_import_date(ctx, stmt): imprev = stmt.search_one('revision-date') if imprev: delete_stmt(stmt, imprev) # noinspection PyUnusedLocal def set_meta_details(ctx, stmt, arg): (newarg, ignore) = get_arg_value(arg, stmt.arg) if newarg is not None: stmt.arg = newarg # XXX note that this logic relies on there already being at least one # revision statement; --lint checks this so it should be OK def set_revision_details(ctx, stmt, lastrev): revision_done = False # relevant options opts = { 'olddate': ctx.opts.edit_delete_revisions_after, 'newdate': ctx.opts.edit_revision_date, 'description': ctx.opts.edit_revision_description, 'reference': ctx.opts.edit_revision_reference } # the logic is quite tricky; here's what we want to achieve: # * 'olddate' is the date of the oldest revision to be retained; if not # supplied, any existing revisions are deleted # * if 'newdate' is supplied, it's the date of the next published # revision and is to be inserted at the start of any remaining # revisions # * reuse rather than delete the oldest revision statement, purely in # order to retain any blank lines after it # default action is to do nothing action = '' #sys.stderr.write('revision %s (lastrev %s)\n' % (stmt.arg, lastrev)) # only adjust revisions if either olddate or newdate is supplied olddate = opts.get('olddate', None) newdate = opts.get('newdate', None) if olddate is not None or newdate is not None: # determine whether to delete this old revision if olddate is None or stmt.arg > olddate: action = 'delete' #sys.stderr.write('-> delete (olddate %s)\n' % olddate) # determine whether to insert the new revision if newdate is not None and (action != 'delete' or lastrev): action = 'replace' if action == 'delete' else 'insert' #sys.stderr.write('-> %s (newdate %s)\n' % (action, newdate)) # if deleting, return an empty list replstmts = None if action == 'delete': replstmts = [] # replace and insert logic is quite similar: # * if replacing, modify this statement and return a list containing # only it # * if inserting, create a new statement and return a list containing # the new and the original statement elif action == 'replace' or action == 'insert': if action == 'replace': revstmt = stmt revstmt.arg = newdate else: revstmt = statements.new_statement(stmt.top, stmt.parent, None, 'revision', newdate) other_keywords = set(opts.keys()) - {'olddate', 'newdate'} for keyword in other_keywords: update_or_add_stmt(revstmt, keyword, opts[keyword]) if action == 'replace': replstmts = [revstmt] else: replstmts = [revstmt, stmt] revision_done = True #sys.stderr.write( # '= %s\n' % ([s.arg for s in replstmts] if replstmts else None)) return replstmts, revision_done def get_arg_value(arg, currarg=None): if arg is None or arg[0] not in ['%', '@']: return arg, True else: replace = False try: argval = '' specs = arg.split('+') for spec in specs: if argval != '': argval += '\n\n' if spec[0] not in ['%', '@']: argval += spec elif spec[0] == '%': if spec == '%SUMMARY': summary = get_arg_summary(currarg) if summary: argval += summary elif spec.startswith('%SUBST/'): (ignore, old, new) = spec.split('/') if currarg is None: if argval == '': argval = None else: argval = currarg.replace(old, new) replace = True elif spec == '%DELETE': argval = '' replace = True else: argval += spec elif spec[0] == '@': argval += open(spec[1:], 'r').read().rstrip() return argval, replace except IOError as e: raise error.EmitError(str(e)) def get_arg_summary(arg): lines = arg.splitlines() summary = '' prev = '' discard_prev = False for line in lines: if line.strip().startswith('Copyright '): if prev.strip() == '': discard_prev = True break if prev != '': summary += prev prev = '' if summary != '': prev += '\n' prev += line if prev and not discard_prev: summary += prev return summary if summary else 'TBD' # XXX should insert in canonical order; currently (apart from the hack noted # below) just appending; should look into doing the same as yang.py, which # does: substmts = grammar.sort_canonical(stmt.keyword, stmt.substmts) def update_or_add_stmt(stmt, keyword, arg, index=None): child = stmt.search_one(keyword) currarg = child.arg if child else None (argval, replace) = get_arg_value(arg, currarg) if argval is None: child = None elif child: if not replace and child.arg and child.arg != argval and child.arg \ != 'TBD': sys.stderr.write('%s: not replacing existing %s %r with %r\n' % ( child.pos, keyword, child.arg, argval)) else: child.arg = argval else: child = statements.new_statement(stmt.top, stmt, None, keyword, argval) if index is None: index = len(stmt.substmts) # XXX this hack ensures that 'reference' is always last if index > 0 and stmt.substmts[index - 1].keyword == 'reference': index -= 1 stmt.substmts.insert(index, child) return child def delete_stmt(parent, stmt): if stmt in parent.substmts: idx = parent.substmts.index(stmt) del parent.substmts[idx] del stmt
isc
mbj4668/pyang
pyang/syntax.py
1
15064
"""Description of YANG & YIN syntax.""" import os import re import shlex import sys import datetime ### Regular expressions - constraints on arguments # keywords and identifiers identifier = r"[_A-Za-z][._\-A-Za-z0-9]*" prefix = identifier keyword = '((' + prefix + '):)?(' + identifier + ')' comment = r'(/\*([^*]|[\r\n\s]|(\*+([^*/]|[\r\n\s])))*\*+/)|(//.*)|(/\*.*)' # no group version of keyword keyword_ng = '(?:(' + prefix + '):)?(?:' + identifier + ')' re_keyword = re.compile(keyword) re_keyword_start = re.compile('^' + keyword) re_comment = re.compile(comment) pos_integer = r"[1-9][0-9]*" nonneg_integer = r"(0|([1-9][0-9]*))" integer_ = r"[+-]?" + nonneg_integer decimal_ = integer_ + r"(\.[0-9]+)?" length_str = r'((min|max|[0-9]+)\s*' \ r'(\.\.\s*' \ r'(min|max|[0-9]+)\s*)?)' length_expr = length_str + r'(\|\s*' + length_str + r')*' re_length_part = re.compile(length_str) range_str = r'((min|max|((\+|\-)?[0-9]+(\.[0-9]+)?))\s*' \ r'(\.\.\s*' \ r'(min|max|(\+|\-)?[0-9]+(\.[0-9]+)?)\s*)?)' range_expr = range_str + r'(\|\s*' + range_str + r')*' re_range_part = re.compile(range_str) re_identifier = re.compile("^" + identifier + "$") # path and unique node_id = keyword_ng rel_path_keyexpr = r"(\.\./)+(" + node_id + "/)*" + node_id path_key_expr = r"(current\s*\(\s*\)/" + rel_path_keyexpr + ")" path_equality_expr = node_id + r"\s*=\s*" + path_key_expr path_predicate = r"\s*\[\s*" + path_equality_expr + r"\s*\]\s*" absolute_path_arg = "(?:/" + node_id + "(" + path_predicate + ")*)+" descendant_path_arg = node_id + "(" + path_predicate + ")*" + \ "(?:" + absolute_path_arg + ")?" relative_path_arg = r"(\.\./)*" + descendant_path_arg deref_path_arg = r"deref\s*\(\s*(?:" + relative_path_arg + \ r")\s*\)/\.\./" + relative_path_arg path_arg = "(" + absolute_path_arg + "|" + relative_path_arg + "|" + \ deref_path_arg + ")" absolute_schema_nodeid = "(/" + node_id + ")+" descendant_schema_nodeid = node_id + "(" + absolute_schema_nodeid + ")?" schema_nodeid = "("+absolute_schema_nodeid+"|"+descendant_schema_nodeid+")" unique_arg = descendant_schema_nodeid + \ r"(\s+" + descendant_schema_nodeid + r")*" key_arg = node_id + r"(\s+" + node_id + r")*" re_schema_node_id_part = re.compile('/' + keyword) # URI - RFC 3986, Appendix A scheme = "[A-Za-z][-+.A-Za-z0-9]*" unreserved = "[-._~A-Za-z0-9]" pct_encoded = "%[0-9A-F]{2}" sub_delims = "[!$&'()*+,;=]" pchar = ("(" + unreserved + "|" + pct_encoded + "|" + sub_delims + "|[:@])") segment = pchar + "*" segment_nz = pchar + "+" userinfo = ("(" + unreserved + "|" + pct_encoded + "|" + sub_delims + "|:)*") dec_octet = "([0-9]|[1-9][0-9]|1[0-9]{2}|2[0-4][0-9]|25[0-5])" ipv4address = "(" + dec_octet + r"\.){3}" + dec_octet h16 = "[0-9A-F]{1,4}" ls32 = "(" + h16 + ":" + h16 + "|" + ipv4address + ")" ipv6address = ( "((" + h16 + ":){6}" + ls32 + "|::(" + h16 + ":){5}" + ls32 + "|(" + h16 + ")?::(" + h16 + ":){4}" + ls32 + "|((" + h16 + ":)?" + h16 + ")?::(" + h16 + ":){3}" + ls32 + "|((" + h16 + ":){,2}" + h16 + ")?::(" + h16 + ":){2}" + ls32 + "|((" + h16 + ":){,3}" + h16 + ")?::" + h16 + ":" + ls32 + "|((" + h16 + ":){,4}" + h16 + ")?::" + ls32 + "|((" + h16 + ":){,5}" + h16 + ")?::" + h16 + "|((" + h16 + ":){,6}" + h16 + ")?::)") ipvfuture = r"v[0-9A-F]+\.(" + unreserved + "|" + sub_delims + "|:)+" ip_literal = r"\[(" + ipv6address + "|" + ipvfuture + r")\]" reg_name = "(" + unreserved + "|" + pct_encoded + "|" + sub_delims + ")*" host = "(" + ip_literal + "|" + ipv4address + "|" + reg_name + ")" port = "[0-9]*" authority = "(" + userinfo + "@)?" + host + "(:" + port + ")?" path_abempty = "(/" + segment + ")*" path_absolute = "/(" + segment_nz + "(/" + segment + ")*)?" path_rootless = segment_nz + "(/" + segment + ")*" path_empty = pchar + "{0}" hier_part = ("(" + "//" + authority + path_abempty + "|" + path_absolute + "|" + path_rootless + "|" + path_empty + ")") query = "(" + pchar + "|[/?])*" fragment = query uri = (scheme + ":" + hier_part + r"(\?" + query + ")?" + "(#" + fragment + ")?") # Date date = r"([1-2][0-9]{3})-(0[1-9]|1[012])-(0[1-9]|[12][0-9]|3[01])" re_nonneg_integer = re.compile("^" + nonneg_integer + "$") re_integer = re.compile("^" + integer_ + "$") re_decimal = re.compile("^" + decimal_ + "$") re_uri = re.compile("^" + uri + "$") re_boolean = re.compile(r"^(true|false)$") re_version = re.compile(r"^(1|(1\.1))$") re_date = re.compile("^" + date +"$") re_status = re.compile(r"^(current|obsolete|deprecated)$") re_key = re.compile("^" + key_arg + "$") re_length = re.compile("^" + length_expr + "$") re_range = re.compile("^" + range_expr + "$") re_pos_integer = re.compile(r"^(unbounded|" + pos_integer + r")$") re_ordered_by = re.compile(r"^(user|system)$") re_modifier = re.compile(r"^(invert-match)$") re_node_id = re.compile("^" + node_id + "$") re_path = re.compile("^" + path_arg + "$") re_absolute_path = re.compile("^" + absolute_path_arg + "$") re_unique = re.compile("^" + unique_arg + "$") re_schema_nodeid = re.compile("^" + schema_nodeid + "$") re_absolute_schema_nodeid = re.compile("^" + absolute_schema_nodeid + "$") re_descendant_schema_nodeid = re.compile("^" + descendant_schema_nodeid + "$") re_deviate = re.compile(r"^(add|delete|replace|not-supported)$") # Not part of YANG syntax per se but useful for pyang in several places re_filename = re.compile( r"^(?:.*" + re.escape(os.sep) + r")?" + # ignore all before os.sep r"([^@]*?)" + # putative module name r"(?:@([^.]*?))?" + # putative revision r"(?:\.yang|\.yin)*" + # foo@bar.yang.yin.yang.yin ? r"\.(yang|yin)$") # actual final extension arg_type_map = { "identifier": lambda s: re_identifier.search(s) is not None, "non-negative-integer": lambda s: re_nonneg_integer.search(s) is not None, "integer": lambda s: re_integer.search(s) is not None, "uri": lambda s: re_uri.search(s) is not None, "boolean": lambda s: re_boolean.search(s) is not None, "version": lambda s: re_version.search(s) is not None, "date": lambda s: chk_date_arg(s), "status-arg": lambda s: re_status.search(s) is not None, "key-arg": lambda s: re_key.search(s) is not None, "length-arg": lambda s: re_length.search(s) is not None, "range-arg": lambda s: re_range.search(s) is not None, "max-value": lambda s: re_pos_integer.search(s) is not None, "ordered-by-arg": lambda s: re_ordered_by.search(s) is not None, "modifier-arg": lambda s: re_modifier.search(s) is not None, "identifier-ref": lambda s: re_node_id.search(s) is not None, "path-arg": lambda s: re_path.search(s) is not None, "absolute-path-arg": lambda s: re_absolute_path.search(s) is not None, "unique-arg": lambda s: re_unique.search(s) is not None, "absolute-schema-nodeid": lambda s: \ re_absolute_schema_nodeid.search(s) is not None, "descendant-schema-nodeid": lambda s: \ re_descendant_schema_nodeid.search(s) is not None, "schema-nodeid": lambda s: \ re_schema_nodeid.search(s) is not None, "enum-arg": lambda s: chk_enum_arg(s), "fraction-digits-arg": lambda s: chk_fraction_digits_arg(s), "if-feature-expr": lambda s: chk_if_feature_expr(s), "deviate-arg": lambda s: re_deviate.search(s) is not None, "_comment": lambda s: re_comment.search(s) is not None, } """Argument type definitions. Regular expressions for all argument types except plain string that are checked directly by the parser. """ def chk_date_arg(s): """Checks if the string `s` is a valid date string. Return True of False.""" match = re_date.match(s) if match is None: return False comp = match.groups() try: datetime.date(int(comp[0]), int(comp[1]), int(comp[2])) return True except ValueError: return False def chk_enum_arg(s): """Checks if the string `s` is a valid enum string. Return True or False.""" if len(s) == 0 or s[0].isspace() or s[-1].isspace(): return False else: return True def chk_fraction_digits_arg(s): """Checks if the string `s` is a valid fraction-digits argument. Return True or False.""" try: v = int(s) if v >= 1 and v <= 18: return True else: return False except ValueError: return False def chk_if_feature_expr(s): return parse_if_feature_expr(s) is not None # if-feature-expr = "(" if-feature-expr ")" / # if-feature-expr sep boolean-operator sep # if-feature-expr / # not-keyword sep if-feature-expr / # identifier-ref-arg # # Rewrite to: # x = y ("and"/"or" y)* # y = "not" x / # "(" x ")" # identifier # # Expr :: ('not', Expr, None) # | ('and'/'or', Expr, Expr) # | Identifier def parse_if_feature_expr(s): try: # Encoding to ascii works for valid if-feature-exprs, since all # pars are YANG identifiers (or the boolean keywords). # The reason for this fix is that in Python < 2.7.3, shlex would return # erroneous tokens if a unicode string was passed. # Also, shlex uses cStringIO internally which doesn't handle unicode # characters outside the ascii range anyway. if sys.version < '3': sx = shlex.shlex(s.encode("ascii")) else: sx = shlex.shlex(s) except UnicodeEncodeError: return None sx.wordchars += ":-" # need to handle prefixes and '-' in the name operators = [None] operands = [] precedence = {'not':3, 'and':2, 'or':1, None:0} def x(): y() tok = sx.get_token() while tok in ('and', 'or'): push_operator(tok) y() tok = sx.get_token() sx.push_token(tok) while operators[-1] is not None: pop_operator() def y(): tok = sx.get_token() if tok == 'not': push_operator(tok) x() elif tok == '(': operators.append(None) x() tok = sx.get_token() if tok != ')': raise ValueError operators.pop() elif is_identifier(tok): operands.append(tok) else: raise ValueError def push_operator(op): while op_gt(operators[-1], op): pop_operator() operators.append(op) def pop_operator(): op = operators.pop() if op == 'not': operands.append((op, operands.pop(), None)) else: operands.append((op, operands.pop(), operands.pop())) def op_gt(op1, op2): return precedence[op1] > precedence[op2] def is_identifier(tok): return re_node_id.search(tok) is not None try: x() if sx.get_token() != '': raise ValueError return operands[-1] except ValueError: return None def add_arg_type(arg_type, regexp): """Add a new arg_type to the map. Used by extension plugins to register their own argument types.""" arg_type_map[arg_type] = regexp # keyword argument-name yin-element yin_map = \ {'action': ('name', False), 'anydata': ('name', False), 'anyxml': ('name', False), 'argument': ('name', False), 'augment': ('target-node', False), 'base': ('name', False), 'belongs-to': ('module', False), 'bit': ('name', False), 'case': ('name', False), 'choice': ('name', False), 'config': ('value', False), 'contact': ('text', True), 'container': ('name', False), 'default': ('value', False), 'description': ('text', True), 'deviate': ('value', False), 'deviation': ('target-node', False), 'enum': ('name', False), 'error-app-tag': ('value', False), 'error-message': ('value', True), 'extension': ('name', False), 'feature': ('name', False), 'fraction-digits': ('value', False), 'grouping': ('name', False), 'identity': ('name', False), 'if-feature': ('name', False), 'import': ('module', False), 'include': ('module', False), 'input': (None, None), 'key': ('value', False), 'leaf': ('name', False), 'leaf-list': ('name', False), 'length': ('value', False), 'list': ('name', False), 'mandatory': ('value', False), 'max-elements': ('value', False), 'min-elements': ('value', False), 'modifier': ('value', False), 'module': ('name', False), 'must': ('condition', False), 'namespace': ('uri', False), 'notification': ('name', False), 'ordered-by': ('value', False), 'organization': ('text', True), 'output': (None, None), 'path': ('value', False), 'pattern': ('value', False), 'position': ('value', False), 'presence': ('value', False), 'prefix': ('value', False), 'range': ('value', False), 'reference': ('text', True), 'refine': ('target-node', False), 'require-instance': ('value', False), 'revision': ('date', False), 'revision-date': ('date', False), 'rpc': ('name', False), 'status': ('value', False), 'submodule': ('name', False), 'type': ('name', False), 'typedef': ('name', False), 'unique': ('tag', False), 'units': ('name', False), 'uses': ('name', False), 'value': ('value', False), 'when': ('condition', False), 'yang-version': ('value', False), 'yin-element': ('value', False), } """Mapping of statements to the YIN representation of their arguments. The values are pairs whose first component specifies whether the argument is stored in a subelement and the second component is the name of the attribute or subelement carrying the argument. See YANG specification. """
isc
mbj4668/pyang
pyang/translators/yin.py
1
6251
"""YIN output plugin""" from xml.sax.saxutils import quoteattr from xml.sax.saxutils import escape import optparse import re from .. import plugin from .. import util from .. import grammar from .. import syntax from .. import statements yin_namespace = "urn:ietf:params:xml:ns:yang:yin:1" def pyang_plugin_init(): plugin.register_plugin(YINPlugin()) class YINPlugin(plugin.PyangPlugin): def add_opts(self, optparser): optlist = [ optparse.make_option("--yin-canonical", dest="yin_canonical", action="store_true", help="Print in canonical order"), optparse.make_option("--yin-pretty-strings", dest="yin_pretty_strings", action="store_true", help="Pretty print strings"), ] g = optparser.add_option_group("YIN output specific options") g.add_options(optlist) def add_output_format(self, fmts): fmts['yin'] = self def emit(self, ctx, modules, fd): module = modules[0] emit_yin(ctx, module, fd) def emit_yin(ctx, module, fd): fd.write('<?xml version="1.0" encoding="UTF-8"?>\n') fd.write('<%s name="%s"\n' % (module.keyword, module.arg)) fd.write(' ' * len(module.keyword) + ' xmlns="%s"' % yin_namespace) prefix = module.search_one('prefix') if prefix is not None: namespace = module.search_one('namespace') fd.write('\n') fd.write(' ' * len(module.keyword)) fd.write(' xmlns:' + prefix.arg + '=' + quoteattr(namespace.arg)) else: belongs_to = module.search_one('belongs-to') if belongs_to is not None: prefix = belongs_to.search_one('prefix') if prefix is not None: # read the parent module in order to find the namespace uri res = ctx.read_module(belongs_to.arg, extra={'no_include':True}) if res is not None: namespace = res.search_one('namespace') if namespace is None or namespace.arg is None: pass else: # success - namespace found fd.write('\n') fd.write(' ' * len(module.keyword)) fd.write(' xmlns:' + prefix.arg + '=' + quoteattr(namespace.arg)) for imp in module.search('import'): prefix = imp.search_one('prefix') if prefix is not None: rev = None r = imp.search_one('revision-date') if r is not None: rev = r.arg mod = statements.modulename_to_module(module, imp.arg, rev) if mod is not None: ns = mod.search_one('namespace') if ns is not None: fd.write('\n') fd.write(' ' * len(module.keyword)) fd.write(' xmlns:' + prefix.arg + '=' + quoteattr(ns.arg)) fd.write('>\n') if ctx.opts.yin_canonical: substmts = grammar.sort_canonical(module.keyword, module.substmts) else: substmts = module.substmts for s in substmts: emit_stmt(ctx, module, s, fd, ' ', ' ') fd.write('</%s>\n' % module.keyword) def emit_stmt(ctx, module, stmt, fd, indent, indentstep): if util.is_prefixed(stmt.raw_keyword): # this is an extension. need to find its definition (prefix, identifier) = stmt.raw_keyword tag = prefix + ':' + identifier if stmt.i_extension is not None: ext_arg = stmt.i_extension.search_one('argument') if ext_arg is not None: yin_element = ext_arg.search_one('yin-element') if yin_element is not None and yin_element.arg == 'true': argname = prefix + ':' + ext_arg.arg argiselem = True else: # explicit false or no yin-element given argname = ext_arg.arg argiselem = False else: argiselem = False argname = None else: argiselem = False argname = None else: (argname, argiselem) = syntax.yin_map[stmt.raw_keyword] tag = stmt.raw_keyword if argiselem is False or argname is None: if argname is None: attr = '' else: attr = ' ' + argname + '=' + quoteattr(stmt.arg) if len(stmt.substmts) == 0: fd.write(indent + '<' + tag + attr + '/>\n') else: fd.write(indent + '<' + tag + attr + '>\n') for s in stmt.substmts: emit_stmt(ctx, module, s, fd, indent + indentstep, indentstep) fd.write(indent + '</' + tag + '>\n') else: fd.write(indent + '<' + tag + '>\n') if ctx.opts.yin_pretty_strings: # since whitespace is significant in XML, the current # code is strictly speaking incorrect. But w/o the whitespace, # it looks too ugly. fd.write(indent + indentstep + '<' + argname + '>\n') fd.write(fmt_text(indent + indentstep + indentstep, stmt.arg)) fd.write('\n' + indent + indentstep + '</' + argname + '>\n') else: fd.write(indent + indentstep + '<' + argname + '>' + \ escape(stmt.arg) + \ '</' + argname + '>\n') if ctx.opts.yin_canonical: substmts = grammar.sort_canonical(stmt.keyword, stmt.substmts) else: substmts = stmt.substmts for s in substmts: emit_stmt(ctx, module, s, fd, indent + indentstep, indentstep) fd.write(indent + '</' + tag + '>\n') def fmt_text(indent, data): res = [] for line in re.split("(\n)", escape(data)): if line == '': continue if line == '\n': res.extend(line) else: res.extend(indent + line) return ''.join(res)
isc
rdegges/django-twilio
test_project/test_app/client.py
1
1200
# -*- coding: utf-8 -*- from django.test import TestCase from django.contrib.auth.models import User from django.conf import settings from twilio.rest import Client from django_dynamic_fixture import G from django_twilio.client import twilio_client from django_twilio.models import Credential from django_twilio.utils import discover_twilio_credentials class TwilioClientTestCase(TestCase): def test_twilio_client_exists(self): self.assertIsInstance(twilio_client, Client) def test_twilio_client_sets_credentials(self): self.assertEqual( twilio_client.auth, (settings.TWILIO_ACCOUNT_SID, settings.TWILIO_AUTH_TOKEN) ) def test_twilio_client_with_credentials_model(self): self.user = G(User, username='test', password='pass') self.credentials = G( Credential, name='Test Credentials', account_sid='AAA', auth_token='BBB', user=self.user, ) credentials = discover_twilio_credentials(user=self.user) self.assertEqual(credentials[0], self.credentials.account_sid) self.assertEqual(credentials[1], self.credentials.auth_token)
unlicense
rdegges/django-twilio
test_project/settings.py
1
6114
# -*- coding: utf-8 -*- import sys # Django settings for test_project project. import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) import django import packaging.version BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) DEBUG = True ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'db.sqlite3', } } # Local time zone for this installation. Choices can be found here: # http://en.wikipedia.org/wiki/List_of_tz_zones_by_name # although not all choices may be available on all operating systems. # On Unix systems, a value of None will cause Django to use the same # timezone as the operating system. # If running in a Windows environment this must be set to the same as your # system time zone. TIME_ZONE = 'America/Los_Angeles' # Language code for this installation. All choices can be found here: # http://www.i18nguy.com/unicode/language-identifiers.html LANGUAGE_CODE = 'en-us' SITE_ID = 1 # If you set this to False, Django will make some optimizations so as not # to load the internationalization machinery. USE_I18N = True # If you set this to False, Django will not format dates, numbers and # calendars according to the current locale USE_L10N = True # Absolute filesystem path to the directory that will hold user-uploaded files. # Example: "/home/media/media.lawrence.com/media/" MEDIA_ROOT = '' # URL that handles the media served from MEDIA_ROOT. Make sure to use a # trailing slash. # Examples: "http://media.lawrence.com/media/", "http://example.com/media/" MEDIA_URL = '' # Absolute path to the directory static files should be collected to. # Don't put anything in this directory yourself; store your static files # in apps' "static/" subdirectories and in STATICFILES_DIRS. # Example: "/home/media/media.lawrence.com/static/" STATIC_ROOT = '' # URL prefix for static files. # Example: "http://media.lawrence.com/static/" STATIC_URL = '/static/' # URL prefix for admin static files -- CSS, JavaScript and images. # Make sure to use a trailing slash. # Examples: "http://foo.com/static/admin/", "/static/admin/". ADMIN_MEDIA_PREFIX = '/static/admin/' # Additional locations of static files STATICFILES_DIRS = ( # Put strings here, like "/home/html/static" or "C:/www/django/static". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. ) # List of finder classes that know how to find static files in # various locations. STATICFILES_FINDERS = ( 'django.contrib.staticfiles.finders.FileSystemFinder', 'django.contrib.staticfiles.finders.AppDirectoriesFinder', ) # Make this unique, and don't share it with anybody. SECRET_KEY = 'j1wd@qqodn-r9h&o@0jj!uw^#pm5wcdu2^cdsax=hm+-mk705p' # This is a temporary shim to allow the old style MIDDLEWARE_CLASSES to work # We will forge a plan to remove at least the unsupported versions soon. # Django 2.0 is the future, but 1.11 is still supported. # This test_project though is simple enough that the restrictions are small. if packaging.version.Version(django.__version__) < packaging.version.Version('2.0'): MIDDLEWARE_CLASSES = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', ) else: MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'test_project.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [os.path.join(BASE_DIR, 'templates')] , 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] INSTALLED_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.staticfiles', 'django.contrib.admin', 'django.contrib.admindocs', # django-twilio, of course! 'django_twilio', ) # A sample logging configuration. The only tangible logging # performed by this configuration is to send an email to # the site admins on every HTTP 500 error. # See http://docs.djangoproject.com/en/dev/topics/logging for # more details on how to customize your logging configuration. LOGGING = { 'version': 1, 'disable_existing_loggers': False, 'handlers': { 'mail_admins': { 'level': 'ERROR', 'class': 'django.utils.log.AdminEmailHandler' } }, 'loggers': { 'django.request': { 'handlers': ['mail_admins'], 'level': 'ERROR', 'propagate': True, }, } } # django-twilio account credentials. These fields are required to use the REST # API (initiate outbound calls and SMS messages). TWILIO_ACCOUNT_SID = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' TWILIO_AUTH_TOKEN = 'YYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYY' # The default callerid will be used for all outgoing phone calls and SMS # messages if not explicitly specified. This number must be previously # validated with twilio in order to work. See # https://www.twilio.com/user/account/phone-numbers# TWILIO_DEFAULT_CALLERID = 'NNNNNNNNNN'
unlicense
mozilla-services/buildhub
jobs/tests/test_lambda_s3_event_functional.py
1
4418
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, you can obtain one at http://mozilla.org/MPL/2.0/. import unittest import os import json import kinto_http from decouple import config from buildhub import lambda_s3_event here = os.path.dirname(__file__) server = config('SERVER_URL', default='http://localhost:8888/v1') bid = 'build-hub' cid = 'releases' class LambdaTest(unittest.TestCase): def setUp(self): filename = os.path.join(here, 'data', 's3-event-simple.json') self.event = json.load(open(filename, 'r')) def test_load_into_kinto(self): lambda_s3_event.lambda_handler(self.event, None) rid = 'firefox_54-0_win64_fr' client = kinto_http.Client(server_url=server) record = client.get_record(bucket=bid, collection=cid, id=rid)['data'] record.pop('last_modified') assert record == { 'id': 'firefox_54-0_win64_fr', 'source': { 'repository': ( 'https://hg.mozilla.org/releases/mozilla-release' ), 'revision': 'e832ed037a3c23004be73178e546d240e57b6ee1', 'product': 'firefox', 'tree': 'releases/mozilla-release' }, 'download': { 'mimetype': 'application/msdos-windows', 'url': 'https://archive.mozilla.org/pub/firefox/releases/' '54.0/win64/fr/Firefox Setup 54.0.exe', 'size': 51001024, 'date': '2017-08-08T17:06:52Z' }, 'target': { 'locale': 'fr', 'platform': 'win64', 'os': 'win', 'version': '54.0', 'channel': 'release' }, 'build': { 'as': 'ml64.exe', 'cc': ( 'c:/builds/moz2_slave/m-rel-w64-00000000000000000000/' 'build/src/vs2015u3/VC/bin/amd64/cl.exe' ), 'cxx': ( 'c:/builds/moz2_slave/m-rel-w64-00000000000000000000/' 'build/src/vs2015u3/VC/bin/amd64/cl.exe' ), 'date': '2017-06-08T10:58:25Z', 'host': 'x86_64-pc-mingw32', 'id': '20170608105825', 'number': 3, 'target': 'x86_64-pc-mingw32' } } rid = 'firefox_nightly_2017-10-29-22-01-12_58-0a1_linux-i686_en-us' record = client.get_record(bucket=bid, collection=cid, id=rid)['data'] record.pop('last_modified') assert record == { 'build': { 'as': '$(CC)', 'cc': ( '/usr/bin/ccache ' '/builds/worker/workspace/build/src/gcc/bin/gcc -m32 ' '-march=pentium-m -std=gnu99' ), 'cxx': ( '/usr/bin/ccache ' '/builds/worker/workspace/build/src/gcc/bin/g++ -m32 ' '-march=pentium-m -std=gnu++11' ), 'date': '2017-10-29T22:01:12Z', 'host': 'i686-pc-linux-gnu', 'id': '20171029220112', 'target': 'i686-pc-linux-gnu', }, 'download': { 'date': '2017-10-29T17:06:52Z', 'mimetype': 'application/x-bzip2', 'size': 51001024, 'url': ( 'https://archive.mozilla.org/pub/firefox/nightly/2017/10/' '2017-10-29-22-01-12-mozilla-central/firefox-58.0a1.' 'en-US.linux-i686.tar.bz2' ) }, 'id': ( 'firefox_nightly_2017-10-29-22-01-12_58-0a1_linux-i686_en-us' ), 'source': { 'product': 'firefox', 'repository': 'https://hg.mozilla.org/mozilla-central', 'revision': 'd3910b7628b8066d3f30d58b17b5824b05768854', 'tree': 'mozilla-central' }, 'target': { 'channel': 'nightly', 'locale': 'en-US', 'os': 'linux', 'platform': 'linux-i686', 'version': '58.0a1' } }
mpl-2.0
pikepdf/pikepdf
src/pikepdf/models/encryption.py
1
5651
# SPDX-FileCopyrightText: 2022 James R. Barlow # SPDX-License-Identifier: MPL-2.0 """For managing PDF encryption.""" from __future__ import annotations import sys from typing import TYPE_CHECKING, Any, NamedTuple, cast if sys.version_info >= (3, 8): from typing import Literal else: from typing_extensions import Literal # pragma: no cover if TYPE_CHECKING: from pikepdf._qpdf import EncryptionMethod class Permissions(NamedTuple): """ Stores the user-level permissions for an encrypted PDF. A compliant PDF reader/writer should enforce these restrictions on people who have the user password and not the owner password. In practice, either password is sufficient to decrypt all document contents. A person who has the owner password should be allowed to modify the document in any way. pikepdf does not enforce the restrictions in any way; it is up to application developers to enforce them as they see fit. Unencrypted PDFs implicitly have all permissions allowed. Permissions can only be changed when a PDF is saved. """ accessibility: bool = True """Can users use screen readers and accessibility tools to read the PDF?""" extract: bool = True """Can users extract contents?""" modify_annotation: bool = True """Can users modify annotations?""" modify_assembly: bool = False """Can users arrange document contents?""" modify_form: bool = True """Can users fill out forms?""" modify_other: bool = True """Can users modify the document?""" print_lowres: bool = True """Can users print the document at low resolution?""" print_highres: bool = True """Can users print the document at high resolution?""" DEFAULT_PERMISSIONS = Permissions() class EncryptionInfo: """ Reports encryption information for an encrypted PDF. This information may not be changed, except when a PDF is saved. This object is not used to specify the encryption settings to save a PDF, due to non-overlapping information requirements. """ def __init__(self, encdict: dict[str, Any]): """ Initialize EncryptionInfo. Generally pikepdf will initialize and return it. Args: encdict: Python dictionary containing encryption settings. """ self._encdict = encdict @property def R(self) -> int: """Revision number of the security handler.""" return int(self._encdict['R']) @property def V(self) -> int: """Version of PDF password algorithm.""" return int(self._encdict['V']) @property def P(self) -> int: """Return encoded permission bits. See :meth:`Pdf.allow` instead. """ return int(self._encdict['P']) @property def stream_method(self) -> EncryptionMethod: """Encryption method used to encode streams.""" return cast('EncryptionMethod', self._encdict['stream']) @property def string_method(self) -> EncryptionMethod: """Encryption method used to encode strings.""" return cast('EncryptionMethod', self._encdict['string']) @property def file_method(self) -> EncryptionMethod: """Encryption method used to encode the whole file.""" return cast('EncryptionMethod', self._encdict['file']) @property def user_password(self) -> bytes: """If possible, return the user password. The user password can only be retrieved when a PDF is opened with the owner password and when older versions of the encryption algorithm are used. The password is always returned as ``bytes`` even if it has a clear Unicode representation. """ return bytes(self._encdict['user_passwd']) @property def encryption_key(self) -> bytes: """Return the RC4 or AES encryption key used for this file.""" return bytes(self._encdict['encryption_key']) @property def bits(self) -> int: """Return the number of bits in the encryption algorithm. e.g. if the algorithm is AES-256, this returns 256. """ return len(self._encdict['encryption_key']) * 8 class Encryption(NamedTuple): """Specify the encryption settings to apply when a PDF is saved.""" owner: str = '' """The owner password to use. This allows full control of the file. If blank, the PDF will be encrypted and present as "(SECURED)" in PDF viewers. If the owner password is blank, the user password should be as well.""" user: str = '' """The user password to use. With this password, some restrictions will be imposed by a typical PDF reader. If blank, the PDF can be opened by anyone, but only modified as allowed by the permissions in ``allow``.""" R: Literal[2, 3, 4, 5, 6] = 6 """Select the security handler algorithm to use. Choose from: ``2``, ``3``, ``4`` or ``6``. By default, the highest version of is selected (``6``). ``5`` is a deprecated algorithm that should not be used.""" allow: Permissions = DEFAULT_PERMISSIONS """The permissions to set. If omitted, all permissions are granted to the user.""" aes: bool = True """If True, request the AES algorithm. If False, use RC4. If omitted, AES is selected whenever possible (R >= 4).""" metadata: bool = True """If True, also encrypt the PDF metadata. If False, metadata is not encrypted. Reading document metadata without decryption may be desirable in some cases. Requires ``aes=True``. If omitted, metadata is encrypted whenever possible."""
mpl-2.0
mail-in-a-box/mailinabox
management/dns_update.py
1
46993
#!/usr/local/lib/mailinabox/env/bin/python # Creates DNS zone files for all of the domains of all of the mail users # and mail aliases and restarts nsd. ######################################################################## import sys, os, os.path, urllib.parse, datetime, re, hashlib, base64 import ipaddress import rtyaml import dns.resolver from utils import shell, load_env_vars_from_file, safe_domain_name, sort_domains from ssl_certificates import get_ssl_certificates, check_certificate # From https://stackoverflow.com/questions/3026957/how-to-validate-a-domain-name-using-regex-php/16491074#16491074 # This regular expression matches domain names according to RFCs, it also accepts fqdn with an leading dot, # underscores, as well as asteriks which are allowed in domain names but not hostnames (i.e. allowed in # DNS but not in URLs), which are common in certain record types like for DKIM. DOMAIN_RE = "^(?!\-)(?:[*][.])?(?:[a-zA-Z\d\-_]{0,62}[a-zA-Z\d_]\.){1,126}(?!\d+)[a-zA-Z\d_]{1,63}(\.?)$" def get_dns_domains(env): # Add all domain names in use by email users and mail aliases, any # domains we serve web for (except www redirects because that would # lead to infinite recursion here) and ensure PRIMARY_HOSTNAME is in the list. from mailconfig import get_mail_domains from web_update import get_web_domains domains = set() domains |= set(get_mail_domains(env)) domains |= set(get_web_domains(env, include_www_redirects=False)) domains.add(env['PRIMARY_HOSTNAME']) return domains def get_dns_zones(env): # What domains should we create DNS zones for? Never create a zone for # a domain & a subdomain of that domain. domains = get_dns_domains(env) # Exclude domains that are subdomains of other domains we know. Proceed # by looking at shorter domains first. zone_domains = set() for domain in sorted(domains, key=lambda d : len(d)): for d in zone_domains: if domain.endswith("." + d): # We found a parent domain already in the list. break else: # 'break' did not occur: there is no parent domain. zone_domains.add(domain) # Make a nice and safe filename for each domain. zonefiles = [] for domain in zone_domains: zonefiles.append([domain, safe_domain_name(domain) + ".txt"]) # Sort the list so that the order is nice and so that nsd.conf has a # stable order so we don't rewrite the file & restart the service # meaninglessly. zone_order = sort_domains([ zone[0] for zone in zonefiles ], env) zonefiles.sort(key = lambda zone : zone_order.index(zone[0]) ) return zonefiles def do_dns_update(env, force=False): # Write zone files. os.makedirs('/etc/nsd/zones', exist_ok=True) zonefiles = [] updated_domains = [] for (domain, zonefile, records) in build_zones(env): # The final set of files will be signed. zonefiles.append((domain, zonefile + ".signed")) # See if the zone has changed, and if so update the serial number # and write the zone file. if not write_nsd_zone(domain, "/etc/nsd/zones/" + zonefile, records, env, force): # Zone was not updated. There were no changes. continue # Mark that we just updated this domain. updated_domains.append(domain) # Sign the zone. # # Every time we sign the zone we get a new result, which means # we can't sign a zone without bumping the zone's serial number. # Thus we only sign a zone if write_nsd_zone returned True # indicating the zone changed, and thus it got a new serial number. # write_nsd_zone is smart enough to check if a zone's signature # is nearing expiration and if so it'll bump the serial number # and return True so we get a chance to re-sign it. sign_zone(domain, zonefile, env) # Write the main nsd.conf file. if write_nsd_conf(zonefiles, list(get_custom_dns_config(env)), env): # Make sure updated_domains contains *something* if we wrote an updated # nsd.conf so that we know to restart nsd. if len(updated_domains) == 0: updated_domains.append("DNS configuration") # Tell nsd to reload changed zone files. if len(updated_domains) > 0: # 'reconfig' is needed if there are added or removed zones, but # it may not reload existing zones, so we call 'reload' too. If # nsd isn't running, nsd-control fails, so in that case revert # to restarting nsd to make sure it is running. Restarting nsd # should also refresh everything. try: shell('check_call', ["/usr/sbin/nsd-control", "reconfig"]) shell('check_call', ["/usr/sbin/nsd-control", "reload"]) except: shell('check_call', ["/usr/sbin/service", "nsd", "restart"]) # Write the OpenDKIM configuration tables for all of the mail domains. from mailconfig import get_mail_domains if write_opendkim_tables(get_mail_domains(env), env): # Settings changed. Kick opendkim. shell('check_call', ["/usr/sbin/service", "opendkim", "restart"]) if len(updated_domains) == 0: # If this is the only thing that changed? updated_domains.append("OpenDKIM configuration") # Clear bind9's DNS cache so our own DNS resolver is up to date. # (ignore errors with trap=True) shell('check_call', ["/usr/sbin/rndc", "flush"], trap=True) if len(updated_domains) == 0: # if nothing was updated (except maybe OpenDKIM's files), don't show any output return "" else: return "updated DNS: " + ",".join(updated_domains) + "\n" ######################################################################## def build_zones(env): # What domains (and their zone filenames) should we build? domains = get_dns_domains(env) zonefiles = get_dns_zones(env) # Create a dictionary of domains to a set of attributes for each # domain, such as whether there are mail users at the domain. from mailconfig import get_mail_domains from web_update import get_web_domains mail_domains = set(get_mail_domains(env)) mail_user_domains = set(get_mail_domains(env, users_only=True)) # i.e. will log in for mail, Nextcloud web_domains = set(get_web_domains(env)) auto_domains = web_domains - set(get_web_domains(env, include_auto=False)) domains |= auto_domains # www redirects not included in the initial list, see above # Add ns1/ns2+PRIMARY_HOSTNAME which must also have A/AAAA records # when the box is acting as authoritative DNS server for its domains. for ns in ("ns1", "ns2"): d = ns + "." + env["PRIMARY_HOSTNAME"] domains.add(d) auto_domains.add(d) domains = { domain: { "user": domain in mail_user_domains, "mail": domain in mail_domains, "web": domain in web_domains, "auto": domain in auto_domains, } for domain in domains } # For MTA-STS, we'll need to check if the PRIMARY_HOSTNAME certificate is # singned and valid. Check that now rather than repeatedly for each domain. domains[env["PRIMARY_HOSTNAME"]]["certificate-is-valid"] = is_domain_cert_signed_and_valid(env["PRIMARY_HOSTNAME"], env) # Load custom records to add to zones. additional_records = list(get_custom_dns_config(env)) # Build DNS records for each zone. for domain, zonefile in zonefiles: # Build the records to put in the zone. records = build_zone(domain, domains, additional_records, env) yield (domain, zonefile, records) def build_zone(domain, domain_properties, additional_records, env, is_zone=True): records = [] # For top-level zones, define the authoritative name servers. # # Normally we are our own nameservers. Some TLDs require two distinct IP addresses, # so we allow the user to override the second nameserver definition so that # secondary DNS can be set up elsewhere. # # 'False' in the tuple indicates these records would not be used if the zone # is managed outside of the box. if is_zone: # Obligatory NS record to ns1.PRIMARY_HOSTNAME. records.append((None, "NS", "ns1.%s." % env["PRIMARY_HOSTNAME"], False)) # NS record to ns2.PRIMARY_HOSTNAME or whatever the user overrides. # User may provide one or more additional nameservers secondary_ns_list = get_secondary_dns(additional_records, mode="NS") \ or ["ns2." + env["PRIMARY_HOSTNAME"]] for secondary_ns in secondary_ns_list: records.append((None, "NS", secondary_ns+'.', False)) # In PRIMARY_HOSTNAME... if domain == env["PRIMARY_HOSTNAME"]: # Set the A/AAAA records. Do this early for the PRIMARY_HOSTNAME so that the user cannot override them # and we can provide different explanatory text. records.append((None, "A", env["PUBLIC_IP"], "Required. Sets the IP address of the box.")) if env.get("PUBLIC_IPV6"): records.append((None, "AAAA", env["PUBLIC_IPV6"], "Required. Sets the IPv6 address of the box.")) # Add a DANE TLSA record for SMTP. records.append(("_25._tcp", "TLSA", build_tlsa_record(env), "Recommended when DNSSEC is enabled. Advertises to mail servers connecting to the box that mandatory encryption should be used.")) # Add a DANE TLSA record for HTTPS, which some browser extensions might make use of. records.append(("_443._tcp", "TLSA", build_tlsa_record(env), "Optional. When DNSSEC is enabled, provides out-of-band HTTPS certificate validation for a few web clients that support it.")) # Add a SSHFP records to help SSH key validation. One per available SSH key on this system. for value in build_sshfp_records(): records.append((None, "SSHFP", value, "Optional. Provides an out-of-band method for verifying an SSH key before connecting. Use 'VerifyHostKeyDNS yes' (or 'VerifyHostKeyDNS ask') when connecting with ssh.")) # Add DNS records for any subdomains of this domain. We should not have a zone for # both a domain and one of its subdomains. if is_zone: # don't recurse when we're just loading data for a subdomain subdomains = [d for d in domain_properties if d.endswith("." + domain)] for subdomain in subdomains: subdomain_qname = subdomain[0:-len("." + domain)] subzone = build_zone(subdomain, domain_properties, additional_records, env, is_zone=False) for child_qname, child_rtype, child_value, child_explanation in subzone: if child_qname == None: child_qname = subdomain_qname else: child_qname += "." + subdomain_qname records.append((child_qname, child_rtype, child_value, child_explanation)) has_rec_base = list(records) # clone current state def has_rec(qname, rtype, prefix=None): for rec in has_rec_base: if rec[0] == qname and rec[1] == rtype and (prefix is None or rec[2].startswith(prefix)): return True return False # The user may set other records that don't conflict with our settings. # Don't put any TXT records above this line, or it'll prevent any custom TXT records. for qname, rtype, value in filter_custom_records(domain, additional_records): # Don't allow custom records for record types that override anything above. # But allow multiple custom records for the same rtype --- see how has_rec_base is used. if has_rec(qname, rtype): continue # The "local" keyword on A/AAAA records are short-hand for our own IP. # This also flags for web configuration that the user wants a website here. if rtype == "A" and value == "local": value = env["PUBLIC_IP"] if rtype == "AAAA" and value == "local": if "PUBLIC_IPV6" in env: value = env["PUBLIC_IPV6"] else: continue records.append((qname, rtype, value, "(Set by user.)")) # Add A/AAAA defaults if not overridden by the user's custom settings (and not otherwise configured). # Any CNAME or A record on the qname overrides A and AAAA. But when we set the default A record, # we should not cause the default AAAA record to be skipped because it thinks a custom A record # was set. So set has_rec_base to a clone of the current set of DNS settings, and don't update # during this process. has_rec_base = list(records) a_expl = "Required. May have a different value. Sets the IP address that %s resolves to for web hosting and other services besides mail. The A record must be present but its value does not affect mail delivery." % domain if domain_properties[domain]["auto"]: if domain.startswith("ns1.") or domain.startswith("ns2."): a_expl = False # omit from 'External DNS' page since this only applies if box is its own DNS server if domain.startswith("www."): a_expl = "Optional. Sets the IP address that %s resolves to so that the box can provide a redirect to the parent domain." % domain if domain.startswith("mta-sts."): a_expl = "Optional. MTA-STS Policy Host serving /.well-known/mta-sts.txt." if domain.startswith("autoconfig."): a_expl = "Provides email configuration autodiscovery support for Thunderbird Autoconfig." if domain.startswith("autodiscover."): a_expl = "Provides email configuration autodiscovery support for Z-Push ActiveSync Autodiscover." defaults = [ (None, "A", env["PUBLIC_IP"], a_expl), (None, "AAAA", env.get('PUBLIC_IPV6'), "Optional. Sets the IPv6 address that %s resolves to, e.g. for web hosting. (It is not necessary for receiving mail on this domain.)" % domain), ] for qname, rtype, value, explanation in defaults: if value is None or value.strip() == "": continue # skip IPV6 if not set if not is_zone and qname == "www": continue # don't create any default 'www' subdomains on what are themselves subdomains # Set the default record, but not if: # (1) there is not a user-set record of the same type already # (2) there is not a CNAME record already, since you can't set both and who knows what takes precedence # (2) there is not an A record already (if this is an A record this is a dup of (1), and if this is an AAAA record then don't set a default AAAA record if the user sets a custom A record, since the default wouldn't make sense and it should not resolve if the user doesn't provide a new AAAA record) if not has_rec(qname, rtype) and not has_rec(qname, "CNAME") and not has_rec(qname, "A"): records.append((qname, rtype, value, explanation)) # Don't pin the list of records that has_rec checks against anymore. has_rec_base = records if domain_properties[domain]["mail"]: # The MX record says where email for the domain should be delivered: Here! if not has_rec(None, "MX", prefix="10 "): records.append((None, "MX", "10 %s." % env["PRIMARY_HOSTNAME"], "Required. Specifies the hostname (and priority) of the machine that handles @%s mail." % domain)) # SPF record: Permit the box ('mx', see above) to send mail on behalf of # the domain, and no one else. # Skip if the user has set a custom SPF record. if not has_rec(None, "TXT", prefix="v=spf1 "): records.append((None, "TXT", 'v=spf1 mx -all', "Recommended. Specifies that only the box is permitted to send @%s mail." % domain)) # Append the DKIM TXT record to the zone as generated by OpenDKIM. # Skip if the user has set a DKIM record already. opendkim_record_file = os.path.join(env['STORAGE_ROOT'], 'mail/dkim/mail.txt') with open(opendkim_record_file) as orf: m = re.match(r'(\S+)\s+IN\s+TXT\s+\( ((?:"[^"]+"\s+)+)\)', orf.read(), re.S) val = "".join(re.findall(r'"([^"]+)"', m.group(2))) if not has_rec(m.group(1), "TXT", prefix="v=DKIM1; "): records.append((m.group(1), "TXT", val, "Recommended. Provides a way for recipients to verify that this machine sent @%s mail." % domain)) # Append a DMARC record. # Skip if the user has set a DMARC record already. if not has_rec("_dmarc", "TXT", prefix="v=DMARC1; "): records.append(("_dmarc", "TXT", 'v=DMARC1; p=quarantine;', "Recommended. Specifies that mail that does not originate from the box but claims to be from @%s or which does not have a valid DKIM signature is suspect and should be quarantined by the recipient's mail system." % domain)) if domain_properties[domain]["user"]: # Add CardDAV/CalDAV SRV records on the non-primary hostname that points to the primary hostname # for autoconfiguration of mail clients (so only domains hosting user accounts need it). # The SRV record format is priority (0, whatever), weight (0, whatever), port, service provider hostname (w/ trailing dot). if domain != env["PRIMARY_HOSTNAME"]: for dav in ("card", "cal"): qname = "_" + dav + "davs._tcp" if not has_rec(qname, "SRV"): records.append((qname, "SRV", "0 0 443 " + env["PRIMARY_HOSTNAME"] + ".", "Recommended. Specifies the hostname of the server that handles CardDAV/CalDAV services for email addresses on this domain.")) # If this is a domain name that there are email addresses configured for, i.e. "something@" # this domain name, then the domain name is a MTA-STS (https://tools.ietf.org/html/rfc8461) # Policy Domain. # # A "_mta-sts" TXT record signals the presence of a MTA-STS policy. The id field helps clients # cache the policy. It should be stable so we don't update DNS unnecessarily but change when # the policy changes. It must be at most 32 letters and numbers, so we compute a hash of the # policy file. # # The policy itself is served at the "mta-sts" (no underscore) subdomain over HTTPS. Therefore # the TLS certificate used by Postfix for STARTTLS must be a valid certificate for the MX # domain name (PRIMARY_HOSTNAME) *and* the TLS certificate used by nginx for HTTPS on the mta-sts # subdomain must be valid certificate for that domain. Do not set an MTA-STS policy if either # certificate in use is not valid (e.g. because it is self-signed and a valid certificate has not # yet been provisioned). Since we cannot provision a certificate without A/AAAA records, we # always set them (by including them in the www domains) --- only the TXT records depend on there # being valid certificates. mta_sts_records = [ ] if domain_properties[domain]["mail"] \ and domain_properties[env["PRIMARY_HOSTNAME"]]["certificate-is-valid"] \ and is_domain_cert_signed_and_valid("mta-sts." + domain, env): # Compute an up-to-32-character hash of the policy file. We'll take a SHA-1 hash of the policy # file (20 bytes) and encode it as base-64 (28 bytes, using alphanumeric alternate characters # instead of '+' and '/' which are not allowed in an MTA-STS policy id) but then just take its # first 20 characters, which is more than sufficient to change whenever the policy file changes # (and ensures any '=' padding at the end of the base64 encoding is dropped). with open("/var/lib/mailinabox/mta-sts.txt", "rb") as f: mta_sts_policy_id = base64.b64encode(hashlib.sha1(f.read()).digest(), altchars=b"AA").decode("ascii")[0:20] mta_sts_records.extend([ ("_mta-sts", "TXT", "v=STSv1; id=" + mta_sts_policy_id, "Optional. Part of the MTA-STS policy for incoming mail. If set, a MTA-STS policy must also be published.") ]) # Enable SMTP TLS reporting (https://tools.ietf.org/html/rfc8460) if the user has set a config option. # Skip if the rules below if the user has set a custom _smtp._tls record. if env.get("MTA_STS_TLSRPT_RUA") and not has_rec("_smtp._tls", "TXT", prefix="v=TLSRPTv1;"): mta_sts_records.append(("_smtp._tls", "TXT", "v=TLSRPTv1; rua=" + env["MTA_STS_TLSRPT_RUA"], "Optional. Enables MTA-STS reporting.")) for qname, rtype, value, explanation in mta_sts_records: if not has_rec(qname, rtype): records.append((qname, rtype, value, explanation)) # Add no-mail-here records for any qname that has an A or AAAA record # but no MX record. This would include domain itself if domain is a # non-mail domain and also may include qnames from custom DNS records. # Do this once at the end of generating a zone. if is_zone: qnames_with_a = set(qname for (qname, rtype, value, explanation) in records if rtype in ("A", "AAAA")) qnames_with_mx = set(qname for (qname, rtype, value, explanation) in records if rtype == "MX") for qname in qnames_with_a - qnames_with_mx: # Mark this domain as not sending mail with hard-fail SPF and DMARC records. d = (qname+"." if qname else "") + domain if not has_rec(qname, "TXT", prefix="v=spf1 "): records.append((qname, "TXT", 'v=spf1 -all', "Recommended. Prevents use of this domain name for outbound mail by specifying that no servers are valid sources for mail from @%s. If you do send email from this domain name you should either override this record such that the SPF rule does allow the originating server, or, take the recommended approach and have the box handle mail for this domain (simply add any receiving alias at this domain name to make this machine treat the domain name as one of its mail domains)." % d)) if not has_rec("_dmarc" + ("."+qname if qname else ""), "TXT", prefix="v=DMARC1; "): records.append(("_dmarc" + ("."+qname if qname else ""), "TXT", 'v=DMARC1; p=reject;', "Recommended. Prevents use of this domain name for outbound mail by specifying that the SPF rule should be honoured for mail from @%s." % d)) # And with a null MX record (https://explained-from-first-principles.com/email/#null-mx-record) if not has_rec(qname, "MX"): records.append((qname, "MX", '0 .', "Recommended. Prevents use of this domain name for incoming mail.")) # Sort the records. The None records *must* go first in the nsd zone file. Otherwise it doesn't matter. records.sort(key = lambda rec : list(reversed(rec[0].split(".")) if rec[0] is not None else "")) return records def is_domain_cert_signed_and_valid(domain, env): cert = get_ssl_certificates(env).get(domain) if not cert: return False # no certificate provisioned cert_status = check_certificate(domain, cert['certificate'], cert['private-key']) return cert_status[0] == 'OK' ######################################################################## def build_tlsa_record(env): # A DANE TLSA record in DNS specifies that connections on a port # must use TLS and the certificate must match a particular criteria. # # Thanks to http://blog.huque.com/2012/10/dnssec-and-certificates.html # and https://community.letsencrypt.org/t/please-avoid-3-0-1-and-3-0-2-dane-tlsa-records-with-le-certificates/7022 # for explaining all of this! Also see https://tools.ietf.org/html/rfc6698#section-2.1 # and https://github.com/mail-in-a-box/mailinabox/issues/268#issuecomment-167160243. # # There are several criteria. We used to use "3 0 1" criteria, which # meant to pin a leaf (3) certificate (0) with SHA256 hash (1). But # certificates change, and especially as we move to short-lived certs # they change often. The TLSA record handily supports the criteria of # a leaf certificate (3)'s subject public key (1) with SHA256 hash (1). # The subject public key is the public key portion of the private key # that generated the CSR that generated the certificate. Since we # generate a private key once the first time Mail-in-a-Box is set up # and reuse it for all subsequent certificates, the TLSA record will # remain valid indefinitely. from ssl_certificates import load_cert_chain, load_pem from cryptography.hazmat.primitives.serialization import Encoding, PublicFormat fn = os.path.join(env["STORAGE_ROOT"], "ssl", "ssl_certificate.pem") cert = load_pem(load_cert_chain(fn)[0]) subject_public_key = cert.public_key().public_bytes(Encoding.DER, PublicFormat.SubjectPublicKeyInfo) # We could have also loaded ssl_private_key.pem and called priv_key.public_key().public_bytes(...) pk_hash = hashlib.sha256(subject_public_key).hexdigest() # Specify the TLSA parameters: # 3: Match the (leaf) certificate. (No CA, no trust path needed.) # 1: Match its subject public key. # 1: Use SHA256. return "3 1 1 " + pk_hash def build_sshfp_records(): # The SSHFP record is a way for us to embed this server's SSH public # key fingerprint into the DNS so that remote hosts have an out-of-band # method to confirm the fingerprint. See RFC 4255 and RFC 6594. This # depends on DNSSEC. # # On the client side, set SSH's VerifyHostKeyDNS option to 'ask' to # include this info in the key verification prompt or 'yes' to trust # the SSHFP record. # # See https://github.com/xelerance/sshfp for inspiriation. algorithm_number = { "ssh-rsa": 1, "ssh-dss": 2, "ecdsa-sha2-nistp256": 3, "ssh-ed25519": 4, } # Get our local fingerprints by running ssh-keyscan. The output looks # like the known_hosts file: hostname, keytype, fingerprint. The order # of the output is arbitrary, so sort it to prevent spurrious updates # to the zone file (that trigger bumping the serial number). However, # if SSH has been configured to listen on a nonstandard port, we must # specify that port to sshkeyscan. port = 22 with open('/etc/ssh/sshd_config', 'r') as f: for line in f: s = line.rstrip().split() if len(s) == 2 and s[0] == 'Port': try: port = int(s[1]) except ValueError: pass break keys = shell("check_output", ["ssh-keyscan", "-t", "rsa,dsa,ecdsa,ed25519", "-p", str(port), "localhost"]) keys = sorted(keys.split("\n")) for key in keys: if key.strip() == "" or key[0] == "#": continue try: host, keytype, pubkey = key.split(" ") yield "%d %d ( %s )" % ( algorithm_number[keytype], 2, # specifies we are using SHA-256 on next line hashlib.sha256(base64.b64decode(pubkey)).hexdigest().upper(), ) except: # Lots of things can go wrong. Don't let it disturb the DNS # zone. pass ######################################################################## def write_nsd_zone(domain, zonefile, records, env, force): # On the $ORIGIN line, there's typically a ';' comment at the end explaining # what the $ORIGIN line does. Any further data after the domain confuses # ldns-signzone, however. It used to say '; default zone domain'. # # The SOA contact address for all of the domains on this system is hostmaster # @ the PRIMARY_HOSTNAME. Hopefully that's legit. # # For the refresh through TTL fields, a good reference is: # https://www.ripe.net/publications/docs/ripe-203 # # A hash of the available DNSSEC keys are added in a comment so that when # the keys change we force a re-generation of the zone which triggers # re-signing it. zone = """ $ORIGIN {domain}. $TTL 86400 ; default time to live @ IN SOA ns1.{primary_domain}. hostmaster.{primary_domain}. ( __SERIAL__ ; serial number 7200 ; Refresh (secondary nameserver update interval) 3600 ; Retry (when refresh fails, how often to try again, should be lower than the refresh) 1209600 ; Expire (when refresh fails, how long secondary nameserver will keep records around anyway) 86400 ; Negative TTL (how long negative responses are cached) ) """ # Replace replacement strings. zone = zone.format(domain=domain, primary_domain=env["PRIMARY_HOSTNAME"]) # Add records. for subdomain, querytype, value, explanation in records: if subdomain: zone += subdomain zone += "\tIN\t" + querytype + "\t" if querytype == "TXT": # Divide into 255-byte max substrings. v2 = "" while len(value) > 0: s = value[0:255] value = value[255:] s = s.replace('\\', '\\\\') # escape backslashes s = s.replace('"', '\\"') # escape quotes s = '"' + s + '"' # wrap in quotes v2 += s + " " value = v2 zone += value + "\n" # Append a stable hash of DNSSEC signing keys in a comment. zone += "\n; DNSSEC signing keys hash: {}\n".format(hash_dnssec_keys(domain, env)) # DNSSEC requires re-signing a zone periodically. That requires # bumping the serial number even if no other records have changed. # We don't see the DNSSEC records yet, so we have to figure out # if a re-signing is necessary so we can prematurely bump the # serial number. force_bump = False if not os.path.exists(zonefile + ".signed"): # No signed file yet. Shouldn't normally happen unless a box # is going from not using DNSSEC to using DNSSEC. force_bump = True else: # We've signed the domain. Check if we are close to the expiration # time of the signature. If so, we'll force a bump of the serial # number so we can re-sign it. with open(zonefile + ".signed") as f: signed_zone = f.read() expiration_times = re.findall(r"\sRRSIG\s+SOA\s+\d+\s+\d+\s\d+\s+(\d{14})", signed_zone) if len(expiration_times) == 0: # weird force_bump = True else: # All of the times should be the same, but if not choose the soonest. expiration_time = min(expiration_times) expiration_time = datetime.datetime.strptime(expiration_time, "%Y%m%d%H%M%S") if expiration_time - datetime.datetime.now() < datetime.timedelta(days=3): # We're within three days of the expiration, so bump serial & resign. force_bump = True # Set the serial number. serial = datetime.datetime.now().strftime("%Y%m%d00") if os.path.exists(zonefile): # If the zone already exists, is different, and has a later serial number, # increment the number. with open(zonefile) as f: existing_zone = f.read() m = re.search(r"(\d+)\s*;\s*serial number", existing_zone) if m: # Clear out the serial number in the existing zone file for the # purposes of seeing if anything *else* in the zone has changed. existing_serial = m.group(1) existing_zone = existing_zone.replace(m.group(0), "__SERIAL__ ; serial number") # If the existing zone is the same as the new zone (modulo the serial number), # there is no need to update the file. Unless we're forcing a bump. if zone == existing_zone and not force_bump and not force: return False # If the existing serial is not less than a serial number # based on the current date plus 00, increment it. Otherwise, # the serial number is less than our desired new serial number # so we'll use the desired new number. if existing_serial >= serial: serial = str(int(existing_serial) + 1) zone = zone.replace("__SERIAL__", serial) # Write the zone file. with open(zonefile, "w") as f: f.write(zone) return True # file is updated def get_dns_zonefile(zone, env): for domain, fn in get_dns_zones(env): if zone == domain: break else: raise ValueError("%s is not a domain name that corresponds to a zone." % zone) nsd_zonefile = "/etc/nsd/zones/" + fn with open(nsd_zonefile, "r") as f: return f.read() ######################################################################## def write_nsd_conf(zonefiles, additional_records, env): # Write the list of zones to a configuration file. nsd_conf_file = "/etc/nsd/nsd.conf.d/zones.conf" nsdconf = "" # Append the zones. for domain, zonefile in zonefiles: nsdconf += """ zone: name: %s zonefile: %s """ % (domain, zonefile) # If custom secondary nameservers have been set, allow zone transfers # and, if not a subnet, notifies to them. for ipaddr in get_secondary_dns(additional_records, mode="xfr"): if "/" not in ipaddr: nsdconf += "\n\tnotify: %s NOKEY" % (ipaddr) nsdconf += "\n\tprovide-xfr: %s NOKEY\n" % (ipaddr) # Check if the file is changing. If it isn't changing, # return False to flag that no change was made. if os.path.exists(nsd_conf_file): with open(nsd_conf_file) as f: if f.read() == nsdconf: return False # Write out new contents and return True to signal that # configuration changed. with open(nsd_conf_file, "w") as f: f.write(nsdconf) return True ######################################################################## def find_dnssec_signing_keys(domain, env): # For key that we generated (one per algorithm)... d = os.path.join(env['STORAGE_ROOT'], 'dns/dnssec') keyconfs = [f for f in os.listdir(d) if f.endswith(".conf")] for keyconf in keyconfs: # Load the file holding the KSK and ZSK key filenames. keyconf_fn = os.path.join(d, keyconf) keyinfo = load_env_vars_from_file(keyconf_fn) # Skip this key if the conf file has a setting named DOMAINS, # holding a comma-separated list of domain names, and if this # domain is not in the list. This allows easily disabling a # key by setting "DOMAINS=" or "DOMAINS=none", other than # deleting the key's .conf file, which might result in the key # being regenerated next upgrade. Keys should be disabled if # they are not needed to reduce the DNSSEC query response size. if "DOMAINS" in keyinfo and domain not in [dd.strip() for dd in keyinfo["DOMAINS"].split(",")]: continue for keytype in ("KSK", "ZSK"): yield keytype, keyinfo[keytype] def hash_dnssec_keys(domain, env): # Create a stable (by sorting the items) hash of all of the private keys # that will be used to sign this domain. keydata = [] for keytype, keyfn in sorted(find_dnssec_signing_keys(domain, env)): oldkeyfn = os.path.join(env['STORAGE_ROOT'], 'dns/dnssec', keyfn + ".private") keydata.append(keytype) keydata.append(keyfn) with open(oldkeyfn, "r") as fr: keydata.append( fr.read() ) keydata = "".join(keydata).encode("utf8") return hashlib.sha1(keydata).hexdigest() def sign_zone(domain, zonefile, env): # Sign the zone with all of the keys that were generated during # setup so that the user can choose which to use in their DS record at # their registrar, and also to support migration to newer algorithms. # In order to use the key files generated at setup which are for # the domain _domain_, we have to re-write the files and place # the actual domain name in it, so that ldns-signzone works. # # Patch each key, storing the patched version in /tmp for now. # Each key has a .key and .private file. Collect a list of filenames # for all of the keys (and separately just the key-signing keys). all_keys = [] ksk_keys = [] for keytype, keyfn in find_dnssec_signing_keys(domain, env): newkeyfn = '/tmp/' + keyfn.replace("_domain_", domain) for ext in (".private", ".key"): # Copy the .key and .private files to /tmp to patch them up. # # Use os.umask and open().write() to securely create a copy that only # we (root) can read. oldkeyfn = os.path.join(env['STORAGE_ROOT'], 'dns/dnssec', keyfn + ext) with open(oldkeyfn, "r") as fr: keydata = fr.read() keydata = keydata.replace("_domain_", domain) prev_umask = os.umask(0o77) # ensure written file is not world-readable try: with open(newkeyfn + ext, "w") as fw: fw.write(keydata) finally: os.umask(prev_umask) # other files we write should be world-readable # Put the patched key filename base (without extension) into the list of keys we'll sign with. all_keys.append(newkeyfn) if keytype == "KSK": ksk_keys.append(newkeyfn) # Do the signing. expiry_date = (datetime.datetime.now() + datetime.timedelta(days=30)).strftime("%Y%m%d") shell('check_call', ["/usr/bin/ldns-signzone", # expire the zone after 30 days "-e", expiry_date, # use NSEC3 "-n", # zonefile to sign "/etc/nsd/zones/" + zonefile, ] # keys to sign with (order doesn't matter -- it'll figure it out) + all_keys ) # Create a DS record based on the patched-up key files. The DS record is specific to the # zone being signed, so we can't use the .ds files generated when we created the keys. # The DS record points to the KSK only. Write this next to the zone file so we can # get it later to give to the user with instructions on what to do with it. # # Generate a DS record for each key. There are also several possible hash algorithms that may # be used, so we'll pre-generate all for each key. One DS record per line. Only one # needs to actually be deployed at the registrar. We'll select the preferred one # in the status checks. with open("/etc/nsd/zones/" + zonefile + ".ds", "w") as f: for key in ksk_keys: for digest_type in ('1', '2', '4'): rr_ds = shell('check_output', ["/usr/bin/ldns-key2ds", "-n", # output to stdout "-" + digest_type, # 1=SHA1, 2=SHA256, 4=SHA384 key + ".key" ]) f.write(rr_ds) # Remove the temporary patched key files. for fn in all_keys: os.unlink(fn + ".private") os.unlink(fn + ".key") ######################################################################## def write_opendkim_tables(domains, env): # Append a record to OpenDKIM's KeyTable and SigningTable for each domain # that we send mail from (zones and all subdomains). opendkim_key_file = os.path.join(env['STORAGE_ROOT'], 'mail/dkim/mail.private') if not os.path.exists(opendkim_key_file): # Looks like OpenDKIM is not installed. return False config = { # The SigningTable maps email addresses to a key in the KeyTable that # specifies signing information for matching email addresses. Here we # map each domain to a same-named key. # # Elsewhere we set the DMARC policy for each domain such that mail claiming # to be From: the domain must be signed with a DKIM key on the same domain. # So we must have a separate KeyTable entry for each domain. "SigningTable": "".join( "*@{domain} {domain}\n".format(domain=domain) for domain in domains ), # The KeyTable specifies the signing domain, the DKIM selector, and the # path to the private key to use for signing some mail. Per DMARC, the # signing domain must match the sender's From: domain. "KeyTable": "".join( "{domain} {domain}:mail:{key_file}\n".format(domain=domain, key_file=opendkim_key_file) for domain in domains ), } did_update = False for filename, content in config.items(): # Don't write the file if it doesn't need an update. if os.path.exists("/etc/opendkim/" + filename): with open("/etc/opendkim/" + filename) as f: if f.read() == content: continue # The contents needs to change. with open("/etc/opendkim/" + filename, "w") as f: f.write(content) did_update = True # Return whether the files changed. If they didn't change, there's # no need to kick the opendkim process. return did_update ######################################################################## def get_custom_dns_config(env, only_real_records=False): try: custom_dns = rtyaml.load(open(os.path.join(env['STORAGE_ROOT'], 'dns/custom.yaml'))) if not isinstance(custom_dns, dict): raise ValueError() # caught below except: return [ ] for qname, value in custom_dns.items(): if qname == "_secondary_nameserver" and only_real_records: continue # skip fake record # Short form. Mapping a domain name to a string is short-hand # for creating A records. if isinstance(value, str): values = [("A", value)] # A mapping creates multiple records. elif isinstance(value, dict): values = value.items() # No other type of data is allowed. else: raise ValueError() for rtype, value2 in values: if isinstance(value2, str): yield (qname, rtype, value2) elif isinstance(value2, list): for value3 in value2: yield (qname, rtype, value3) # No other type of data is allowed. else: raise ValueError() def filter_custom_records(domain, custom_dns_iter): for qname, rtype, value in custom_dns_iter: # We don't count the secondary nameserver config (if present) as a record - that would just be # confusing to users. Instead it is accessed/manipulated directly via (get/set)_custom_dns_config. if qname == "_secondary_nameserver": continue # Is this record for the domain or one of its subdomains? # If `domain` is None, return records for all domains. if domain is not None and qname != domain and not qname.endswith("." + domain): continue # Turn the fully qualified domain name in the YAML file into # our short form (None => domain, or a relative QNAME) if # domain is not None. if domain is not None: if qname == domain: qname = None else: qname = qname[0:len(qname)-len("." + domain)] yield (qname, rtype, value) def write_custom_dns_config(config, env): # We get a list of (qname, rtype, value) triples. Convert this into a # nice dictionary format for storage on disk. from collections import OrderedDict config = list(config) dns = OrderedDict() seen_qnames = set() # Process the qnames in the order we see them. for qname in [rec[0] for rec in config]: if qname in seen_qnames: continue seen_qnames.add(qname) records = [(rec[1], rec[2]) for rec in config if rec[0] == qname] if len(records) == 1 and records[0][0] == "A": dns[qname] = records[0][1] else: dns[qname] = OrderedDict() seen_rtypes = set() # Process the rtypes in the order we see them. for rtype in [rec[0] for rec in records]: if rtype in seen_rtypes: continue seen_rtypes.add(rtype) values = [rec[1] for rec in records if rec[0] == rtype] if len(values) == 1: values = values[0] dns[qname][rtype] = values # Write. config_yaml = rtyaml.dump(dns) with open(os.path.join(env['STORAGE_ROOT'], 'dns/custom.yaml'), "w") as f: f.write(config_yaml) def set_custom_dns_record(qname, rtype, value, action, env): # validate qname for zone, fn in get_dns_zones(env): # It must match a zone apex or be a subdomain of a zone # that we are otherwise hosting. if qname == zone or qname.endswith("."+zone): break else: # No match. if qname != "_secondary_nameserver": raise ValueError("%s is not a domain name or a subdomain of a domain name managed by this box." % qname) # validate rtype rtype = rtype.upper() if value is not None and qname != "_secondary_nameserver": if not re.search(DOMAIN_RE, qname): raise ValueError("Invalid name.") if rtype in ("A", "AAAA"): if value != "local": # "local" is a special flag for us v = ipaddress.ip_address(value) # raises a ValueError if there's a problem if rtype == "A" and not isinstance(v, ipaddress.IPv4Address): raise ValueError("That's an IPv6 address.") if rtype == "AAAA" and not isinstance(v, ipaddress.IPv6Address): raise ValueError("That's an IPv4 address.") elif rtype in ("CNAME", "NS"): if rtype == "NS" and qname == zone: raise ValueError("NS records can only be set for subdomains.") # ensure value has a trailing dot if not value.endswith("."): value = value + "." if not re.search(DOMAIN_RE, value): raise ValueError("Invalid value.") elif rtype in ("CNAME", "TXT", "SRV", "MX", "SSHFP", "CAA"): # anything goes pass else: raise ValueError("Unknown record type '%s'." % rtype) # load existing config config = list(get_custom_dns_config(env)) # update newconfig = [] made_change = False needs_add = True for _qname, _rtype, _value in config: if action == "add": if (_qname, _rtype, _value) == (qname, rtype, value): # Record already exists. Bail. return False elif action == "set": if (_qname, _rtype) == (qname, rtype): if _value == value: # Flag that the record already exists, don't # need to add it. needs_add = False else: # Drop any other values for this (qname, rtype). made_change = True continue elif action == "remove": if (_qname, _rtype, _value) == (qname, rtype, value): # Drop this record. made_change = True continue if value == None and (_qname, _rtype) == (qname, rtype): # Drop all qname-rtype records. made_change = True continue else: raise ValueError("Invalid action: " + action) # Preserve this record. newconfig.append((_qname, _rtype, _value)) if action in ("add", "set") and needs_add and value is not None: newconfig.append((qname, rtype, value)) made_change = True if made_change: # serialize & save write_custom_dns_config(newconfig, env) return made_change ######################################################################## def get_secondary_dns(custom_dns, mode=None): resolver = dns.resolver.get_default_resolver() resolver.timeout = 10 values = [] for qname, rtype, value in custom_dns: if qname != '_secondary_nameserver': continue for hostname in value.split(" "): hostname = hostname.strip() if mode == None: # Just return the setting. values.append(hostname) continue # This is a hostname. Before including in zone xfr lines, # resolve to an IP address. Otherwise just return the hostname. # It may not resolve to IPv6, so don't throw an exception if it # doesn't. if not hostname.startswith("xfr:"): if mode == "xfr": response = dns.resolver.resolve(hostname+'.', "A", raise_on_no_answer=False) values.extend(map(str, response)) response = dns.resolver.resolve(hostname+'.', "AAAA", raise_on_no_answer=False) values.extend(map(str, response)) continue values.append(hostname) # This is a zone-xfer-only IP address. Do not return if # we're querying for NS record hostnames. Only return if # we're querying for zone xfer IP addresses - return the # IP address. elif mode == "xfr": values.append(hostname[4:]) return values def set_secondary_dns(hostnames, env): if len(hostnames) > 0: # Validate that all hostnames are valid and that all zone-xfer IP addresses are valid. resolver = dns.resolver.get_default_resolver() resolver.timeout = 5 for item in hostnames: if not item.startswith("xfr:"): # Resolve hostname. try: response = resolver.resolve(item, "A") except (dns.resolver.NoNameservers, dns.resolver.NXDOMAIN, dns.resolver.NoAnswer): try: response = resolver.resolve(item, "AAAA") except (dns.resolver.NoNameservers, dns.resolver.NXDOMAIN, dns.resolver.NoAnswer): raise ValueError("Could not resolve the IP address of %s." % item) else: # Validate IP address. try: if "/" in item[4:]: v = ipaddress.ip_network(item[4:]) # raises a ValueError if there's a problem else: v = ipaddress.ip_address(item[4:]) # raises a ValueError if there's a problem except ValueError: raise ValueError("'%s' is not an IPv4 or IPv6 address or subnet." % item[4:]) # Set. set_custom_dns_record("_secondary_nameserver", "A", " ".join(hostnames), "set", env) else: # Clear. set_custom_dns_record("_secondary_nameserver", "A", None, "set", env) # Apply. return do_dns_update(env) def get_custom_dns_records(custom_dns, qname, rtype): for qname1, rtype1, value in custom_dns: if qname1 == qname and rtype1 == rtype: yield value return None ######################################################################## def build_recommended_dns(env): ret = [] for (domain, zonefile, records) in build_zones(env): # remove records that we don't dislay records = [r for r in records if r[3] is not False] # put Required at the top, then Recommended, then everythiing else records.sort(key = lambda r : 0 if r[3].startswith("Required.") else (1 if r[3].startswith("Recommended.") else 2)) # expand qnames for i in range(len(records)): if records[i][0] == None: qname = domain else: qname = records[i][0] + "." + domain records[i] = { "qname": qname, "rtype": records[i][1], "value": records[i][2], "explanation": records[i][3], } # return ret.append((domain, records)) return ret if __name__ == "__main__": from utils import load_environment env = load_environment() if sys.argv[-1] == "--lint": write_custom_dns_config(get_custom_dns_config(env), env) else: for zone, records in build_recommended_dns(env): for record in records: print("; " + record['explanation']) print(record['qname'], record['rtype'], record['value'], sep="\t") print()
cc0-1.0
mcedit/pymclevel
mce.py
2
47597
#!/usr/bin/env python import mclevelbase import mclevel import materials import infiniteworld import sys import os from box import BoundingBox, Vector import numpy from numpy import zeros, bincount import logging import itertools import traceback import shlex import operator import codecs from math import floor try: import readline # if available, used by raw_input() except: pass class UsageError(RuntimeError): pass class BlockMatchError(RuntimeError): pass class PlayerNotFound(RuntimeError): pass class mce(object): """ Block commands: {commandPrefix}clone <sourceBox> <destPoint> [noair] [nowater] {commandPrefix}fill <blockType> [ <box> ] {commandPrefix}replace <blockType> [with] <newBlockType> [ <box> ] {commandPrefix}export <filename> <sourceBox> {commandPrefix}import <filename> <destPoint> [noair] [nowater] {commandPrefix}createChest <point> <item> [ <count> ] {commandPrefix}analyze Player commands: {commandPrefix}player [ <player> [ <point> ] ] {commandPrefix}spawn [ <point> ] Entity commands: {commandPrefix}removeEntities [ <EntityID> ] {commandPrefix}dumpSigns [ <filename> ] {commandPrefix}dumpChests [ <filename> ] Chunk commands: {commandPrefix}createChunks <box> {commandPrefix}deleteChunks <box> {commandPrefix}prune <box> {commandPrefix}relight [ <box> ] World commands: {commandPrefix}create <filename> {commandPrefix}dimension [ <dim> ] {commandPrefix}degrief {commandPrefix}time [ <time> ] {commandPrefix}worldsize {commandPrefix}heightmap <filename> {commandPrefix}randomseed [ <seed> ] {commandPrefix}gametype [ <player> [ <gametype> ] ] Editor commands: {commandPrefix}save {commandPrefix}reload {commandPrefix}load <filename> | <world number> {commandPrefix}execute <filename> {commandPrefix}quit Informational: {commandPrefix}blocks [ <block name> | <block ID> ] {commandPrefix}help [ <command> ] **IMPORTANT** {commandPrefix}box Type 'box' to learn how to specify points and areas. """ random_seed = os.getenv('MCE_RANDOM_SEED', None) last_played = os.getenv("MCE_LAST_PLAYED", None) def commandUsage(self, command): " returns usage info for the named command - just give the docstring for the handler func " func = getattr(self, "_" + command) return func.__doc__ commands = [ "clone", "fill", "replace", "export", "execute", "import", "createchest", "player", "spawn", "removeentities", "dumpsigns", "dumpchests", "createchunks", "deletechunks", "prune", "relight", "create", "degrief", "time", "worldsize", "heightmap", "randomseed", "gametype", "save", "load", "reload", "dimension", "repair", "quit", "exit", "help", "blocks", "analyze", "region", "debug", "log", "box", ] debug = False needsSave = False def readInt(self, command): try: val = int(command.pop(0)) except ValueError: raise UsageError("Cannot understand numeric input") return val def prettySplit(self, command): cmdstring = " ".join(command) lex = shlex.shlex(cmdstring) lex.whitespace_split = True lex.whitespace += "()," command[:] = list(lex) def readBox(self, command): self.prettySplit(command) sourcePoint = self.readIntPoint(command) if command[0].lower() == "to": command.pop(0) sourcePoint2 = self.readIntPoint(command) sourceSize = sourcePoint2 - sourcePoint else: sourceSize = self.readIntPoint(command, isPoint=False) if len([p for p in sourceSize if p <= 0]): raise UsageError("Box size cannot be zero or negative") box = BoundingBox(sourcePoint, sourceSize) return box def readIntPoint(self, command, isPoint=True): point = self.readPoint(command, isPoint) point = map(int, map(floor, point)) return Vector(*point) def readPoint(self, command, isPoint=True): self.prettySplit(command) try: word = command.pop(0) if isPoint and (word in self.level.players): x, y, z = self.level.getPlayerPosition(word) if len(command) and command[0].lower() == "delta": command.pop(0) try: x += int(command.pop(0)) y += int(command.pop(0)) z += int(command.pop(0)) except ValueError: raise UsageError("Error decoding point input (expected a number).") return x, y, z except IndexError: raise UsageError("Error decoding point input (expected more values).") try: try: x = float(word) except ValueError: if isPoint: raise PlayerNotFound(word) raise y = float(command.pop(0)) z = float(command.pop(0)) except ValueError: raise UsageError("Error decoding point input (expected a number).") except IndexError: raise UsageError("Error decoding point input (expected more values).") return x, y, z def readBlockInfo(self, command): keyword = command.pop(0) matches = self.level.materials.blocksMatching(keyword) blockInfo = None if len(matches): if len(matches) == 1: blockInfo = matches[0] # eat up more words that possibly specify a block. stop eating when 0 matching blocks. while len(command): newMatches = self.level.materials.blocksMatching(keyword + " " + command[0]) if len(newMatches) == 1: blockInfo = newMatches[0] if len(newMatches) > 0: matches = newMatches keyword = keyword + " " + command.pop(0) else: break else: try: data = 0 if ":" in keyword: blockID, data = map(int, keyword.split(":")) else: blockID = int(keyword) blockInfo = self.level.materials.blockWithID(blockID, data) except ValueError: blockInfo = None if blockInfo is None: print "Ambiguous block specifier: ", keyword if len(matches): print "Matches: " for m in matches: if m == self.level.materials.defaultName: continue print "{0:3}:{1:<2} : {2}".format(m.ID, m.blockData, m.name) else: print "No blocks matched." raise BlockMatchError return blockInfo def readBlocksToCopy(self, command): blocksToCopy = range(materials.id_limit) while len(command): word = command.pop() if word == "noair": blocksToCopy.remove(0) if word == "nowater": blocksToCopy.remove(8) blocksToCopy.remove(9) return blocksToCopy def _box(self, command): """ Boxes: Many commands require a <box> as arguments. A box can be specified with a point and a size: (12, 5, 15), (5, 5, 5) or with two points, making sure to put the keyword "to" between them: (12, 5, 15) to (17, 10, 20) The commas and parentheses are not important. You may add them for improved readability. Points: Points and sizes are triplets of numbers ordered X Y Z. X is position north-south, increasing southward. Y is position up-down, increasing upward. Z is position east-west, increasing westward. Players: A player's name can be used as a point - it will use the position of the player's head. Use the keyword 'delta' after the name to specify a point near the player. Example: codewarrior delta 0 5 0 This refers to a point 5 blocks above codewarrior's head. """ raise UsageError def _debug(self, command): self.debug = not self.debug print "Debug", ("disabled", "enabled")[self.debug] def _log(self, command): """ log [ <number> ] Get or set the log threshold. 0 logs everything; 50 only logs major errors. """ if len(command): try: logging.getLogger().level = int(command[0]) except ValueError: raise UsageError("Cannot understand numeric input.") else: print "Log level: {0}".format(logging.getLogger().level) def _clone(self, command): """ clone <sourceBox> <destPoint> [noair] [nowater] Clone blocks in a cuboid starting at sourcePoint and extending for sourceSize blocks in each direction. Blocks and entities in the area are cloned at destPoint. """ if len(command) == 0: self.printUsage("clone") return box = self.readBox(command) destPoint = self.readPoint(command) destPoint = map(int, map(floor, destPoint)) blocksToCopy = self.readBlocksToCopy(command) tempSchematic = self.level.extractSchematic(box) self.level.copyBlocksFrom(tempSchematic, BoundingBox((0, 0, 0), box.origin), destPoint, blocksToCopy) self.needsSave = True print "Cloned 0 blocks." def _fill(self, command): """ fill <blockType> [ <box> ] Fill blocks with blockType in a cuboid starting at point and extending for size blocks in each direction. Without a destination, fills the whole world. blockType and may be a number from 0-255 or a name listed by the 'blocks' command. """ if len(command) == 0: self.printUsage("fill") return blockInfo = self.readBlockInfo(command) if len(command): box = self.readBox(command) else: box = None print "Filling with {0}".format(blockInfo.name) self.level.fillBlocks(box, blockInfo) self.needsSave = True print "Filled {0} blocks.".format("all" if box is None else box.volume) def _replace(self, command): """ replace <blockType> [with] <newBlockType> [ <box> ] Replace all blockType blocks with newBlockType in a cuboid starting at point and extending for size blocks in each direction. Without a destination, replaces blocks over the whole world. blockType and newBlockType may be numbers from 0-255 or names listed by the 'blocks' command. """ if len(command) == 0: self.printUsage("replace") return blockInfo = self.readBlockInfo(command) if command[0].lower() == "with": command.pop(0) newBlockInfo = self.readBlockInfo(command) if len(command): box = self.readBox(command) else: box = None print "Replacing {0} with {1}".format(blockInfo.name, newBlockInfo.name) self.level.fillBlocks(box, newBlockInfo, blocksToReplace=[blockInfo]) self.needsSave = True print "Done." def _createchest(self, command): """ createChest <point> <item> [ <count> ] Create a chest filled with the specified item. Stacks are 64 if count is not given. """ point = map(lambda x: int(floor(float(x))), self.readPoint(command)) itemID = self.readInt(command) count = 64 if len(command): count = self.readInt(command) chest = mclevel.MCSchematic.chestWithItemID(itemID, count) self.level.copyBlocksFrom(chest, chest.bounds, point) self.needsSave = True def _analyze(self, command): """ analyze Counts all of the block types in every chunk of the world. """ blockCounts = zeros((65536,), 'uint64') sizeOnDisk = 0 print "Analyzing {0} chunks...".format(self.level.chunkCount) # for input to bincount, create an array of uint16s by # shifting the data left and adding the blocks for i, cPos in enumerate(self.level.allChunks, 1): ch = self.level.getChunk(*cPos) btypes = numpy.array(ch.Data.ravel(), dtype='uint16') btypes <<= 12 btypes += ch.Blocks.ravel() counts = bincount(btypes) blockCounts[:counts.shape[0]] += counts if i % 100 == 0: logging.info("Chunk {0}...".format(i)) for blockID in range(materials.id_limit): block = self.level.materials.blockWithID(blockID, 0) if block.hasVariants: for data in range(16): i = (data << 12) + blockID if blockCounts[i]: idstring = "({id}:{data})".format(id=blockID, data=data) print "{idstring:9} {name:30}: {count:<10}".format( idstring=idstring, name=self.level.materials.blockWithID(blockID, data).name, count=blockCounts[i]) else: count = int(sum(blockCounts[(d << 12) + blockID] for d in range(16))) if count: idstring = "({id})".format(id=blockID) print "{idstring:9} {name:30}: {count:<10}".format( idstring=idstring, name=self.level.materials.blockWithID(blockID, 0).name, count=count) self.needsSave = True def _export(self, command): """ export <filename> <sourceBox> Exports blocks in the specified region to a file in schematic format. This file can be imported with mce or MCEdit. """ if len(command) == 0: self.printUsage("export") return filename = command.pop(0) box = self.readBox(command) tempSchematic = self.level.extractSchematic(box) tempSchematic.saveToFile(filename) print "Exported {0} blocks.".format(tempSchematic.bounds.volume) def _import(self, command): """ import <filename> <destPoint> [noair] [nowater] Imports a level or schematic into this world, beginning at destPoint. Supported formats include - Alpha single or multiplayer world folder containing level.dat, - Zipfile containing Alpha world folder, - Classic single-player .mine, - Classic multiplayer server_level.dat, - Indev .mclevel - Schematic from RedstoneSim, MCEdit, mce - .inv from INVEdit (appears as a chest) """ if len(command) == 0: self.printUsage("import") return filename = command.pop(0) destPoint = self.readPoint(command) blocksToCopy = self.readBlocksToCopy(command) importLevel = mclevel.fromFile(filename) self.level.copyBlocksFrom(importLevel, importLevel.bounds, destPoint, blocksToCopy, create=True) self.needsSave = True print "Imported {0} blocks.".format(importLevel.bounds.volume) def _player(self, command): """ player [ <player> [ <point> ] ] Move the named player to the specified point. Without a point, prints the named player's position. Without a player, prints all players and positions. In a single-player world, the player is named Player. """ if len(command) == 0: print "Players: " for player in self.level.players: print " {0}: {1}".format(player, self.level.getPlayerPosition(player)) return player = command.pop(0) if len(command) == 0: print "Player {0}: {1}".format(player, self.level.getPlayerPosition(player)) return point = self.readPoint(command) self.level.setPlayerPosition(point, player) self.needsSave = True print "Moved player {0} to {1}".format(player, point) def _spawn(self, command): """ spawn [ <point> ] Move the world's spawn point. Without a point, prints the world's spawn point. """ if len(command): point = self.readPoint(command) point = map(int, map(floor, point)) self.level.setPlayerSpawnPosition(point) self.needsSave = True print "Moved spawn point to ", point else: print "Spawn point: ", self.level.playerSpawnPosition() def _dumpsigns(self, command): """ dumpSigns [ <filename> ] Saves the text and location of every sign in the world to a text file. With no filename, saves signs to <worldname>.signs Output is newline-delimited. 5 lines per sign. Coordinates are on the first line, followed by four lines of sign text. For example: [229, 118, -15] "To boldly go where no man has gone before." Coordinates are ordered the same as point inputs: [North/South, Down/Up, East/West] """ if len(command): filename = command[0] else: filename = self.level.displayName + ".signs" # We happen to encode the output file in UTF-8 too, although # we could use another UTF encoding. The '-sig' encoding puts # a signature at the start of the output file that tools such # as Microsoft Windows Notepad and Emacs understand to mean # the file has UTF-8 encoding. outFile = codecs.open(filename, "w", encoding='utf-8-sig') print "Dumping signs..." signCount = 0 for i, cPos in enumerate(self.level.allChunks): try: chunk = self.level.getChunk(*cPos) except mclevelbase.ChunkMalformed: continue for tileEntity in chunk.TileEntities: if tileEntity["id"].value == "Sign": signCount += 1 outFile.write(str(map(lambda x: tileEntity[x].value, "xyz")) + "\n") for i in range(4): signText = tileEntity["Text{0}".format(i + 1)].value outFile.write(signText + u"\n") if i % 100 == 0: print "Chunk {0}...".format(i) print "Dumped {0} signs to {1}".format(signCount, filename) outFile.close() def _region(self, command): """ region [rx rz] List region files in this world. """ level = self.level assert(isinstance(level, mclevel.MCInfdevOldLevel)) assert level.version def getFreeSectors(rf): runs = [] start = None count = 0 for i, free in enumerate(rf.freeSectors): if free: if start is None: start = i count = 1 else: count += 1 else: if start is None: pass else: runs.append((start, count)) start = None count = 0 return runs def printFreeSectors(runs): for i, (start, count) in enumerate(runs): if i % 4 == 3: print "" print "{start:>6}+{count:<4}".format(**locals()), print "" if len(command): if len(command) > 1: rx, rz = map(int, command[:2]) print "Calling allChunks to preload region files: %d chunks" % len(level.allChunks) rf = level.regionFiles.get((rx, rz)) if rf is None: print "Region {rx},{rz} not found.".format(**locals()) return print "Region {rx:6}, {rz:6}: {used}/{sectors} sectors".format(used=rf.usedSectors, sectors=rf.sectorCount) print "Offset Table:" for cx in range(32): for cz in range(32): if cz % 4 == 0: print "" print "{0:3}, {1:3}: ".format(cx, cz), off = rf.getOffset(cx, cz) sector, length = off >> 8, off & 0xff print "{sector:>6}+{length:<2} ".format(**locals()), print "" runs = getFreeSectors(rf) if len(runs): print "Free sectors:", printFreeSectors(runs) else: if command[0] == "free": print "Calling allChunks to preload region files: %d chunks" % len(level.allChunks) for (rx, rz), rf in level.regionFiles.iteritems(): runs = getFreeSectors(rf) if len(runs): print "R {0:3}, {1:3}:".format(rx, rz), printFreeSectors(runs) else: print "Calling allChunks to preload region files: %d chunks" % len(level.allChunks) coords = (r for r in level.regionFiles) for i, (rx, rz) in enumerate(coords): print "({rx:6}, {rz:6}): {count}, ".format(count=level.regionFiles[rx, rz].chunkCount), if i % 5 == 4: print "" def _repair(self, command): """ repair Attempt to repair inconsistent region files. MAKE A BACKUP. WILL DELETE YOUR DATA. Scans for and repairs errors in region files: Deletes chunks whose sectors overlap with another chunk Rearranges chunks that are in the wrong slot in the offset table Deletes completely unreadable chunks Only usable with region-format saves. """ if self.level.version: self.level.preloadRegions() for rf in self.level.regionFiles.itervalues(): rf.repair() def _dumpchests(self, command): """ dumpChests [ <filename> ] Saves the content and location of every chest in the world to a text file. With no filename, saves signs to <worldname>.chests Output is delimited by brackets and newlines. A set of coordinates in brackets begins a chest, followed by a line for each inventory slot. For example: [222, 51, 22] 2 String 3 String 3 Iron bar Coordinates are ordered the same as point inputs: [North/South, Down/Up, East/West] """ from items import items if len(command): filename = command[0] else: filename = self.level.displayName + ".chests" outFile = file(filename, "w") print "Dumping chests..." chestCount = 0 for i, cPos in enumerate(self.level.allChunks): try: chunk = self.level.getChunk(*cPos) except mclevelbase.ChunkMalformed: continue for tileEntity in chunk.TileEntities: if tileEntity["id"].value == "Chest": chestCount += 1 outFile.write(str(map(lambda x: tileEntity[x].value, "xyz")) + "\n") itemsTag = tileEntity["Items"] if len(itemsTag): for itemTag in itemsTag: try: id = itemTag["id"].value damage = itemTag["Damage"].value item = items.findItem(id, damage) itemname = item.name except KeyError: itemname = "Unknown Item {0}".format(itemTag) except Exception, e: itemname = repr(e) outFile.write("{0} {1}\n".format(itemTag["Count"].value, itemname)) else: outFile.write("Empty Chest\n") if i % 100 == 0: print "Chunk {0}...".format(i) print "Dumped {0} chests to {1}".format(chestCount, filename) outFile.close() def _removeentities(self, command): """ removeEntities [ [except] [ <EntityID> [ <EntityID> ... ] ] ] Remove all entities matching one or more entity IDs. With the except keyword, removes all entities not matching one or more entity IDs. Without any IDs, removes all entities in the world, except for Paintings. Known Mob Entity IDs: Mob Monster Creeper Skeleton Spider Giant Zombie Slime Pig Sheep Cow Chicken Known Item Entity IDs: Item Arrow Snowball Painting Known Vehicle Entity IDs: Minecart Boat Known Dynamic Tile Entity IDs: PrimedTnt FallingSand """ ENT_MATCHTYPE_ANY = 0 ENT_MATCHTYPE_EXCEPT = 1 ENT_MATCHTYPE_NONPAINTING = 2 def match(entityID, matchType, matchWords): if ENT_MATCHTYPE_ANY == matchType: return entityID.lower() in matchWords elif ENT_MATCHTYPE_EXCEPT == matchType: return not (entityID.lower() in matchWords) else: # ENT_MATCHTYPE_EXCEPT == matchType return entityID != "Painting" removedEntities = {} match_words = [] if len(command): if command[0].lower() == "except": command.pop(0) print "Removing all entities except ", command match_type = ENT_MATCHTYPE_EXCEPT else: print "Removing {0}...".format(", ".join(command)) match_type = ENT_MATCHTYPE_ANY match_words = map(lambda x: x.lower(), command) else: print "Removing all entities except Painting..." match_type = ENT_MATCHTYPE_NONPAINTING for cx, cz in self.level.allChunks: chunk = self.level.getChunk(cx, cz) entitiesRemoved = 0 for entity in list(chunk.Entities): entityID = entity["id"].value if match(entityID, match_type, match_words): removedEntities[entityID] = removedEntities.get(entityID, 0) + 1 chunk.Entities.remove(entity) entitiesRemoved += 1 if entitiesRemoved: chunk.chunkChanged(False) if len(removedEntities) == 0: print "No entities to remove." else: print "Removed entities:" for entityID in sorted(removedEntities.keys()): print " {0}: {1:6}".format(entityID, removedEntities[entityID]) self.needsSave = True def _createchunks(self, command): """ createChunks <box> Creates any chunks not present in the specified region. New chunks are filled with only air. New chunks are written to disk immediately. """ if len(command) == 0: self.printUsage("createchunks") return box = self.readBox(command) chunksCreated = self.level.createChunksInBox(box) print "Created {0} chunks." .format(len(chunksCreated)) self.needsSave = True def _deletechunks(self, command): """ deleteChunks <box> Removes all chunks contained in the specified region. Chunks are deleted from disk immediately. """ if len(command) == 0: self.printUsage("deletechunks") return box = self.readBox(command) deletedChunks = self.level.deleteChunksInBox(box) print "Deleted {0} chunks." .format(len(deletedChunks)) def _prune(self, command): """ prune <box> Removes all chunks not contained in the specified region. Useful for enforcing a finite map size. Chunks are deleted from disk immediately. """ if len(command) == 0: self.printUsage("prune") return box = self.readBox(command) i = 0 for cx, cz in list(self.level.allChunks): if cx < box.mincx or cx >= box.maxcx or cz < box.mincz or cz >= box.maxcz: self.level.deleteChunk(cx, cz) i += 1 print "Pruned {0} chunks." .format(i) def _relight(self, command): """ relight [ <box> ] Recalculates lights in the region specified. If omitted, recalculates the entire world. """ if len(command): box = self.readBox(command) chunks = itertools.product(range(box.mincx, box.maxcx), range(box.mincz, box.maxcz)) else: chunks = self.level.allChunks self.level.generateLights(chunks) print "Relit 0 chunks." self.needsSave = True def _create(self, command): """ create [ <filename> ] Create and load a new Minecraft Alpha world. This world will have no chunks and a random terrain seed. If run from the shell, filename is not needed because you already specified a filename earlier in the command. For example: mce.py MyWorld create """ if len(command) < 1: raise UsageError("Expected a filename") filename = command[0] if not os.path.exists(filename): os.mkdir(filename) if not os.path.isdir(filename): raise IOError("{0} already exists".format(filename)) if mclevel.MCInfdevOldLevel.isLevel(filename): raise IOError("{0} is already a Minecraft Alpha world".format(filename)) level = mclevel.MCInfdevOldLevel(filename, create=True) self.level = level def _degrief(self, command): """ degrief [ <height> ] Reverse a few forms of griefing by removing Adminium, Obsidian, Fire, and Lava wherever they occur above the specified height. Without a height, uses height level 32. Removes natural surface lava. Also see removeEntities """ box = self.level.bounds box = BoundingBox(box.origin + (0, 32, 0), box.size - (0, 32, 0)) if len(command): try: box.miny = int(command[0]) except ValueError: pass print "Removing grief matter and surface lava above height {0}...".format(box.miny) self.level.fillBlocks(box, self.level.materials.Air, blocksToReplace=[self.level.materials.Bedrock, self.level.materials.Obsidian, self.level.materials.Fire, self.level.materials.LavaActive, self.level.materials.Lava, ] ) self.needsSave = True def _time(self, command): """ time [time of day] Set or display the time of day. Acceptable values are "morning", "noon", "evening", "midnight", or a time of day such as 8:02, 12:30 PM, or 16:45. """ ticks = self.level.Time timeOfDay = ticks % 24000 ageInTicks = ticks - timeOfDay if len(command) == 0: days = ageInTicks / 24000 hours = timeOfDay / 1000 clockHours = (hours + 6) % 24 ampm = ("AM", "PM")[clockHours > 11] minutes = (timeOfDay % 1000) / 60 print "It is {0}:{1:02} {2} on Day {3}".format(clockHours % 12 or 12, minutes, ampm, days) else: times = {"morning": 6, "noon": 12, "evening": 18, "midnight": 24} word = command[0] minutes = 0 if word in times: hours = times[word] else: try: if ":" in word: h, m = word.split(":") hours = int(h) minutes = int(m) else: hours = int(word) except Exception, e: raise UsageError(("Cannot interpret time, ", e)) if len(command) > 1: if command[1].lower() == "pm": hours += 12 ticks = ageInTicks + hours * 1000 + minutes * 1000 / 60 - 6000 if ticks < 0: ticks += 18000 ampm = ("AM", "PM")[hours > 11 and hours < 24] print "Changed time to {0}:{1:02} {2}".format(hours % 12 or 12, minutes, ampm) self.level.Time = ticks self.needsSave = True def _randomseed(self, command): """ randomseed [ <seed> ] Set or display the world's random seed, a 64-bit integer that uniquely defines the world's terrain. """ if len(command): try: seed = long(command[0]) except ValueError: raise UsageError("Expected a long integer.") self.level.RandomSeed = seed self.needsSave = True else: print "Random Seed: ", self.level.RandomSeed def _gametype(self, command): """ gametype [ <player> [ <gametype> ] ] Set or display the player's game type, an integer that identifies whether their game is survival (0) or creative (1). On single-player worlds, the player is just 'Player'. """ if len(command) == 0: print "Players: " for player in self.level.players: print " {0}: {1}".format(player, self.level.getPlayerGameType(player)) return player = command.pop(0) if len(command) == 0: print "Player {0}: {1}".format(player, self.level.getPlayerGameType(player)) return try: gametype = int(command[0]) except ValueError: raise UsageError("Expected an integer.") self.level.setPlayerGameType(gametype, player) self.needsSave = True def _worldsize(self, command): """ worldsize Computes and prints the dimensions of the world. For infinite worlds, also prints the most negative corner. """ bounds = self.level.bounds if isinstance(self.level, mclevel.MCInfdevOldLevel): print "\nWorld size: \n {0[0]:7} north to south\n {0[2]:7} east to west\n".format(bounds.size) print "Smallest and largest points: ({0[0]},{0[2]}), ({1[0]},{1[2]})".format(bounds.origin, bounds.maximum) else: print "\nWorld size: \n {0[0]:7} wide\n {0[1]:7} tall\n {0[2]:7} long\n".format(bounds.size) def _heightmap(self, command): """ heightmap <filename> Takes a png and imports it as the terrain starting at chunk 0,0. Data is internally converted to greyscale and scaled to the maximum height. The game will fill the terrain with trees and mineral deposits the next time you play the level. Please please please try out a small test image before using a big source. Using the levels tool to get a good heightmap is an art, not a science. A smaller map lets you experiment and get it right before having to blow all night generating the really big map. Requires the PIL library. """ if len(command) == 0: self.printUsage("heightmap") return if not sys.stdin.isatty() or raw_input( "This will destroy a large portion of the map and may take a long time. Did you really want to do this?" ).lower() in ("yes", "y", "1", "true"): from PIL import Image import datetime filename = command.pop(0) imgobj = Image.open(filename) greyimg = imgobj.convert("L") # luminance del imgobj width, height = greyimg.size water_level = 64 xchunks = (height + 15) / 16 zchunks = (width + 15) / 16 start = datetime.datetime.now() for cx in range(xchunks): for cz in range(zchunks): try: self.level.createChunk(cx, cz) except: pass c = self.level.getChunk(cx, cz) imgarray = numpy.asarray(greyimg.crop((cz * 16, cx * 16, cz * 16 + 16, cx * 16 + 16))) imgarray = imgarray / 2 # scale to 0-127 for x in range(16): for z in range(16): if z + (cz * 16) < width - 1 and x + (cx * 16) < height - 1: # world dimension X goes north-south # first array axis goes up-down h = imgarray[x, z] c.Blocks[x, z, h + 1:] = 0 # air c.Blocks[x, z, h:h + 1] = 2 # grass c.Blocks[x, z, h - 4:h] = 3 # dirt c.Blocks[x, z, :h - 4] = 1 # rock if h < water_level: c.Blocks[x, z, h + 1:water_level] = 9 # water if h < water_level + 2: c.Blocks[x, z, h - 2:h + 1] = 12 # sand if it's near water level c.Blocks[x, z, 0] = 7 # bedrock c.chunkChanged() c.TerrainPopulated = False # the quick lighting from chunkChanged has already lit this simple terrain completely c.needsLighting = False logging.info("%s Just did chunk %d,%d" % (datetime.datetime.now().strftime("[%H:%M:%S]"), cx, cz)) logging.info("Done with mapping!") self.needsSave = True stop = datetime.datetime.now() logging.info("Took %s." % str(stop - start)) spawnz = width / 2 spawnx = height / 2 spawny = greyimg.getpixel((spawnx, spawnz)) logging.info("You probably want to change your spawn point. I suggest {0}".format((spawnx, spawny, spawnz))) def _execute(self, command): """ execute <filename> Execute all commands in a file and save. """ if len(command) == 0: print "You must give the file with commands to execute" else: commandFile = open(command[0], "r") commandsFromFile = commandFile.readlines() for commandFromFile in commandsFromFile: print commandFromFile self.processCommand(commandFromFile) self._save("") def _quit(self, command): """ quit [ yes | no ] Quits the program. Without 'yes' or 'no', prompts to save before quitting. In batch mode, an end of file automatically saves the level. """ if len(command) == 0 or not (command[0].lower() in ("yes", "no")): if raw_input("Save before exit? ").lower() in ("yes", "y", "1", "true"): self._save(command) raise SystemExit if len(command) and command[0].lower == "yes": self._save(command) raise SystemExit def _exit(self, command): self._quit(command) def _save(self, command): if self.needsSave: self.level.generateLights() self.level.saveInPlace() self.needsSave = False def _load(self, command): """ load [ <filename> | <world number> ] Loads another world, discarding all changes to this world. """ if len(command) == 0: self.printUsage("load") self.loadWorld(command[0]) def _reload(self, command): self.level = mclevel.fromFile(self.level.filename) def _dimension(self, command): """ dimension [ <dim> ] Load another dimension, a sub-world of this level. Without options, lists all of the dimensions found in this world. <dim> can be a number or one of these keywords: nether, hell, slip: DIM-1 earth, overworld, parent: parent world end: DIM1 """ if len(command): if command[0].lower() in ("earth", "overworld", "parent"): if self.level.parentWorld: self.level = self.level.parentWorld return else: print "You are already on earth." return elif command[0].lower() in ("hell", "nether", "slip"): dimNo = -1 elif command[0].lower() == "end": dimNo = 1 else: dimNo = self.readInt(command) if dimNo in self.level.dimensions: self.level = self.level.dimensions[dimNo] return if self.level.parentWorld: print u"Parent world: {0} ('dimension parent' to return)".format(self.level.parentWorld.displayName) if len(self.level.dimensions): print u"Dimensions in {0}:".format(self.level.displayName) for k in self.level.dimensions: print "{0}: {1}".format(k, infiniteworld.MCAlphaDimension.dimensionNames.get(k, "Unknown")) def _help(self, command): if len(command): self.printUsage(command[0]) else: self.printUsage() def _blocks(self, command): """ blocks [ <block name> | <block ID> ] Prints block IDs matching the name, or the name matching the ID. With nothing, prints a list of all blocks. """ searchName = None if len(command): searchName = " ".join(command) try: searchNumber = int(searchName) except ValueError: searchNumber = None matches = self.level.materials.blocksMatching(searchName) else: matches = [b for b in self.level.materials.allBlocks if b.ID == searchNumber] # print "{0:3}: {1}".format(searchNumber, self.level.materials.names[searchNumber]) # return else: matches = self.level.materials.allBlocks print "{id:9} : {name} {aka}".format(id="(ID:data)", name="Block name", aka="[Other names]") for b in sorted(matches): idstring = "({ID}:{data})".format(ID=b.ID, data=b.blockData) aka = b.aka and " [{aka}]".format(aka=b.aka) or "" print "{idstring:9} : {name} {aka}".format(idstring=idstring, name=b.name, aka=aka) def printUsage(self, command=""): if command.lower() in self.commands: print "Usage: ", self.commandUsage(command.lower()) else: print self.__doc__.format(commandPrefix=("", "mce.py <world> ")[not self.batchMode]) def printUsageAndQuit(self): self.printUsage() raise SystemExit def loadWorld(self, world): worldpath = os.path.expanduser(world) if os.path.exists(worldpath): self.level = mclevel.fromFile(worldpath) else: self.level = mclevel.loadWorld(world) level = None batchMode = False def run(self): logging.basicConfig(format=u'%(levelname)s:%(message)s') logging.getLogger().level = logging.INFO sys.argv.pop(0) if len(sys.argv): world = sys.argv.pop(0) if world.lower() in ("-h", "--help"): self.printUsageAndQuit() if len(sys.argv) and sys.argv[0].lower() == "create": # accept the syntax, "mce world3 create" self._create([world]) print "Created world {0}".format(world) sys.exit(0) else: self.loadWorld(world) else: self.batchMode = True self.printUsage() while True: try: world = raw_input("Please enter world name or path to world folder: ") self.loadWorld(world) except EOFError, e: print "End of input." raise SystemExit except Exception, e: print "Cannot open {0}: {1}".format(world, e) else: break if len(sys.argv): # process one command from command line try: self.processCommand(" ".join(sys.argv)) except UsageError: self.printUsageAndQuit() self._save([]) else: # process many commands on standard input, maybe interactively command = [""] self.batchMode = True while True: try: command = raw_input(u"{0}> ".format(self.level.displayName)) print self.processCommand(command) except EOFError, e: print "End of file. Saving automatically." self._save([]) raise SystemExit except Exception, e: if self.debug: traceback.print_exc() print 'Exception during command: {0!r}'.format(e) print "Use 'debug' to enable tracebacks." # self.printUsage() def processCommand(self, command): command = command.strip() if len(command) == 0: return if command[0] == "#": return commandWords = command.split() keyword = commandWords.pop(0).lower() if not keyword in self.commands: matches = filter(lambda x: x.startswith(keyword), self.commands) if len(matches) == 1: keyword = matches[0] elif len(matches): print "Ambiguous command. Matches: " for k in matches: print " ", k return else: raise UsageError("Command {0} not recognized.".format(keyword)) func = getattr(self, "_" + keyword) try: func(commandWords) except PlayerNotFound, e: print "Cannot find player {0}".format(e.args[0]) self._player([]) except UsageError, e: print e if self.debug: traceback.print_exc() self.printUsage(keyword) def main(argv): profile = os.getenv("MCE_PROFILE", None) editor = mce() if profile: print "Profiling enabled" import cProfile cProfile.runctx('editor.run()', locals(), globals(), profile) else: editor.run() return 0 if __name__ == '__main__': sys.exit(main(sys.argv))
isc
mcedit/pymclevel
box.py
3
6660
from collections import namedtuple import itertools import math _Vector = namedtuple("_Vector", ("x", "y", "z")) class Vector(_Vector): __slots__ = () def __add__(self, other): return Vector(self[0] + other[0], self[1] + other[1], self[2] + other[2]) def __sub__(self, other): return Vector(self[0] - other[0], self[1] - other[1], self[2] - other[2]) def __mul__(self, other): if isinstance(other, (int, float)): return Vector(self[0] * other, self[1] * other, self[2] * other) return Vector(self[0] * other[0], self[1] * other[1], self[2] * other[2]) def __truediv__(self, other): if isinstance(other, (int, float)): return Vector(self[0] / other, self[1] / other, self[2] / other) return Vector(self[0] / other[0], self[1] / other[1], self[2] / other[2]) __div__ = __truediv__ def length(self): return math.sqrt(self[0] * self[0] + self[1] * self[1] + self[2] * self[2]) def normalize(self): l = self.length() if l == 0: return self return self / l def intfloor(self): return Vector(*[int(math.floor(p)) for p in self]) class BoundingBox (object): type = int def __init__(self, origin=(0, 0, 0), size=(0, 0, 0)): if isinstance(origin, BoundingBox): self._origin = origin._origin self._size = origin._size else: self._origin, self._size = Vector(*(self.type(a) for a in origin)), Vector(*(self.type(a) for a in size)) def __repr__(self): return "BoundingBox({0}, {1})".format(self.origin, self.size) @property def origin(self): "The smallest position in the box" return self._origin @property def size(self): "The size of the box" return self._size @property def width(self): "The dimension along the X axis" return self._size.x @property def height(self): "The dimension along the Y axis" return self._size.y @property def length(self): "The dimension along the Z axis" return self._size.z @property def minx(self): return self.origin.x @property def miny(self): return self.origin.y @property def minz(self): return self.origin.z @property def maxx(self): return self.origin.x + self.size.x @property def maxy(self): return self.origin.y + self.size.y @property def maxz(self): return self.origin.z + self.size.z @property def maximum(self): "The largest point of the box; origin plus size." return self._origin + self._size @property def volume(self): "The volume of the box in blocks" return self.size.x * self.size.y * self.size.z @property def positions(self): """iterate through all of the positions within this selection box""" return itertools.product( xrange(self.minx, self.maxx), xrange(self.miny, self.maxy), xrange(self.minz, self.maxz) ) def intersect(self, box): """ Return a box containing the area self and box have in common. Box will have zero volume if there is no common area. """ if (self.minx > box.maxx or self.maxx < box.minx or self.miny > box.maxy or self.maxy < box.miny or self.minz > box.maxz or self.maxz < box.minz): #Zero size intersection. return BoundingBox() origin = Vector( max(self.minx, box.minx), max(self.miny, box.miny), max(self.minz, box.minz), ) maximum = Vector( min(self.maxx, box.maxx), min(self.maxy, box.maxy), min(self.maxz, box.maxz), ) #print "Intersect of {0} and {1}: {2}".format(self, box, newbox) return BoundingBox(origin, maximum - origin) def union(self, box): """ Return a box large enough to contain both self and box. """ origin = Vector( min(self.minx, box.minx), min(self.miny, box.miny), min(self.minz, box.minz), ) maximum = Vector( max(self.maxx, box.maxx), max(self.maxy, box.maxy), max(self.maxz, box.maxz), ) return BoundingBox(origin, maximum - origin) def expand(self, dx, dy=None, dz=None): """ Return a new box with boundaries expanded by dx, dy, dz. If only dx is passed, expands by dx in all dimensions. """ if dz is None: dz = dx if dy is None: dy = dx origin = self.origin - (dx, dy, dz) size = self.size + (dx * 2, dy * 2, dz * 2) return BoundingBox(origin, size) def __contains__(self, pos): x, y, z = pos if x < self.minx or x >= self.maxx: return False if y < self.miny or y >= self.maxy: return False if z < self.minz or z >= self.maxz: return False return True def __cmp__(self, b): return cmp((self.origin, self.size), (b.origin, b.size)) # --- Chunk positions --- @property def mincx(self): "The smallest chunk position contained in this box" return self.origin.x >> 4 @property def mincz(self): "The smallest chunk position contained in this box" return self.origin.z >> 4 @property def maxcx(self): "The largest chunk position contained in this box" return ((self.origin.x + self.size.x - 1) >> 4) + 1 @property def maxcz(self): "The largest chunk position contained in this box" return ((self.origin.z + self.size.z - 1) >> 4) + 1 def chunkBox(self, level): """Returns this box extended to the chunk boundaries of the given level""" box = self return BoundingBox((box.mincx << 4, 0, box.mincz << 4), (box.maxcx - box.mincx << 4, level.Height, box.maxcz - box.mincz << 4)) @property def chunkPositions(self): #iterate through all of the chunk positions within this selection box return itertools.product(xrange(self.mincx, self.maxcx), xrange(self.mincz, self.maxcz)) @property def chunkCount(self): return (self.maxcx - self.mincx) * (self.maxcz - self.mincz) @property def isChunkAligned(self): return (self.origin.x & 0xf == 0) and (self.origin.z & 0xf == 0) class FloatBox (BoundingBox): type = float
isc
mcedit/mcedit
albow/file_dialogs.py
1
9481
# -*- coding: utf-8 -*- # # Albow - File Dialogs # import os from pygame import draw, Rect from pygame.locals import * from albow.widget import Widget from albow.dialogs import Dialog, ask, alert from albow.controls import Label, Button from albow.fields import TextField from albow.layout import Row, Column from albow.palette_view import PaletteView from albow.theme import ThemeProperty class DirPathView(Widget): def __init__(self, width, client, **kwds): Widget.__init__(self, **kwds) self.set_size_for_text(width) self.client = client def draw(self, surf): frame = self.get_margin_rect() image = self.font.render(self.client.directory, True, self.fg_color) tw = image.get_width() mw = frame.width if tw <= mw: x = 0 else: x = mw - tw surf.blit(image, (frame.left + x, frame.top)) class FileListView(PaletteView): #scroll_button_color = (255, 255, 0) def __init__(self, width, client, **kwds): font = self.predict_font(kwds) h = font.get_linesize() d = 2 * self.predict(kwds, 'margin') PaletteView.__init__(self, (width - d, h), 10, 1, scrolling=True, **kwds) self.client = client self.selection = None self.names = [] def update(self): client = self.client dir = client.directory suffixes = client.suffixes def filter(name): path = os.path.join(dir, name) return os.path.isdir(path) or self.client.filter(path) try: names = [name for name in os.listdir(dir) if filter(name)] #if not name.startswith(".") and filter(name)] except EnvironmentError, e: alert(u"%s: %s" % (dir, e)) names = [] self.names = sorted(names) self.selection = None self.scroll = 0 def num_items(self): return len(self.names) #def draw_prehighlight(self, surf, item_no, rect): # draw.rect(surf, self.sel_color, rect) def draw_item(self, surf, item_no, rect): font = self.font color = self.fg_color buf = self.font.render(self.names[item_no], True, color) surf.blit(buf, rect) def click_item(self, item_no, e): self.selection = item_no self.client.dir_box_click(e.num_clicks == 2) def item_is_selected(self, item_no): return item_no == self.selection def get_selected_name(self): sel = self.selection if sel is not None: return self.names[sel] else: return "" class FileDialog(Dialog): box_width = 250 default_prompt = None up_button_text = ThemeProperty("up_button_text") def __init__(self, prompt=None, suffixes=None, **kwds): Dialog.__init__(self, **kwds) label = None d = self.margin self.suffixes = suffixes or ("",) up_button = Button(self.up_button_text, action=self.go_up) dir_box = DirPathView(self.box_width - up_button.width - 10, self) self.dir_box = dir_box top_row = Row([dir_box, up_button]) list_box = FileListView(self.box_width - 16, self) self.list_box = list_box ctrls = [top_row, list_box] prompt = prompt or self.default_prompt if prompt: label = Label(prompt) if self.saving: filename_box = TextField(self.box_width) filename_box.change_action = self.update filename_box._enter_action = filename_box.enter_action filename_box.enter_action = self.enter_action self.filename_box = filename_box ctrls.append(Column([label, filename_box], align='l', spacing=0)) else: if label: ctrls.insert(0, label) ok_button = Button(self.ok_label, action=self.ok, enable=self.ok_enable) self.ok_button = ok_button cancel_button = Button("Cancel", action=self.cancel) vbox = Column(ctrls, align='l', spacing=d) vbox.topleft = (d, d) y = vbox.bottom + d ok_button.topleft = (vbox.left, y) cancel_button.topright = (vbox.right, y) self.add(vbox) self.add(ok_button) self.add(cancel_button) self.shrink_wrap() self._directory = None self.directory = os.getcwdu() #print "FileDialog: cwd =", repr(self.directory) ### if self.saving: filename_box.focus() def get_directory(self): return self._directory def set_directory(self, x): x = os.path.abspath(x) while not os.path.exists(x): y = os.path.dirname(x) if y == x: x = os.getcwdu() break x = y if self._directory != x: self._directory = x self.list_box.update() self.update() directory = property(get_directory, set_directory) def filter(self, path): suffixes = self.suffixes if not suffixes or os.path.isdir(path): #return os.path.isfile(path) return True for suffix in suffixes: if path.endswith(suffix.lower()): return True def update(self): pass def go_up(self): self.directory = os.path.dirname(self.directory) self.list_box.scroll_to_item(0) def dir_box_click(self, double): if double: name = self.list_box.get_selected_name() path = os.path.join(self.directory, name) suffix = os.path.splitext(name)[1] if suffix not in self.suffixes and os.path.isdir(path): self.directory = path else: self.double_click_file(name) self.update() def enter_action(self): self.filename_box._enter_action() self.ok() def ok(self): self.dir_box_click(True) #self.dismiss(True) def cancel(self): self.dismiss(False) def key_down(self, evt): k = evt.key if k == K_RETURN or k == K_KP_ENTER: self.dir_box_click(True) if k == K_ESCAPE: self.cancel() class FileSaveDialog(FileDialog): saving = True default_prompt = "Save as:" ok_label = "Save" def get_filename(self): return self.filename_box.value def set_filename(self, x): dsuf = self.suffixes[0] if x.endswith(dsuf): x = x[:-len(dsuf)] self.filename_box.value = x filename = property(get_filename, set_filename) def get_pathname(self): path = os.path.join(self.directory, self.filename_box.value) suffixes = self.suffixes if suffixes and not path.endswith(suffixes[0]): path = path + suffixes[0] return path pathname = property(get_pathname) def double_click_file(self, name): self.filename_box.value = name def ok(self): path = self.pathname if os.path.exists(path): answer = ask("Replace existing '%s'?" % os.path.basename(path)) if answer != "OK": return #FileDialog.ok(self) self.dismiss(True) def update(self): FileDialog.update(self) def ok_enable(self): return self.filename_box.text != "" class FileOpenDialog(FileDialog): saving = False ok_label = "Open" def get_pathname(self): name = self.list_box.get_selected_name() if name: return os.path.join(self.directory, name) else: return None pathname = property(get_pathname) #def update(self): # FileDialog.update(self) def ok_enable(self): path = self.pathname enabled = self.item_is_choosable(path) return enabled def item_is_choosable(self, path): return bool(path) and self.filter(path) def double_click_file(self, name): self.dismiss(True) class LookForFileDialog(FileOpenDialog): target = None def __init__(self, target, **kwds): FileOpenDialog.__init__(self, **kwds) self.target = target def item_is_choosable(self, path): return path and os.path.basename(path) == self.target def filter(self, name): return name and os.path.basename(name) == self.target def request_new_filename(prompt=None, suffix=None, extra_suffixes=None, directory=None, filename=None, pathname=None): if pathname: directory, filename = os.path.split(pathname) if extra_suffixes: suffixes = extra_suffixes else: suffixes = [] if suffix: suffixes = [suffix] + suffixes dlog = FileSaveDialog(prompt=prompt, suffixes=suffixes) if directory: dlog.directory = directory if filename: dlog.filename = filename if dlog.present(): return dlog.pathname else: return None def request_old_filename(suffixes=None, directory=None): dlog = FileOpenDialog(suffixes=suffixes) if directory: dlog.directory = directory if dlog.present(): return dlog.pathname else: return None def look_for_file_or_directory(target, prompt=None, directory=None): dlog = LookForFileDialog(target=target, prompt=prompt) if directory: dlog.directory = directory if dlog.present(): return dlog.pathname else: return None
isc
mozilla-services/tecken
tecken/upload/views.py
1
18470
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at https://mozilla.org/MPL/2.0/. import re import logging import fnmatch import zipfile import hashlib import os import time import concurrent.futures import markus from encore.concurrent.futures.synchronous import SynchronousExecutor from django import http from django.conf import settings from django.utils import timezone from django.core.exceptions import ImproperlyConfigured from django.views.decorators.csrf import csrf_exempt from tecken.base.decorators import ( api_login_required, api_any_permission_required, api_require_POST, make_tempdir, ) from tecken.base.utils import filesizeformat, invalid_key_name_characters from tecken.upload.forms import UploadByDownloadForm, UploadByDownloadRemoteError from tecken.upload.models import Upload, UploadsCreated from tecken.upload.utils import ( dump_and_extract, UnrecognizedArchiveFileExtension, DuplicateFileDifferentSize, upload_file_upload, ) from tecken.librequests import session_with_retries from tecken.storage import StorageBucket logger = logging.getLogger("tecken") metrics = markus.get_metrics("tecken") class NoPossibleBucketName(Exception): """When you tried to specify a preferred bucket name but it never matched to one you can use.""" _not_hex_characters = re.compile(r"[^a-f0-9]", re.I) # This list of filenames is used to validate a zip and also when iterating # over the extracted zip. # The names of files in this list are considered harmless and something that # can simply be ignored. _ignorable_filenames = (".DS_Store",) def check_symbols_archive_file_listing(file_listings): """return a string (the error) if there was something not as expected""" for file_listing in file_listings: for snippet in settings.DISALLOWED_SYMBOLS_SNIPPETS: if snippet in file_listing.name: return ( f"Content of archive file contains the snippet " f"'{snippet}' which is not allowed" ) # Now check that the filename is matching according to these rules: # 1. Either /<name1>/hex/<name2>, # 2. Or, /<name>-symbols.txt # Anything else should be considered and unrecognized file pattern # and thus rejected. split = file_listing.name.split("/") if split[-1] in _ignorable_filenames: continue if len(split) == 3: # Check the symbol and the filename part of it to make sure # it doesn't contain any, considered, invalid S3 characters # when it'd become a key. if invalid_key_name_characters(split[0] + split[2]): return f"Invalid character in filename {file_listing.name!r}" # Check that the middle part is only hex characters. if not _not_hex_characters.findall(split[1]): continue elif len(split) == 1: if file_listing.name.lower().endswith("-symbols.txt"): continue # If it didn't get "continued" above, it's an unrecognized file # pattern. return ( "Unrecognized file pattern. Should only be <module>/<hex>/<file> " "or <name>-symbols.txt and nothing else. " f"(First unrecognized pattern was {file_listing.name})" ) def get_bucket_info(user, try_symbols=None, preferred_bucket_name=None): """return an object that has 'bucket', 'endpoint_url', 'region'. Only 'bucket' is mandatory in the response object. """ if try_symbols is None: # If it wasn't explicitly passed, we need to figure this out by # looking at the user who uploads. # Namely, we're going to see if the user has the permission # 'upload.upload_symbols'. If the user does, it means the user intends # to *not* upload Try build symbols. # This is based on the axiom that, if the upload is made with an # API token, that API token can't have *both* the # 'upload.upload_symbols' permission *and* the # 'upload.upload_try_symbols' permission. # If the user uploads via the web the user has a choice to check # a checkbox that is off by default. If doing so, the user isn't # using an API token, so the user might have BOTH permissions. # Then the default falls on this NOT being a Try upload. try_symbols = not user.has_perm("upload.upload_symbols") if try_symbols: url = settings.UPLOAD_TRY_SYMBOLS_URL else: url = settings.UPLOAD_DEFAULT_URL exceptions = settings.UPLOAD_URL_EXCEPTIONS if preferred_bucket_name: # If the user has indicated a preferred bucket name, check that they have # permission to use it. for url, _ in get_possible_bucket_urls(user): if preferred_bucket_name in url: return StorageBucket(url, try_symbols=try_symbols) raise NoPossibleBucketName(preferred_bucket_name) else: if user.email.lower() in exceptions: # easy exception = exceptions[user.email.lower()] else: # match against every possible wildcard exception = None # assume no match for email_or_wildcard in settings.UPLOAD_URL_EXCEPTIONS: if fnmatch.fnmatch(user.email.lower(), email_or_wildcard.lower()): # a match! exception = settings.UPLOAD_URL_EXCEPTIONS[email_or_wildcard] break if exception: url = exception return StorageBucket(url, try_symbols=try_symbols) def get_possible_bucket_urls(user): """Return list of possible buckets this user can upload to. If the user is specified in UPLOAD_URL_EXCEPTIONS, then the user can only upload into that bucket. If the user is not specified, then the user can upload to the public bucket. :param user: a django user :return: list of tuples of (url, "private"/"public") """ urls = [] exceptions = settings.UPLOAD_URL_EXCEPTIONS email_lower = user.email.lower() for email_pattern in exceptions: if ( email_lower == email_pattern.lower() or fnmatch.fnmatch(email_lower, email_pattern.lower()) or user.is_superuser ): urls.append((exceptions[email_pattern], "private")) # We use UPLOAD_URL_EXCEPTIONS to specify buckets people can upload into. If a # person is specified in UPLOAD_URL_EXCEPTIONS, then they can only upload to that # bucket. If they are not specified, then they can upload to the public bucket. if not urls: urls.append((settings.UPLOAD_DEFAULT_URL, "public")) return urls def _ignore_member_file(filename): """Return true if the given filename (could be a filepath), should be completely ignored in the upload process. At the moment the list is "allow-list based", meaning all files are processed and uploaded to S3 unless it meets certain checks. """ if filename.lower().endswith("-symbols.txt"): return True if os.path.basename(filename) in _ignorable_filenames: return True return False @metrics.timer_decorator("upload_archive") @api_require_POST @csrf_exempt @api_login_required @api_any_permission_required("upload.upload_symbols", "upload.upload_try_symbols") @make_tempdir(settings.UPLOAD_TEMPDIR_PREFIX) def upload_archive(request, upload_dir): try: for name in request.FILES: upload_ = request.FILES[name] file_listing = dump_and_extract(upload_dir, upload_, name) size = upload_.size url = None redirect_urls = None break else: if request.POST.get("url"): form = UploadByDownloadForm(request.POST) try: is_valid = form.is_valid() except UploadByDownloadRemoteError as exception: return http.JsonResponse({"error": str(exception)}, status=500) if is_valid: url = form.cleaned_data["url"] name = form.cleaned_data["upload"]["name"] size = form.cleaned_data["upload"]["size"] size_fmt = filesizeformat(size) logger.info(f"Download to upload {url} ({size_fmt})") redirect_urls = form.cleaned_data["upload"]["redirect_urls"] or None download_name = os.path.join(upload_dir, name) session = session_with_retries(default_timeout=(5, 300)) with metrics.timer("upload_download_by_url"): response_stream = session.get(url, stream=True) # NOTE(willkg): The UploadByDownloadForm handles most errors # when it does a HEAD, so this mostly covers transient errors # between the HEAD and this GET request. if response_stream.status_code != 200: return http.JsonResponse( { "error": "non-200 status code when retrieving %s" % url }, status=400, ) with open(download_name, "wb") as f: # Read 1MB at a time chunk_size = 1024 * 1024 stream = response_stream.iter_content(chunk_size=chunk_size) count_chunks = 0 start = time.time() for chunk in stream: if chunk: # filter out keep-alive new chunks f.write(chunk) count_chunks += 1 end = time.time() total_size = chunk_size * count_chunks download_speed = size / (end - start) logger.info( f"Read {count_chunks} chunks of " f"{filesizeformat(chunk_size)} each " f"totalling {filesizeformat(total_size)} " f"({filesizeformat(download_speed)}/s)." ) file_listing = dump_and_extract(upload_dir, download_name, name) os.remove(download_name) else: for key, errors in form.errors.as_data().items(): return http.JsonResponse( {"error": errors[0].message}, status=400 ) else: return http.JsonResponse( { "error": ( "Must be multipart form data with at " "least one file" ) }, status=400, ) except zipfile.BadZipfile as exception: return http.JsonResponse({"error": str(exception)}, status=400) except UnrecognizedArchiveFileExtension as exception: return http.JsonResponse( {"error": f'Unrecognized archive file extension "{exception}"'}, status=400 ) except DuplicateFileDifferentSize as exception: return http.JsonResponse({"error": str(exception)}, status=400) error = check_symbols_archive_file_listing(file_listing) if error: return http.JsonResponse({"error": error.strip()}, status=400) # If you pass an extract argument, independent of value, with key 'try' # then we definitely knows this is a Try symbols upload. is_try_upload = request.POST.get("try") # If you have special permission, you can affect which bucket to upload to. preferred_bucket_name = request.POST.get("bucket_name") try: bucket_info = get_bucket_info( request.user, try_symbols=is_try_upload, preferred_bucket_name=preferred_bucket_name, ) except NoPossibleBucketName as exception: logger.warning(f"No possible bucket for {request.user!r} ({exception})") return http.JsonResponse({"error": "No valid bucket"}, status=403) if is_try_upload is None: # If 'is_try_upload' isn't immediately true by looking at the # request.POST parameters, the get_bucket_info() function can # figure it out too. is_try_upload = bucket_info.try_symbols else: # In case it's passed in as a string is_try_upload = bool(is_try_upload) if not bucket_info.exists(): raise ImproperlyConfigured(f"Bucket does not exist: {bucket_info!r}") # Create the client for upload_file_upload # TODO(jwhitlock): implement backend details in StorageBucket API client = bucket_info.get_storage_client( read_timeout=settings.S3_PUT_READ_TIMEOUT, connect_timeout=settings.S3_PUT_CONNECT_TIMEOUT, ) # Use a different client for doing the lookups. # That's because we don't want the size lookup to severly accumulate # in the case of there being some unpredictable slowness. # When that happens the lookup is quickly cancelled and it assumes # the file does not exist. # See http://botocore.readthedocs.io/en/latest/reference/config.html#botocore.config.Config # noqa lookup_client = bucket_info.get_storage_client( read_timeout=settings.S3_LOOKUP_READ_TIMEOUT, connect_timeout=settings.S3_LOOKUP_CONNECT_TIMEOUT, ) # Every key has a prefix. If the StorageBucket instance has it's own prefix # prefix that first :) prefix = settings.SYMBOL_FILE_PREFIX if bucket_info.prefix: prefix = f"{bucket_info.prefix}/{prefix}" # Make a hash string that represents every file listing in the archive. # Do this by making a string first out of all files listed. content = "\n".join( f"{x.name}:{x.size}" for x in sorted(file_listing, key=lambda x: x.name) ) # The MD5 is just used to make the temporary S3 file unique in name # if the client uploads with the same filename in quick succession. content_hash = hashlib.md5(content.encode("utf-8")).hexdigest()[:30] # nosec # Always create the Upload object no matter what happens next. # If all individual file uploads work out, we say this is complete. upload_obj = Upload.objects.create( user=request.user, filename=name, bucket_name=bucket_info.name, bucket_region=bucket_info.region, bucket_endpoint_url=bucket_info.endpoint_url, size=size, download_url=url, redirect_urls=redirect_urls, content_hash=content_hash, try_symbols=is_try_upload, ) ignored_keys = [] skipped_keys = [] if settings.SYNCHRONOUS_UPLOAD_FILE_UPLOAD: # This is only applicable when running unit tests thread_pool = SynchronousExecutor() else: thread_pool = concurrent.futures.ThreadPoolExecutor( max_workers=settings.UPLOAD_FILE_UPLOAD_MAX_WORKERS or None ) file_uploads_created = 0 uploaded_symbol_keys = [] key_to_symbol_keys = {} with thread_pool as executor: future_to_key = {} for member in file_listing: if _ignore_member_file(member.name): ignored_keys.append(member.name) continue key_name = os.path.join(prefix, member.name) # We need to know and remember, for every file attempted, # what that name corresponds to as a "symbol key". # A symbol key is, for example, ('xul.pdb', 'A7D6F1BBA7D6F1BB1') symbol_key = tuple(member.name.split("/")[:2]) key_to_symbol_keys[key_name] = symbol_key future_to_key[ executor.submit( upload_file_upload, client, bucket_info.name, key_name, member.path, upload=upload_obj, client_lookup=lookup_client, ) ] = key_name # Now lets wait for them all to finish and we'll see which ones # were skipped and which ones were created. for future in concurrent.futures.as_completed(future_to_key): file_upload = future.result() if file_upload: file_uploads_created += 1 uploaded_symbol_keys.append(key_to_symbol_keys[file_upload.key]) else: skipped_keys.append(future_to_key[future]) metrics.incr("upload_file_upload_skip", 1) if file_uploads_created: logger.info(f"Created {file_uploads_created} FileUpload objects") else: logger.info(f"No file uploads created for {upload_obj!r}") Upload.objects.filter(id=upload_obj.id).update( skipped_keys=skipped_keys or None, ignored_keys=ignored_keys or None, completed_at=timezone.now(), ) # Re-calculate the UploadsCreated for today. # FIXME(willkg): when/if we get a scheduled task runner, we should move this # to that date = timezone.now().date() with metrics.timer("uploads_created_update"): UploadsCreated.update(date) logger.info(f"UploadsCreated updated for {date!r}") metrics.incr( "upload_uploads", tags=[f"try:{is_try_upload}", f"bucket:{bucket_info.name}"] ) return http.JsonResponse({"upload": _serialize_upload(upload_obj)}, status=201) def _serialize_upload(upload): return { "id": upload.id, "size": upload.size, "filename": upload.filename, "bucket": upload.bucket_name, "region": upload.bucket_region, "download_url": upload.download_url, "try_symbols": upload.try_symbols, "redirect_urls": upload.redirect_urls or [], "completed_at": upload.completed_at, "created_at": upload.created_at, "user": upload.user.email, "skipped_keys": upload.skipped_keys or [], }
mpl-2.0
mozilla-services/tecken
tecken/api/urls.py
1
1554
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at https://mozilla.org/MPL/2.0/. from django.urls import path from . import views app_name = "api" # NOTE(peterbe): The endpoints that start with a '_' are basically only relevant for the # sake of the frontend. Meaning, it doesn't make sense to use them in your curl script, # for example. urlpatterns = [ path("_auth/", views.auth, name="auth"), path("stats/", views.stats, name="stats"), path("stats/uploads", views.stats_uploads, name="stats_uploads"), path("tokens/", views.tokens, name="tokens"), path("tokens/token/<int:id>/extend", views.extend_token, name="extend_token"), path("tokens/token/<int:id>", views.delete_token, name="delete_token"), path( "uploads/_possible_upload_urls/", views.possible_upload_urls, name="possible_upload_urls", ), path("uploads/", views.uploads, name="uploads"), path("uploads/created/", views.uploads_created, name="uploads_created"), path( "uploads/created/backfilled/", views.uploads_created_backfilled, name="uploads_created_backfilled", ), path("uploads/files/", views.upload_files, name="upload_files"), path("uploads/files/file/<int:id>", views.upload_file, name="upload_file"), path("uploads/upload/<int:id>", views.upload, name="upload"), path("downloads/missing/", views.downloads_missing, name="downloads_missing"), ]
mpl-2.0
mozilla-services/tecken
bin/debug-sym-file.py
1
1332
#!/usr/bin/env python # This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at https://mozilla.org/MPL/2.0/. # Prints information about a sym file including whether it kicks up # a parse error. # Usage: debug-sym-file.py [SYMFILE] import os import click import symbolic @click.command() @click.argument("symfile") @click.pass_context def sym_file_debug(ctx, symfile): """Prints information about a sym file including whether it parses correctly.""" # Print size stats = os.stat(symfile) click.echo(f"{symfile}") click.echo(f"size: {stats.st_size:,}") # Print first line with open(symfile, "r") as fp: firstline = fp.readline().strip() parts = firstline.split(" ") click.echo(f"first line: {parts}") # Parse with symbolic and create symcache try: click.echo("parsing with symbolic ...") archive = symbolic.Archive.open(symfile) click.echo("listing objects and making symcaches ...") for obj in archive.iter_objects(): click.echo(f"* {obj.debug_id}") obj.make_symcache() except Exception: click.echo("symbolic can't parse it") raise if __name__ == "__main__": sym_file_debug()
mpl-2.0
pimutils/todoman
tests/test_backend.py
1
3329
from datetime import date from datetime import datetime import icalendar import pytest import pytz from dateutil.tz import tzlocal from freezegun import freeze_time from todoman.model import Todo from todoman.model import VtodoWriter def test_datetime_serialization(todo_factory, tmpdir): now = datetime(2017, 8, 31, 23, 49, 53, tzinfo=pytz.UTC) todo = todo_factory(created_at=now) filename = tmpdir.join("default").join(todo.filename) with open(str(filename)) as f: assert "CREATED:20170831T234953Z\n" in f.readlines() def test_serialize_created_at(todo_factory): now = datetime.now(tz=pytz.UTC) todo = todo_factory(created_at=now) vtodo = VtodoWriter(todo).serialize() assert vtodo.get("created") is not None def test_serialize_dtstart(todo_factory): now = datetime.now(tz=pytz.UTC) todo = todo_factory(start=now) vtodo = VtodoWriter(todo).serialize() assert vtodo.get("dtstart") is not None def test_serializer_raises(todo_factory): todo = todo_factory() writter = VtodoWriter(todo) with pytest.raises(Exception): writter.serialize_field("nonexistant", 7) def test_supported_fields_are_serializeable(): supported_fields = set(Todo.ALL_SUPPORTED_FIELDS) serialized_fields = set(VtodoWriter.FIELD_MAP.keys()) assert supported_fields == serialized_fields def test_vtodo_serialization(todo_factory): """Test VTODO serialization: one field of each type.""" description = "A tea would be nice, thanks." todo = todo_factory( categories=["tea", "drinking", "hot"], description=description, due=datetime(3000, 3, 21), start=date(3000, 3, 21), priority=7, status="IN-PROCESS", summary="Some tea", rrule="FREQ=MONTHLY", ) writer = VtodoWriter(todo) vtodo = writer.serialize() assert [str(c) for c in vtodo.get("categories").cats] == ["tea", "drinking", "hot"] assert str(vtodo.get("description")) == description assert vtodo.get("priority") == 7 assert vtodo.decoded("due") == datetime(3000, 3, 21, tzinfo=tzlocal()) assert vtodo.decoded("dtstart") == date(3000, 3, 21) assert str(vtodo.get("status")) == "IN-PROCESS" assert vtodo.get("rrule") == icalendar.vRecur.from_ical("FREQ=MONTHLY") @freeze_time("2017-04-04 20:11:57") def test_update_last_modified(todo_factory, todos, tmpdir): todo = todo_factory() assert todo.last_modified == datetime.now(tzlocal()) def test_sequence_increment(default_database, todo_factory, todos): todo = todo_factory() assert todo.sequence == 1 default_database.save(todo) assert todo.sequence == 2 # Relaod (and check the caching flow for the sequence) todo = next(todos()) assert todo.sequence == 2 def test_normalize_datetime(): writter = VtodoWriter(None) assert writter.normalize_datetime(date(2017, 6, 17)) == date(2017, 6, 17) assert writter.normalize_datetime(datetime(2017, 6, 17)) == datetime( 2017, 6, 17, tzinfo=tzlocal() ) assert writter.normalize_datetime(datetime(2017, 6, 17, 12, 19)) == datetime( 2017, 6, 17, 12, 19, tzinfo=tzlocal() ) assert writter.normalize_datetime( datetime(2017, 6, 17, 12, tzinfo=tzlocal()) ) == datetime(2017, 6, 17, 12, tzinfo=tzlocal())
isc
pimutils/todoman
todoman/formatters.py
1
9862
import json from datetime import date from datetime import datetime from datetime import timedelta from time import mktime from typing import Iterable from typing import Optional from typing import Union import click import humanize import parsedatetime import pytz from dateutil.tz import tzlocal from todoman.model import Todo from todoman.model import TodoList def rgb_to_ansi(colour: Optional[str]) -> Optional[str]: """ Convert a string containing an RGB colour to ANSI escapes """ if not colour or not colour.startswith("#"): return None r, g, b = colour[1:3], colour[3:5], colour[5:7] if not len(r) == len(g) == len(b) == 2: return None return f"\33[38;2;{int(r, 16)!s};{int(g, 16)!s};{int(b, 16)!s}m" class DefaultFormatter: def __init__( self, date_format="%Y-%m-%d", time_format="%H:%M", dt_separator=" ", tz_override=None, ): self.date_format = date_format self.time_format = time_format self.dt_separator = dt_separator self.datetime_format = dt_separator.join( filter(bool, (date_format, time_format)) ) self.tz = tz_override or tzlocal() self.now = datetime.now().replace(tzinfo=self.tz) self._parsedatetime_calendar = parsedatetime.Calendar( version=parsedatetime.VERSION_CONTEXT_STYLE, ) def simple_action(self, action: str, todo: Todo) -> str: return f'{action} "{todo.summary}"' def compact(self, todo: Todo) -> str: return self.compact_multiple([todo]) def compact_multiple(self, todos: Iterable[Todo], hide_list=False) -> str: # TODO: format lines fuidly and drop the table # it can end up being more readable when too many columns are empty. # show dates that are in the future in yellow (in 24hs) or grey (future) table = [] for todo in todos: completed = "X" if todo.is_completed else " " percent = todo.percent_complete or "" if percent: percent = f" ({percent}%)" if todo.categories: categories = " [" + ", ".join(todo.categories) + "]" else: categories = "" priority = click.style( self.format_priority_compact(todo.priority), fg="magenta", ) due = self.format_datetime(todo.due) or "(no due date)" now = self.now if isinstance(todo.due, datetime) else self.now.date() due_colour = None if todo.due: if todo.due <= now and not todo.is_completed: due_colour = "red" elif todo.due >= now + timedelta(hours=24): due_colour = "white" elif todo.due >= now: due_colour = "yellow" else: due_colour = "white" if due_colour: due = click.style(str(due), fg=due_colour) recurring = "⟳" if todo.is_recurring else "" if hide_list: summary = "{} {}".format( todo.summary, percent, ) else: if not todo.list: raise ValueError("Cannot format todo without a list") summary = "{} {}{}".format( todo.summary, self.format_database(todo.list), percent, ) # TODO: add spaces on the left based on max todos" # FIXME: double space when no priority # split into parts to satisfy linter line too long table.append( f"[{completed}] {todo.id} {priority} {due} " f"{recurring}{summary}{categories}" ) return "\n".join(table) def _format_multiline(self, title: str, value: str) -> str: formatted_title = click.style(title, fg="white") if value.strip().count("\n") == 0: return f"\n\n{formatted_title}: {value}" else: return f"\n\n{formatted_title}:\n{value}" def detailed(self, todo: Todo) -> str: """Returns a detailed representation of a task. :param todo: The todo component. """ extra_lines = [] if todo.description: extra_lines.append(self._format_multiline("Description", todo.description)) if todo.location: extra_lines.append(self._format_multiline("Location", todo.location)) return f"{self.compact(todo)}{''.join(extra_lines)}" def format_datetime(self, dt: Optional[date]) -> Union[str, int, None]: if not dt: return "" elif isinstance(dt, datetime): return dt.strftime(self.datetime_format) elif isinstance(dt, date): return dt.strftime(self.date_format) def format_categories(self, categories): return ", ".join(categories) def parse_categories(self, categories): # existing code assumes categories is list, # but click passes tuple return list(categories) def parse_priority(self, priority: Optional[str]) -> Optional[int]: if priority is None or priority == "": return None if priority == "low": return 9 elif priority == "medium": return 5 elif priority == "high": return 4 elif priority == "none": return 0 else: raise ValueError("Priority has to be one of low, medium, high or none") def format_priority(self, priority: Optional[int]) -> str: if not priority: return "none" elif 1 <= priority <= 4: return "high" elif priority == 5: return "medium" elif 6 <= priority <= 9: return "low" raise ValueError("priority is an invalid value") def format_priority_compact(self, priority: Optional[int]) -> str: if not priority: return "" elif 1 <= priority <= 4: return "!!!" elif priority == 5: return "!!" elif 6 <= priority <= 9: return "!" raise ValueError("priority is an invalid value") def parse_datetime(self, dt: str) -> Optional[date]: if not dt: return None rv = self._parse_datetime_naive(dt) return rv.replace(tzinfo=self.tz) if isinstance(rv, datetime) else rv def _parse_datetime_naive(self, dt: str) -> date: """Parse dt and returns a naive datetime or a date""" try: return datetime.strptime(dt, self.datetime_format) except ValueError: pass try: return datetime.strptime(dt, self.date_format).date() except ValueError: pass try: return datetime.combine( self.now.date(), datetime.strptime(dt, self.time_format).time() ) except ValueError: pass rv, pd_ctx = self._parsedatetime_calendar.parse(dt) if not pd_ctx.hasDateOrTime: raise ValueError(f"Time description not recognized: {dt}") return datetime.fromtimestamp(mktime(rv)) def format_database(self, database: TodoList): return "{}@{}".format( rgb_to_ansi(database.colour) or "", click.style(database.name) ) class HumanizedFormatter(DefaultFormatter): def format_datetime(self, dt: Optional[date]) -> str: if not dt: return "" if isinstance(dt, datetime): rv = humanize.naturaltime(self.now - dt) if " from now" in rv: rv = f"in {rv[:-9]}" elif isinstance(dt, date): rv = humanize.naturaldate(dt) return rv class PorcelainFormatter(DefaultFormatter): def _todo_as_dict(self, todo): return { "completed": todo.is_completed, "due": self.format_datetime(todo.due), "id": todo.id, "list": todo.list.name, "percent": todo.percent_complete, "summary": todo.summary, "categories": todo.categories, "priority": todo.priority, "location": todo.location, "description": todo.description, "completed_at": self.format_datetime(todo.completed_at), } def compact(self, todo: Todo) -> str: return json.dumps(self._todo_as_dict(todo), indent=4, sort_keys=True) def compact_multiple(self, todos: Iterable[Todo], hide_list=False) -> str: data = [self._todo_as_dict(todo) for todo in todos] return json.dumps(data, indent=4, sort_keys=True) def simple_action(self, action: str, todo: Todo) -> str: return self.compact(todo) def parse_priority(self, priority: Optional[str]) -> Optional[int]: if priority is None: return None try: if int(priority) in range(0, 10): return int(priority) else: raise ValueError("Priority has to be in the range 0-9") except ValueError as e: raise click.BadParameter(str(e)) def detailed(self, todo: Todo) -> str: return self.compact(todo) def format_datetime(self, value: Optional[date]) -> Optional[int]: if value: if not isinstance(value, datetime): dt = datetime.fromordinal(value.toordinal()) else: dt = value return int(dt.timestamp()) else: return None def parse_datetime(self, value): if value: return datetime.fromtimestamp(value, tz=pytz.UTC) else: return None
isc
pimutils/todoman
todoman/model.py
1
35300
from __future__ import annotations import logging import os import socket import sqlite3 from datetime import date from datetime import datetime from datetime import time from datetime import timedelta from os.path import normpath from os.path import split from typing import Iterable from uuid import uuid4 import icalendar import pytz from atomicwrites import AtomicWriter from dateutil.rrule import rrulestr from dateutil.tz import tzlocal from todoman import exceptions logger = logging.getLogger(name=__name__) # Initialize this only once # We were doing this all over the place (even if unused!), so at least only do # it once. LOCAL_TIMEZONE = tzlocal() class cached_property: """A read-only @property that is only evaluated once. Only usable on class instances' methods. """ def __init__(self, fget, doc=None): self.__name__ = fget.__name__ self.__module__ = fget.__module__ self.__doc__ = doc or fget.__doc__ self.fget = fget def __get__(self, obj, cls): if obj is None: return self obj.__dict__[self.__name__] = result = self.fget(obj) return result class Todo: """ Represents a task/todo, and wrapps around icalendar.Todo. All text attributes are always treated as text, and "" will be returned if they are not defined. Date attributes are treated as datetime objects, and None will be returned if they are not defined. All datetime objects have tzinfo, either the one defined in the file, or the local system's one. """ categories: list[str] completed_at: datetime | None created_at: datetime | None due: date | None dtstamp: datetime | None last_modified: datetime | None related: list[Todo] rrule: str | None start: date | None def __init__( self, filename: str = None, mtime: int = None, new: bool = False, list: TodoList = None, ): """ Creates a new todo using `todo` as a source. :param str filename: The name of the file for this todo. Defaults to the <uid>.ics :param mtime int: The last modified time for the file backing this Todo. :param bool new: Indicate that a new Todo is being created and should be populated with default values. :param TodoList list: The list to which this Todo belongs. """ self.list = list now = datetime.now(LOCAL_TIMEZONE) self.uid = f"{uuid4().hex}@{socket.gethostname()}" if new: self.created_at = now else: self.created_at = None # Default values for supported fields self.categories = [] self.completed_at = None self.description = "" self.dtstamp = now self.due = None self.id = None self.last_modified = None self.location = "" self.percent_complete = 0 self.priority = 0 self.rrule = "" self.sequence = 0 self.start = None self.status = "NEEDS-ACTION" self.summary = "" self.filename = filename or f"{self.uid}.ics" self.related = [] if os.path.basename(self.filename) != self.filename: raise ValueError(f"Must not be an absolute path: {self.filename}") self.mtime = mtime or datetime.now() def clone(self) -> Todo: """ Returns a clone of this todo Returns a copy of this todo, which is almost identical, except that is has a different UUID and filename. """ todo = Todo(new=True, list=self.list) fields = ( Todo.STRING_FIELDS + Todo.INT_FIELDS + Todo.LIST_FIELDS + Todo.DATETIME_FIELDS ) fields.remove("uid") for field in fields: setattr(todo, field, getattr(self, field)) return todo STRING_FIELDS = [ "description", "location", "status", "summary", "uid", "rrule", ] INT_FIELDS = [ "percent_complete", "priority", "sequence", ] LIST_FIELDS = [ "categories", ] DATETIME_FIELDS = [ "completed_at", "created_at", "dtstamp", "start", "due", "last_modified", ] RRULE_FIELDS = [ "rrule", ] ALL_SUPPORTED_FIELDS = ( DATETIME_FIELDS + INT_FIELDS + LIST_FIELDS + RRULE_FIELDS + STRING_FIELDS ) VALID_STATUSES = ( "CANCELLED", "COMPLETED", "IN-PROCESS", "NEEDS-ACTION", ) def __setattr__(self, name: str, value): """Check type and avoid setting fields to None""" """when that is not a valid attribue.""" v = value if name in Todo.RRULE_FIELDS: if value is None: v = "" else: assert isinstance( value, str ), f"Got {type(value)} for {name} where str was expected" if name in Todo.STRING_FIELDS: if value is None: v = "" else: assert isinstance( value, str ), f"Got {type(value)} for {name} where str was expected" if name in Todo.INT_FIELDS: if value is None: v = 0 else: assert isinstance( value, int ), f"Got {type(value)} for {name} where int was expected" if name in Todo.LIST_FIELDS: if value is None: v = [] else: assert isinstance( value, list ), f"Got {type(value)} for {name} where list was expected" return object.__setattr__(self, name, v) @property def is_completed(self) -> bool: return bool(self.completed_at) or self.status in ("CANCELLED", "COMPLETED") @property def is_recurring(self) -> bool: return bool(self.rrule) def _apply_recurrence_to_dt(self, dt) -> datetime | None: if not dt: return None recurrence = rrulestr(self.rrule, dtstart=dt) if isinstance(dt, date) and not isinstance(dt, datetime): dt = datetime.combine(dt, time.min) return recurrence.after(dt) def _create_next_instance(self): copy = self.clone() copy.due = self._apply_recurrence_to_dt(self.due) copy.start = self._apply_recurrence_to_dt(self.start) assert copy.uid != self.uid # TODO: Push copy's alarms. return copy def complete(self) -> None: """ Immediately completes this todo Immediately marks this todo as completed, sets the percentage to 100% and the completed_at datetime to now. If this todo belongs to a series, newly created todo are added to the ``related`` list. """ if self.is_recurring: related = self._create_next_instance() if related: self.rrule = None self.related.append(related) self.completed_at = datetime.now(tz=LOCAL_TIMEZONE) self.percent_complete = 100 self.status = "COMPLETED" @cached_property def path(self) -> str: if not self.list: raise ValueError("A todo without a list does not have a path.") return os.path.join(self.list.path, self.filename) def cancel(self) -> None: self.status = "CANCELLED" class VtodoWriter: """Writes a Todo as a VTODO file.""" """Maps Todo field names to VTODO field names""" FIELD_MAP = { "summary": "summary", "priority": "priority", "sequence": "sequence", "uid": "uid", "categories": "categories", "completed_at": "completed", "description": "description", "dtstamp": "dtstamp", "start": "dtstart", "due": "due", "location": "location", "percent_complete": "percent-complete", "priority": "priority", "status": "status", "created_at": "created", "last_modified": "last-modified", "rrule": "rrule", } def __init__(self, todo: Todo): self.todo = todo def normalize_datetime(self, dt: date) -> date: """ Eliminate several differences between dates, times and datetimes which are hindering comparison: - Convert everything to datetime - Add missing timezones - Cast to UTC Datetimes are cast to UTC because icalendar doesn't include the VTIMEZONE information upon serialization, and some clients have issues dealing with that. """ if isinstance(dt, date) and not isinstance(dt, datetime): return dt if not dt.tzinfo: dt = dt.replace(tzinfo=LOCAL_TIMEZONE) return dt.astimezone(pytz.UTC) def serialize_field(self, name: str, value): if name in Todo.RRULE_FIELDS: return icalendar.vRecur.from_ical(value) if name in Todo.DATETIME_FIELDS: return self.normalize_datetime(value) if name in Todo.LIST_FIELDS: return value if name in Todo.INT_FIELDS: return int(value) if name in Todo.STRING_FIELDS: return value raise Exception(f"Unknown field {name} serialized.") def set_field(self, name: str, value): # If serialized value is None: self.vtodo.pop(name) if value: logger.debug("Setting field %s to %s.", name, value) self.vtodo.add(name, value) def serialize(self, original=None): """Serialize a Todo into a VTODO.""" if not original: original = icalendar.Todo() self.vtodo = original for source, target in self.FIELD_MAP.items(): self.vtodo.pop(target) if getattr(self.todo, source): self.set_field( target, self.serialize_field(source, getattr(self.todo, source)), ) return self.vtodo def _read(self, path): with open(path, "rb") as f: cal = f.read() cal = icalendar.Calendar.from_ical(cal) for component in cal.walk("VTODO"): return component def write(self): if os.path.exists(self.todo.path): self._write_existing(self.todo.path) else: self._write_new(self.todo.path) return self.vtodo def _write_existing(self, path): original = self._read(path) vtodo = self.serialize(original) with open(path, "rb") as f: cal = icalendar.Calendar.from_ical(f.read()) for index, component in enumerate(cal.subcomponents): if component.get("uid", None) == self.todo.uid: cal.subcomponents[index] = vtodo with AtomicWriter(path, "wb", overwrite=True).open() as f: f.write(cal.to_ical()) def _write_new(self, path): vtodo = self.serialize() c = icalendar.Calendar() c.add_component(vtodo) with AtomicWriter(path, "wb").open() as f: c.add("prodid", "io.barrera.todoman") c.add("version", "2.0") f.write(c.to_ical()) return vtodo class Cache: """ Caches Todos for faster read and simpler querying interface The Cache class persists relevant[1] fields into an SQL database, which is only updated if the actual file has been modified. This greatly increases load times, but, more importantly, provides a simpler interface for filtering/querying/sorting. [1]: Relevant fields are those we show when listing todos, or those which may be used for filtering/sorting. """ SCHEMA_VERSION = 9 def __init__(self, path: str): self.cache_path = str(path) os.makedirs(os.path.dirname(self.cache_path), exist_ok=True) self._conn = sqlite3.connect(self.cache_path) self._conn.row_factory = sqlite3.Row self._conn.execute("PRAGMA foreign_keys = ON") self.create_tables() def save_to_disk(self) -> None: self._conn.commit() def is_latest_version(self): """Checks if the cache DB schema is the latest version.""" try: return self._conn.execute( "SELECT version FROM meta WHERE version = ?", (Cache.SCHEMA_VERSION,), ).fetchone() except sqlite3.OperationalError: return False def drop_tables(self): self._conn.executescript( """ DROP TABLE IF EXISTS todos; DROP TABLE IF EXISTS lists; DROP TABLE IF EXISTS files; DROP TABLE IF EXISTS categories; """ ) def create_tables(self): if self.is_latest_version(): return self.drop_tables() self._conn.execute('CREATE TABLE IF NOT EXISTS meta ("version" INT)') self._conn.execute( "INSERT INTO meta (version) VALUES (?)", (Cache.SCHEMA_VERSION,), ) self._conn.execute( """ CREATE TABLE IF NOT EXISTS lists ( "name" TEXT PRIMARY KEY, "path" TEXT, "colour" TEXT, "mtime" INTEGER, CONSTRAINT path_unique UNIQUE (path) ); """ ) self._conn.execute( """ CREATE TABLE IF NOT EXISTS files ( "path" TEXT PRIMARY KEY, "list_name" TEXT, "mtime" INTEGER, CONSTRAINT path_unique UNIQUE (path), FOREIGN KEY(list_name) REFERENCES lists(name) ON DELETE CASCADE ); """ ) self._conn.execute( """ CREATE TABLE IF NOT EXISTS categories ( "id" INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT, "todos_id" INTEGER NOT NULL, "category" TEXT, CONSTRAINT category_unique UNIQUE (todos_id,category), FOREIGN KEY(todos_id) REFERENCES todos(id) ON DELETE CASCADE ); """ ) self._conn.execute( """ CREATE TABLE IF NOT EXISTS todos ( "file_path" TEXT, "id" INTEGER PRIMARY KEY, "uid" TEXT, "summary" TEXT, "due" INTEGER, "due_dt" INTEGER, "start" INTEGER, "start_dt" INTEGER, "priority" INTEGER, "created_at" INTEGER, "completed_at" INTEGER, "percent_complete" INTEGER, "dtstamp" INTEGER, "status" TEXT, "description" TEXT, "location" TEXT, "sequence" INTEGER, "last_modified" INTEGER, "rrule" TEXT, FOREIGN KEY(file_path) REFERENCES files(path) ON DELETE CASCADE ); """ ) def clear(self): self._conn.close() os.remove(self.cache_path) self._conn = None def add_list(self, name: str, path: str, colour: str, mtime: int): """ Inserts a new list into the cache. Returns the id of the newly inserted list. """ result = self._conn.execute( "SELECT name FROM lists WHERE path = ?", (path,), ).fetchone() if result: return result["name"] try: self._conn.execute( """ INSERT INTO lists ( name, path, colour, mtime ) VALUES (?, ?, ?, ?) """, ( name, path, colour, mtime, ), ) except sqlite3.IntegrityError as e: raise exceptions.AlreadyExists("list", name) from e return self.add_list(name, path, colour, mtime) def add_file(self, list_name: str, path: str, mtime: int): try: self._conn.execute( """ INSERT INTO files ( list_name, path, mtime ) VALUES (?, ?, ?); """, ( list_name, path, mtime, ), ) except sqlite3.IntegrityError as e: raise exceptions.AlreadyExists("file", list_name) from e def add_category(self, todos_id, category): try: self._conn.execute( """ INSERT INTO categories ( todos_id, category ) VALUES (?, ?); """, (todos_id, category), ) except sqlite3.IntegrityError as e: raise exceptions.AlreadyExists("category", category) from e def _serialize_datetime( self, todo: icalendar.Todo, field: str, ) -> tuple[int | None, bool | None]: """ Serialize a todo field in two value, the first one is the corresponding timestamp, the second one is a boolean indicating if the serialized value is a date or a datetime. :param icalendar.Todo todo: An icalendar component object :param str field: The name of the field to serialize """ dt = todo.decoded(field, None) if not dt: return None, None is_date = isinstance(dt, date) and not isinstance(dt, datetime) if is_date: dt = datetime(dt.year, dt.month, dt.day) if not dt.tzinfo: dt = dt.replace(tzinfo=LOCAL_TIMEZONE) return dt.timestamp(), is_date def _serialize_rrule(self, todo, field) -> str | None: rrule = todo.get(field) if not rrule: return None return rrule.to_ical().decode() def add_vtodo(self, todo: icalendar.Todo, file_path: str, id=None) -> int: """ Adds a todo into the cache. :param icalendar.Todo todo: The icalendar component object on which """ sql = """ INSERT INTO todos ( {} file_path, uid, summary, due, due_dt, start, start_dt, priority, created_at, completed_at, percent_complete, dtstamp, status, description, location, sequence, last_modified, rrule ) VALUES ({}?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?) """ due, due_dt = self._serialize_datetime(todo, "due") start, start_dt = self._serialize_datetime(todo, "dtstart") if start and due: start = None if start >= due else start params = [ file_path, todo.get("uid"), todo.get("summary"), due, due_dt, start, start_dt, todo.get("priority", 0) or None, self._serialize_datetime(todo, "created")[0], self._serialize_datetime(todo, "completed")[0], todo.get("percent-complete", None), self._serialize_datetime(todo, "dtstamp")[0], todo.get("status", "NEEDS-ACTION"), todo.get("description", None), todo.get("location", None), todo.get("sequence", 1), self._serialize_datetime(todo, "last-modified")[0], self._serialize_rrule(todo, "rrule"), ] if id: params = [id] + params sql = sql.format("id,\n", "?, ") else: sql = sql.format("", "") cursor = self._conn.cursor() try: cursor.execute(sql, params) rv = cursor.lastrowid assert rv is not None finally: cursor.close() if todo.get("categories"): for category in todo.get("categories").cats: self.add_category(rv, category) return rv def todos( self, lists=(), categories=None, priority=None, location="", grep="", sort=(), reverse=True, due=None, start=None, startable=False, status="NEEDS-ACTION,IN-PROCESS", ) -> Iterable[Todo]: """ Returns filtered cached todos, in a specified order. If no order is specified, todos are sorted by the following fields:: completed_at -priority due -created_at :param list lists: Only return todos for these lists. :param str location: Only return todos with a location containing this string. :param str categories: Only return todos with a category containing this string. :param str grep: Filter common fields with this substring. :param list sort: Order returned todos by these fields. Field names with a ``-`` prepended will be used to sort in reverse order. :param bool reverse: Reverse the order of the todos after sorting. :param int due: Return only todos due within ``due`` hours. :param str priority: Only return todos with priority at least as high as specified. :param tuple(bool, datetime) start: Return only todos before/after ``start`` date :param list(str) status: Return only todos with any of the given statuses. :return: A sorted, filtered list of todos. :rtype: generator """ extra_where = [] params: list = [] if "ANY" not in status: statuses = status.split(",") extra_where.append( "AND (status IN ({}) OR status IS NULL)".format( ", ".join(["?"] * len(statuses)) ) ) params.extend(s.upper() for s in statuses) if lists: lists = [ list_.name if isinstance(list_, TodoList) else list_ for list_ in lists ] q = ", ".join(["?"] * len(lists)) extra_where.append(f"AND files.list_name IN ({q})") params.extend(lists) if categories: category_slots = ", ".join(["?"] * len(categories)) extra_where.append( "AND upper(categories.category) IN ({category_slots})".format( category_slots=category_slots ) ) params = params + [category.upper() for category in categories] if priority: extra_where.append("AND PRIORITY > 0 AND PRIORITY <= ?") params.append(f"{priority}") if location: extra_where.append("AND location LIKE ?") params.append(f"%{location}%") if grep: # # requires sqlite with pcre, which won't be available everywhere: # extra_where.append('AND summary REGEXP ?') # params.append(grep) extra_where.append("AND summary LIKE ?") params.append(f"%{grep}%") if due: max_due = (datetime.now() + timedelta(hours=due)).timestamp() extra_where.append("AND due IS NOT NULL AND due < ?") params.append(max_due) if start: is_before, dt = start dt = dt.timestamp() if is_before: extra_where.append("AND start <= ?") params.append(dt) else: extra_where.append("AND start >= ?") params.append(dt) if startable: extra_where.append("AND (start IS NULL OR start <= ?)") params.append(datetime.now().timestamp()) if sort: order_items = [] for s in sort: if s.startswith("-"): order_items.append(f" {s[1:]} ASC") else: order_items.append(f" {s} DESC") order = ",".join(order_items) else: order = """ completed_at DESC, priority IS NOT NULL, priority DESC, due IS NOT NULL, due DESC, created_at ASC """ if not reverse: # Note the change in case to avoid swapping all of them. sqlite # doesn't care about casing anyway. order = order.replace(" DESC", " asc").replace(" ASC", " desc") query = """ SELECT DISTINCT todos.*, files.list_name, files.path, group_concat(category) AS categories FROM todos, files LEFT JOIN categories ON categories.todos_id = todos.id WHERE todos.file_path = files.path {} GROUP BY uid ORDER BY {} """.format( " ".join(extra_where), order, ) logger.debug(query) logger.debug(params) result = self._conn.execute(query, params) seen_paths = set() warned_paths = set() for row in result: todo = self._todo_from_db(row) path = row["path"] if path in seen_paths and path not in warned_paths: logger.warning( "Todo is in read-only mode because there are multiple todos in %s", path, ) warned_paths.add(path) seen_paths.add(path) yield todo def _datetime_from_db(self, dt) -> datetime | None: if dt: return datetime.fromtimestamp(dt, LOCAL_TIMEZONE) return None def _date_from_db(self, dt, is_date=False) -> date | None: """Deserialise a date (possible datetime).""" if not dt: return dt if is_date: return datetime.fromtimestamp(dt, LOCAL_TIMEZONE).date() else: return datetime.fromtimestamp(dt, LOCAL_TIMEZONE) def _categories_from_db(self, categories): if categories: return categories.split(",") return [] def _todo_from_db(self, row: dict) -> Todo: todo = Todo() todo.id = row["id"] todo.uid = row["uid"] todo.summary = row["summary"] todo.due = self._date_from_db(row["due"], row["due_dt"]) todo.start = self._date_from_db(row["start"], row["start_dt"]) todo.categories = self._categories_from_db(row["categories"]) todo.priority = row["priority"] todo.created_at = self._datetime_from_db(row["created_at"]) todo.completed_at = self._datetime_from_db(row["completed_at"]) todo.dtstamp = self._datetime_from_db(row["dtstamp"]) todo.percent_complete = row["percent_complete"] todo.status = row["status"] todo.description = row["description"] todo.location = row["location"] logger.debug("todo.categories: %s\n", todo.categories) todo.sequence = row["sequence"] todo.last_modified = row["last_modified"] todo.list = self.lists_map[row["list_name"]] todo.filename = os.path.basename(row["path"]) todo.rrule = row["rrule"] return todo def lists(self) -> Iterable[TodoList]: result = self._conn.execute("SELECT * FROM lists") for row in result: yield TodoList( name=row["name"], path=row["path"], colour=row["colour"], ) @cached_property def lists_map(self) -> dict[str, TodoList]: return {list_.name: list_ for list_ in self.lists()} def expire_lists(self, paths: dict[str, int]) -> None: results = self._conn.execute("SELECT path, name, mtime from lists") for result in results: if result["path"] not in paths: self.delete_list(result["name"]) else: mtime = paths.get(result["path"]) if mtime and mtime > result["mtime"]: self.delete_list(result["name"]) def delete_list(self, name: str) -> None: self._conn.execute("DELETE FROM lists WHERE lists.name = ?", (name,)) def todo(self, id: int, read_only=False) -> Todo: # XXX: DON'T USE READ_ONLY result = self._conn.execute( """ SELECT todos.*, files.list_name, files.path, group_concat(category) AS categories FROM todos, files LEFT JOIN categories ON categories.todos_id = todos.id WHERE files.path = todos.file_path AND todos.id = ? GROUP BY uid """, (id,), ).fetchone() if not result: raise exceptions.NoSuchTodo(id) if not read_only: count = self._conn.execute( """ SELECT count(id) AS c FROM files, todos WHERE todos.file_path = files.path AND path=? """, (result["path"],), ).fetchone() if count["c"] > 1: raise exceptions.ReadOnlyTodo(result["path"]) return self._todo_from_db(result) def expire_files(self, paths_to_mtime: dict[str, int]) -> None: """Remove stale cache entries based on the given fresh data.""" result = self._conn.execute("SELECT path, mtime FROM files") for row in result: path, mtime = row["path"], row["mtime"] if paths_to_mtime.get(path, None) != mtime: self.expire_file(path) def expire_file(self, path: str) -> None: self._conn.execute("DELETE FROM files WHERE path = ?", (path,)) class TodoList: def __init__(self, name: str, path: str, colour: str = None): self.path = path self.name = name self.colour = colour @staticmethod def colour_for_path(path: str) -> str | None: try: with open(os.path.join(path, "color")) as f: return f.read().strip() except OSError: logger.debug("No colour for list %s", path) @staticmethod def name_for_path(path: str) -> str: try: with open(os.path.join(path, "displayname")) as f: return f.read().strip() except OSError: return split(normpath(path))[1] @staticmethod def mtime_for_path(path: str) -> int: colour_file = os.path.join(path, "color") display_file = os.path.join(path, "displayname") mtimes = [] if os.path.exists(colour_file): mtimes.append(_getmtime(colour_file)) if os.path.exists(display_file): mtimes.append(_getmtime(display_file)) if mtimes: return max(mtimes) else: return 0 def __eq__(self, other) -> bool: if isinstance(other, TodoList): return self.name == other.name return object.__eq__(self, other) def __str__(self) -> str: return self.name class Database: """ This class is essentially a wrapper around all the lists (which in turn, contain all the todos). Caching in abstracted inside this class, and is transparent to outside classes. """ def __init__(self, paths, cache_path): self.cache = Cache(cache_path) self.paths = [str(path) for path in paths] self.update_cache() def update_cache(self) -> None: paths = {path: TodoList.mtime_for_path(path) for path in self.paths} self.cache.expire_lists(paths) paths_to_mtime = {} paths_to_list_name = {} for path in self.paths: list_name = self.cache.add_list( TodoList.name_for_path(path), path, TodoList.colour_for_path(path), paths[path], ) for entry in os.listdir(path): if not entry.endswith(".ics"): continue entry_path = os.path.join(path, entry) mtime = _getmtime(entry_path) paths_to_mtime[entry_path] = mtime paths_to_list_name[entry_path] = list_name self.cache.expire_files(paths_to_mtime) for entry_path, mtime in paths_to_mtime.items(): list_name = paths_to_list_name[entry_path] try: self.cache.add_file(list_name, entry_path, mtime) except exceptions.AlreadyExists: logger.debug("File already in cache: %s", entry_path) continue try: with open(entry_path, "rb") as f: cal = icalendar.Calendar.from_ical(f.read()) for component in cal.walk("VTODO"): self.cache.add_vtodo(component, entry_path) except Exception: logger.exception("Failed to read entry %s.", entry_path) self.cache.save_to_disk() def todos(self, **kwargs) -> Iterable[Todo]: return self.cache.todos(**kwargs) def todo(self, id: int, **kwargs) -> Todo: return self.cache.todo(id, **kwargs) def lists(self) -> Iterable[TodoList]: return self.cache.lists() def move(self, todo: Todo, new_list: TodoList, from_list: TodoList) -> None: orig_path = os.path.join(from_list.path, todo.filename) dest_path = os.path.join(new_list.path, todo.filename) os.rename(orig_path, dest_path) def delete(self, todo: Todo) -> None: if not todo.list: raise ValueError("Cannot delete Todo without a list.") path = os.path.join(todo.list.path, todo.filename) os.remove(path) def flush(self) -> Iterable[Todo]: for todo in self.todos(status=["ANY"]): if todo.is_completed: yield todo self.delete(todo) self.cache.clear() self.cache = None def save(self, todo: Todo) -> None: if not todo.list: raise ValueError("Cannot save Todo without a list.") for related in todo.related: self.save(related) todo.sequence += 1 todo.last_modified = datetime.now(LOCAL_TIMEZONE) vtodo = VtodoWriter(todo).write() self.cache.expire_file(todo.path) mtime = _getmtime(todo.path) self.cache.add_file(todo.list.name, todo.path, mtime) todo.id = self.cache.add_vtodo(vtodo, todo.path, todo.id) self.cache.save_to_disk() def _getmtime(path: str) -> int: return os.stat(path).st_mtime_ns
isc
pimutils/todoman
tests/test_config.py
2
4629
from unittest.mock import patch import pytest from click.testing import CliRunner from todoman.cli import cli from todoman.configuration import ConfigurationException from todoman.configuration import load_config def test_explicit_nonexistant(runner): result = CliRunner().invoke( cli, env={"TODOMAN_CONFIG": "/nonexistant"}, catch_exceptions=True, ) assert result.exception assert "Configuration file /nonexistant does not exist" in result.output def test_xdg_nonexistant(runner): with patch("xdg.BaseDirectory.xdg_config_dirs", ["/does-not-exist"]): result = CliRunner().invoke( cli, catch_exceptions=True, ) assert result.exception assert "No configuration file found" in result.output def test_xdg_existant(runner, tmpdir, config): with tmpdir.mkdir("todoman").join("config.py").open("w") as f: with config.open() as c: f.write(c.read()) with patch("xdg.BaseDirectory.xdg_config_dirs", [str(tmpdir)]): result = CliRunner().invoke( cli, catch_exceptions=True, ) assert not result.exception assert not result.output.strip() def test_sane_config(config, runner, tmpdir): config.write( 'color = "auto"\n' 'date_format = "%Y-%m-%d"\n' f'path = "{tmpdir}"\n' f'cache_path = "{tmpdir.join("cache.sqlite")}"\n' ) result = runner.invoke(cli) # This is handy for debugging breakage: if result.exception: print(result.output) raise result.exception assert not result.exception def test_invalid_color(config, runner): config.write('color = 12\npath = "/"\n') result = runner.invoke(cli, ["list"]) assert result.exception assert ( "Error: Bad color setting. Invalid type (expected str, got int)." in result.output ) def test_invalid_color_arg(config, runner): config.write('path = "/"\n') result = runner.invoke(cli, ["--color", "12", "list"]) assert result.exception assert "Usage:" in result.output def test_missing_path(config, runner): config.write('color = "auto"\n') result = runner.invoke(cli, ["list"]) assert result.exception assert "Error: Missing 'path' setting." in result.output @pytest.mark.xfail(reason="Not implemented") def test_extra_entry(config, runner): config.write("color = auto\ndate_format = %Y-%m-%d\npath = /\nblah = false\n") result = runner.invoke(cli, ["list"]) assert result.exception assert "Error: Invalid configuration entry" in result.output @pytest.mark.xfail(reason="Not implemented") def test_extra_section(config, runner): config.write("date_format = %Y-%m-%d\npath = /\n[extra]\ncolor = auto\n") result = runner.invoke(cli, ["list"]) assert result.exception assert "Invalid configuration section" in result.output def test_missing_cache_dir(config, runner, tmpdir): cache_dir = tmpdir.join("does").join("not").join("exist") cache_file = cache_dir.join("cache.sqlite") config.write(f'path = "{tmpdir}/*"\ncache_path = "{cache_file}"\n') result = runner.invoke(cli) assert not result.exception assert cache_dir.isdir() assert cache_file.isfile() def test_date_field_in_time_format(config, runner, tmpdir): config.write('path = "/"\ntime_format = "%Y-%m-%d"\n') result = runner.invoke(cli) assert result.exception assert ( "Found date component in `time_format`, please use `date_format` for that." in result.output ) def test_date_field_in_time(config, runner, tmpdir): config.write('path = "/"\ndate_format = "%Y-%d-:%M"\n') result = runner.invoke(cli) assert result.exception assert ( "Found time component in `date_format`, please use `time_format` for that." in result.output ) def test_colour_validation_auto(config): with patch( "todoman.configuration.find_config", return_value=(str(config)), ): cfg = load_config() assert cfg["color"] == "auto" def test_colour_validation_always(config): config.write("color = 'always'\n", "a") with patch( "todoman.configuration.find_config", return_value=(str(config)), ): cfg = load_config() assert cfg["color"] == "always" def test_colour_validation_invalid(config): config.write("color = 'on_weekends_only'\n", "a") with patch( "todoman.configuration.find_config", return_value=(str(config)), ), pytest.raises(ConfigurationException): load_config()
isc
mozilla-services/autopush
autopush/logging.py
1
9732
"""Custom Logging Setup """ import io import json import Queue import pkg_resources import socket import sys import time import threading from typing import Any # noqa import boto3 import raven from raven.transport.twisted import TwistedHTTPTransport from raven.utils.stacks import iter_stack_frames from twisted.internet import reactor from twisted.logger import ( formatEvent, formatEventAsClassicLogText, globalLogBeginner, globalLogPublisher, LogLevel, ILogObserver ) from zope.interface import implementer from autopush.utils import get_ec2_instance_id # A complete set of keys we don't include in Fields from a log event IGNORED_KEYS = frozenset([ "factory", "failure", "format", "isError", "log_failure", "log_format", "log_flattened", "log_level", "log_legacy", "log_logger", "log_source", "log_system", "log_text", "log_time", "log_trace", "message", "message_type", "severity", "task_level", "time", "timestamp", "type", "why", ]) # whether the global LogBeginner.beginLoggingTo has been called: it # should only be called once began_logging = False # an ec2 instance id or falling back to the hostname instance_id_or_hostname = None def begin_or_register(observer, redirectStandardIO=False, **kwargs): # type: (Any, bool, **Any) -> None """Register observer with the global LogPublisher Registers via the global LogBeginner the first time called. """ global began_logging if not began_logging: globalLogBeginner.beginLoggingTo( [observer], redirectStandardIO=redirectStandardIO, **kwargs ) began_logging = True else: globalLogPublisher.addObserver(observer) @implementer(ILogObserver) class PushLogger(object): """Twisted LogObserver implementation Supports firehose delivery, Raven exception reporting, and json/test console debugging output. """ def __init__(self, logger_name, log_level="debug", log_format="json", log_output="stdout", sentry_dsn=None, firehose_delivery_stream=None): self.logger_name = "-".join([ logger_name, pkg_resources.get_distribution("autopush").version ]) self._filename = None self.log_level = LogLevel.lookupByName(log_level) if log_output == "stdout": self._output = sys.stdout elif log_output == "none": self._output = None else: self._filename = log_output self._output = None if log_format == "json": self.format_event = self.json_format else: self.format_event = formatEventAsClassicLogText if sentry_dsn: self.raven_client = raven.Client( release=raven.fetch_package_version("autopush"), transport=TwistedHTTPTransport, enable_breadcrumbs=False, ) else: self.raven_client = None if firehose_delivery_stream: self.firehose = FirehoseProcessor( stream_name=firehose_delivery_stream) else: self.firehose = None def __call__(self, event): if self.raven_client and 'log_failure' in event: self.raven_log(event) if event["log_level"] < self.log_level: return text = self.format_event(event) if self.firehose: self.firehose.process(text) if self._output: self._output.write(unicode(text)) self._output.flush() def raven_log(self, event): f = event["log_failure"] stack = None extra = dict() tb = f.getTracebackObject() if not tb: # include the current stack for at least some # context. sentry's expecting that "Frames should be # sorted from oldest to newest." stack = list(iter_stack_frames())[:-5] # approx. extra = dict(no_failure_tb=True) extra.update( log_format=event.get('log_format'), log_namespace=event.get('log_namespace'), client_info=event.get('client_info'), ) reactor.callFromThread( self.raven_client.captureException, exc_info=(f.type, f.value, tb), stack=stack, extra=extra, ) # just in case del tb def json_format(self, event): error = bool(event.get("isError")) or "log_failure" in event ts = event["log_time"] if error: severity = 3 else: severity = 5 def to_fields(kv): reply = dict() for k, v in kv: if (k not in IGNORED_KEYS and type(v) in (str, unicode, list, int, float, bool)): reply[k] = v return reply msg = { "Hostname": instance_id_or_hostname, "Timestamp": ts * 1000 * 1000 * 1000, "Type": "twisted:log", "Severity": event.get("severity") or severity, "EnvVersion": "2.0", "Fields": to_fields(event.iteritems()), "Logger": self.logger_name, } # flatten the client_info into Fields ci = event.get('client_info') if ci and isinstance(ci, dict): msg['Fields'].update( to_fields(ci.iteritems())) # flatten timings into Fields ti = event.get('timings') if ti and isinstance(ti, dict): msg["Fields"].update( to_fields(ti.iteritems()) ) # Add the nicely formatted message msg["Fields"]["message"] = formatEvent(event) return json.dumps(msg, skipkeys=True) + "\n" def start(self): if self._filename: self._output = io.open(self._filename, "a", encoding="utf-8") if self.firehose: self.firehose.start() begin_or_register(self) def stop(self): globalLogPublisher.removeObserver(self) if self._filename: self._output.close() self._output = None if self.firehose: self.firehose.stop() @classmethod def setup_logging(cls, logger_name, log_level="info", log_format="json", log_output="stdout", sentry_dsn=None, firehose_delivery_stream=None, no_aws=False): global instance_id_or_hostname if not instance_id_or_hostname: instance_id = None if no_aws else get_ec2_instance_id() instance_id_or_hostname = instance_id or socket.getfqdn() pl = cls(logger_name, log_level=log_level, log_format=log_format, log_output=log_output, sentry_dsn=sentry_dsn, firehose_delivery_stream=firehose_delivery_stream) pl.start() reactor.addSystemEventTrigger('before', 'shutdown', pl.stop) return pl class FirehoseProcessor(object): """Batches log events for sending to AWS FireHose""" RECORD_SEPARATOR = u"\x1e" MAX_RECORD_SIZE = 1024 * 1024 MAX_REQUEST_SIZE = 4 * 1024 * 1024 MAX_RECORD_BATCH = 500 MAX_INTERVAL = 30 def __init__(self, stream_name, maxsize=0): self._records = Queue.Queue(maxsize=maxsize) self._prepped = [] self._total_size = 0 self._thread = None self._client = boto3.client("firehose") self._run = False self._stream_name = stream_name def start(self): self._thread = threading.Thread(target=self._worker) self._thread.start() def stop(self): self._records.put_nowait(None) self._thread.join() self._thread = None def process(self, record): try: self._records.put_nowait(record) except Queue.Full: # Drop extra records pass def _worker(self): self._last_send = time.time() while True: time_since_sent = time.time() - self._last_send remaining_wait = self.MAX_INTERVAL - time_since_sent try: record = self._records.get(timeout=remaining_wait) except Queue.Empty: # Send the records self._send_record_batch() continue if record is None: # Stop signal so we exit break # Is this record going to put us over our request size? rec_size = len(record) + 1 if self._total_size + rec_size >= self.MAX_REQUEST_SIZE: self._send_record_batch() # Store this record self._prepped.append(record) self._total_size += rec_size if len(self._prepped) >= self.MAX_RECORD_BATCH: self._send_record_batch() # We're done running, send any remaining self._send_record_batch() def _send_record_batch(self): self._last_send = time.time() if not self._prepped: return # Attempt to send the record batch twice, or give up tries = 0 while tries < 3: response = self._client.put_record_batch( DeliveryStreamName=self._stream_name, Records=[{"Data": bytes(self.RECORD_SEPARATOR + record)} for record in self._prepped] ) if response["FailedPutCount"] > 0: tries += 1 else: break self._prepped = [] self._total_size = 0
mpl-2.0
mozilla-services/autopush
autopush/jwt.py
1
4878
import base64 import binascii import json import os from cryptography.exceptions import InvalidSignature from cryptography.hazmat.primitives.asymmetric import ec, utils from cryptography.hazmat.primitives import hashes from pyasn1.error import PyAsn1Error from twisted.logger import Logger from typing import Tuple # noqa from autopush.types import JSONDict # noqa # temporarily toggleable for easily enabling on production _JWT_MEMORY_PRESSURE = os.environ.get('_JWT_MEMORY_PRESSURE', 0) if _JWT_MEMORY_PRESSURE != 0: # pragma: nocover try: from __pypy__ import add_memory_pressure except ImportError: _JWT_MEMORY_PRESSURE = 0 else: try: _JWT_MEMORY_PRESSURE = int(_JWT_MEMORY_PRESSURE) except ValueError: _JWT_MEMORY_PRESSURE = 2496 def repad(string): # type: (str) -> str """Adds padding to strings for base64 decoding""" if len(string) % 4: string += '===='[len(string) % 4:] return string class VerifyJWT(object): """Minimally verify a Vapid JWT object. Why hand roll? Most python JWT libraries either use a python elliptic curve library directly, or call one that does, or is abandoned, or a dozen other reasons. After spending half a day looking for reasonable replacements, I decided to just write the functions we need directly. THIS IS NOT A FULL JWT REPLACEMENT. """ @staticmethod def extract_signature(auth): # type: (str) -> Tuple[str, str] """Fix the JWT auth token. The JWA spec defines the signature to be a pair of 32octet encoded longs. The `ecdsa` library signs using a raw, 32octet pair of values (s, r). Cryptography, which uses OpenSSL, uses a DER sequence of (s, r). This function converts the raw ecdsa to DER. :param auth: A JWT authorization token. :type auth: str :return tuple containing the signature material and signature """ payload, asig = auth.encode('utf8').rsplit(".", 1) sig = base64.urlsafe_b64decode(repad(asig)) if len(sig) != 64: return payload, sig encoded = utils.encode_dss_signature( s=int(binascii.hexlify(sig[32:]), 16), r=int(binascii.hexlify(sig[:32]), 16) ) return payload, encoded @staticmethod def extract_assertion(token): # type (str) -> JSONDict """Extract the assertion dictionary from the passed token. This does NOT do validation. :param token: Partial or full VAPID auth token :type token: str :return dict of the VAPID claims """ return json.loads( base64.urlsafe_b64decode( repad(token.split('.')[1]).encode('utf8'))) @staticmethod def validate_and_extract_assertion(token, key): # type (str, str) -> JSONDict """Decode a web token into a assertion dictionary. This attempts to rectify both ecdsa and openssl generated signatures. We use the built-in cryptography library since it wraps libssl and is faster than the python only approach. :param token: VAPID auth token :type token: str :param key: bitarray containing public key :type key: str or bitarray :return dict of the VAPID claims :raise InvalidSignature """ # convert the signature if needed. try: sig_material, signature = VerifyJWT.extract_signature(token) pkey = ec.EllipticCurvePublicKey.from_encoded_point( ec.SECP256R1(), key ) # cffi issue #320: public_key & verify allocate approx. if _JWT_MEMORY_PRESSURE: # pragma: nocover add_memory_pressure(_JWT_MEMORY_PRESSURE) # NOTE: verify() will take any string as the signature. It appears # to be doing lazy verification and matching strings rather than # comparing content values. If the signatures start failing for # some unknown reason in the future, decode the signature and # make sure it matches how we're reconstructing it. # This will raise an InvalidSignature exception if failure. # It will be captured externally. pkey.verify( signature, sig_material.encode('utf8'), ec.ECDSA(hashes.SHA256())) return VerifyJWT.extract_assertion(sig_material) except InvalidSignature: raise except (ValueError, TypeError, binascii.Error, PyAsn1Error): raise InvalidSignature() except Exception: # pragma: no cover Logger().failure("Unexpected error processing JWT") raise InvalidSignature()
mpl-2.0
mozilla-services/autopush
autopush/http.py
1
7885
"""HTTP Server Protocol Factories on top of cyclone""" from typing import ( # noqa Any, Callable, Dict, Optional, Sequence, Tuple, Type ) import cyclone.web from twisted.internet import reactor from twisted.web.client import ( _HTTP11ClientFactory, Agent, HTTPConnectionPool, ) from autopush.base import BaseHandler from autopush.config import AutopushConfig # noqa from autopush.db import DatabaseManager from autopush.router import routers_from_config from autopush.router.interface import IRouter # noqa from autopush.ssl import AutopushSSLContextFactory # noqa from autopush.web.health import ( HealthHandler, MemUsageHandler, StatusHandler ) from autopush.web.base import NotFoundHandler from autopush.web.limitedhttpconnection import LimitedHTTPConnection from autopush.web.log_check import LogCheckHandler from autopush.web.message import MessageHandler from autopush.web.registration import ( ChannelRegistrationHandler, NewRegistrationHandler, SubRegistrationHandler, UaidRegistrationHandler, ) from autopush.web.webpush import WebPushHandler from autopush.websocket import ( NotificationHandler, RouterHandler, ) from autopush.websocket import PushServerProtocol # noqa from autopush.web.dockerflow import VersionHandler, LBHeartbeatHandler APHandlers = Sequence[Tuple[str, Type[BaseHandler]]] CycloneLogger = Callable[[BaseHandler], None] def skip_request_logging(handler): # type: (cyclone.web.RequestHandler) -> None """Skip cyclone's request logging""" class BaseHTTPFactory(cyclone.web.Application): ap_handlers = None # type: APHandlers health_ap_handlers = ( (r"^/status", StatusHandler), (r"^/health", HealthHandler), # DockerFlow checks (r"^/__version__", VersionHandler), (r"^/__heartbeat__", StatusHandler), (r"^/__lbheartbeat__", LBHeartbeatHandler), ) def __init__(self, conf, # type: AutopushConfig db, # type: DatabaseManager handlers=None, # type: APHandlers log_function=skip_request_logging, # type: CycloneLogger **kwargs): # type: (...) -> None self.conf = conf self.db = db self.noisy = conf.debug cyclone.web.Application.__init__( self, handlers=self.ap_handlers if handlers is None else handlers, default_host=self._hostname, debug=conf.debug, log_function=log_function, **kwargs ) def add_health_handlers(self): """Add the health check HTTP handlers""" self.add_handlers(".*$", self.health_ap_handlers) @property def _hostname(self): return self.conf.hostname @classmethod def for_handler(cls, handler_cls, # Type[BaseHTTPFactory] conf, # type: AutopushConfig db=None, # type: Optional[DatabaseManager] **kwargs): # type: (...) -> BaseHTTPFactory """Create a cyclone app around a specific handler_cls for tests. Creates an uninitialized (no setup() called) DatabaseManager from conf if one isn't specified. handler_cls must be included in ap_handlers or a ValueError is thrown. """ if 'handlers' in kwargs: # pragma: nocover raise ValueError("handler_cls incompatibile with handlers kwarg") for pattern, handler in cls.ap_handlers + cls.health_ap_handlers: if handler is handler_cls: if db is None: db = DatabaseManager.from_config(conf) return cls._for_handler( conf, db=db, handlers=[(pattern, handler)], **kwargs ) raise ValueError("{!r} not in ap_handlers".format( handler_cls)) # pragma: nocover @classmethod def _for_handler(cls, conf, **kwargs): # type: (AutopushConfig, **Any) -> BaseHTTPFactory """Create an instance w/ default kwargs for for_handler""" raise NotImplementedError # pragma: nocover class EndpointHTTPFactory(BaseHTTPFactory): ap_handlers = ( (r"/wpush/(?:(?P<api_ver>v\d+)\/)?(?P<token>[^\/]+)", WebPushHandler), (r"/m/(?P<message_id>[^\/]+)", MessageHandler), (r"/v1/(?P<type>[^\/]+)/(?P<app_id>[^\/]+)/registration", NewRegistrationHandler), (r"/v1/(?P<type>[^\/]+)/(?P<app_id>[^\/]+)/registration/" r"(?P<uaid>[^\/]+)", UaidRegistrationHandler), (r"/v1/(?P<type>[^\/]+)/(?P<app_id>[^\/]+)/registration/" r"(?P<uaid>[^\/]+)/subscription", SubRegistrationHandler), (r"/v1/(?P<type>[^\/]+)/(?P<app_id>[^\/]+)/registration/" r"(?P<uaid>[^\/]+)/subscription/(?P<chid>[^\/]+)", ChannelRegistrationHandler), (r"/v1/err(?:/(?P<err_type>[^\/]+))?", LogCheckHandler), (r".*", NotFoundHandler), ) protocol = LimitedHTTPConnection def __init__(self, conf, # type: AutopushConfig db, # type: DatabaseManager routers, # type: Dict[str, IRouter] **kwargs): # type: (...) -> None self.ap_handlers = tuple(self.ap_handlers) BaseHTTPFactory.__init__(self, conf, db=db, **kwargs) self.routers = routers def ssl_cf(self): # type: () -> Optional[AutopushSSLContextFactory] """Build our SSL Factory (if configured). Configured from the ssl_key/cert/dh_param and client_cert values. """ conf = self.conf return conf.ssl.cf(require_peer_certs=conf.enable_tls_auth) @classmethod def _for_handler(cls, conf, db, routers=None, **kwargs): if routers is None: routers = routers_from_config( conf, db=db, agent=agent_from_config(conf) ) return cls(conf, db=db, routers=routers, **kwargs) class InternalRouterHTTPFactory(BaseHTTPFactory): ap_handlers = ( (r"/push/([^\/]+)", RouterHandler), (r"/notif/([^\/]+)(?:/(\d+))?", NotificationHandler), ) def __init__(self, conf, # type: AutopushConfig db, # type: DatabaseManager clients, # type: Dict[str, PushServerProtocol] **kwargs): # type: (...) -> None BaseHTTPFactory.__init__(self, conf, db, **kwargs) self.clients = clients @property def _hostname(self): return self.conf.router_hostname def ssl_cf(self): # type: () -> Optional[AutopushSSLContextFactory] """Build our SSL Factory (if configured). Configured from the router_ssl_key/cert and ssl_dh_param values. """ return self.conf.router_ssl.cf() @classmethod def _for_handler(cls, conf, db, clients=None, **kwargs): if clients is None: clients = {} return cls(conf, db=db, clients=clients, **kwargs) class MemUsageHTTPFactory(BaseHTTPFactory): ap_handlers = ( (r"^/_memusage", MemUsageHandler), ) class QuietClientFactory(_HTTP11ClientFactory): """Silence the start/stop factory messages.""" noisy = False def agent_from_config(conf): # type: (AutopushConfig) -> Agent """Create a twisted.web.client Agent from the given config""" # Use a persistent connection pool for HTTP requests. pool = HTTPConnectionPool(reactor) if not conf.debug: pool._factory = QuietClientFactory return Agent(reactor, connectTimeout=conf.connect_timeout, pool=pool)
mpl-2.0
mozilla-services/autopush
autopush/ssl.py
1
5050
"""Custom SSL configuration""" from __future__ import absolute_import import socket # noqa import ssl from typing import ( # noqa Any, Dict, FrozenSet, Optional, Tuple, ) from OpenSSL import SSL from twisted.internet.ssl import DefaultOpenSSLContextFactory try: SSL_PROTO = ssl.PROTOCOL_TLS except AttributeError: # pragma: nocover SSL_PROTO = ssl.PROTOCOL_SSLv23 MOZILLA_INTERMEDIATE_CIPHERS = ( 'ECDHE-RSA-AES128-GCM-SHA256:' 'ECDHE-ECDSA-AES128-GCM-SHA256:' 'ECDHE-RSA-AES256-GCM-SHA384:' 'ECDHE-ECDSA-AES256-GCM-SHA384:' 'DHE-RSA-AES128-GCM-SHA256:' 'DHE-DSS-AES128-GCM-SHA256:' 'ECDHE-RSA-AES128-SHA256:' 'ECDHE-ECDSA-AES128-SHA256:' 'ECDHE-RSA-AES128-SHA:' 'ECDHE-ECDSA-AES128-SHA:' 'ECDHE-RSA-AES256-SHA384:' 'ECDHE-ECDSA-AES256-SHA384:' 'ECDHE-RSA-AES256-SHA:' 'ECDHE-ECDSA-AES256-SHA:' 'DHE-RSA-AES128-SHA256:' 'DHE-RSA-AES128-SHA:' 'DHE-DSS-AES128-SHA256:' 'DHE-RSA-AES256-SHA256:' 'DHE-DSS-AES256-SHA:' 'DHE-RSA-AES256-SHA:' 'AES128-GCM-SHA256:' 'AES256-GCM-SHA384:' 'AES128-SHA256:' 'AES256-SHA256:' 'AES128-SHA:' 'AES256-SHA:' 'AES:' 'CAMELLIA:DES-CBC3-SHA:' '!aNULL:!eNULL:!EXPORT:!DES:!RC4:!MD5:!PSK:!aECDH:' '!EDH-DSS-DES-CBC3-SHA:!EDH-RSA-DES-CBC3-SHA:!KRB5-DES-CBC3-SHA' ) class AutopushSSLContextFactory(DefaultOpenSSLContextFactory): """A SSL context factory""" def __init__(self, *args, **kwargs): self.dh_file = kwargs.pop('dh_file', None) self.require_peer_certs = kwargs.pop('require_peer_certs', False) DefaultOpenSSLContextFactory.__init__(self, *args, **kwargs) def cacheContext(self): """Setup the main context factory with custom SSL settings""" if self._context is None: ctx = self._contextFactory(self.sslmethod) ctx.set_cipher_list(MOZILLA_INTERMEDIATE_CIPHERS) ctx.set_options(SSL.OP_CIPHER_SERVER_PREFERENCE) ctx.set_options(SSL.OP_NO_SSLv2) ctx.set_options(SSL.OP_NO_SSLv3) ctx.set_options(SSL.OP_NO_COMPRESSION) ctx.set_mode(SSL.MODE_RELEASE_BUFFERS) ctx.set_options(SSL.OP_ALL & ~SSL.OP_MICROSOFT_BIG_SSLV3_BUFFER) ctx.use_certificate_chain_file(self.certificateFileName) ctx.use_privatekey_file(self.privateKeyFileName) if self.dh_file: ctx.load_tmp_dh(self.dh_file) if self.require_peer_certs: # Require peer certs but only for use by # RequestHandlers ctx.set_verify( SSL.VERIFY_PEER | SSL.VERIFY_CLIENT_ONCE, self._allow_peer) self._context = ctx def _allow_peer(self, conn, cert, errno, depth, preverify_ok): # skip verification: we only care about whitelisted signatures # on file return True def monkey_patch_ssl_wrap_socket(): """Replace ssl.wrap_socket with ssl_wrap_socket_cached""" ssl.wrap_socket = ssl_wrap_socket_cached def undo_monkey_patch_ssl_wrap_socket(): """Undo monkey_patch_ssl_wrap_socket""" ssl.wrap_socket = _orig_ssl_wrap_socket _CacheKey = FrozenSet[Tuple[str, Any]] _sslcontext_cache = {} # type: Dict[_CacheKey, ssl.SSLContext] _orig_ssl_wrap_socket = ssl.wrap_socket def ssl_wrap_socket_cached( sock, # type: socket.socket keyfile=None, # type: Optional[str] certfile=None, # type: Optional[str] server_side=False, # type: bool cert_reqs=ssl.CERT_NONE, # type: int ssl_version=SSL_PROTO, # type: int ca_certs=None, # type: Optional[str] do_handshake_on_connect=True, # type: bool suppress_ragged_eofs=True, # type: bool ciphers=None # type: Optional[str] ): # type: (...) -> ssl.SSLSocket """ssl.wrap_socket replacement that caches SSLContexts""" key_kwargs = ( ('keyfile', keyfile), ('certfile', certfile), ('cert_reqs', cert_reqs), ('ssl_version', ssl_version), ('ca_certs', ca_certs), ('ciphers', ciphers), ) key = frozenset(key_kwargs) context = _sslcontext_cache.get(key) if context is not None: return context.wrap_socket( sock, server_side=server_side, do_handshake_on_connect=do_handshake_on_connect, suppress_ragged_eofs=suppress_ragged_eofs ) wrapped = _orig_ssl_wrap_socket( sock, keyfile=keyfile, certfile=certfile, server_side=server_side, cert_reqs=cert_reqs, ssl_version=ssl_version, ca_certs=ca_certs, do_handshake_on_connect=do_handshake_on_connect, suppress_ragged_eofs=suppress_ragged_eofs, ciphers=ciphers ) _sslcontext_cache[key] = wrapped.context return wrapped
mpl-2.0
mozilla-services/autopush
autopush/router/apns2.py
1
7384
import json from collections import deque from decimal import Decimal import hyper.tls from hyper import HTTP20Connection from hyper.http20.exceptions import HTTP20Error from autopush.exceptions import RouterException SANDBOX = 'api.development.push.apple.com' SERVER = 'api.push.apple.com' APNS_MAX_CONNECTIONS = 20 # These values are defined by APNs as header values that should be sent. # The hyper library requires that all header values be strings. # These values should be considered "opaque" to APNs. # see https://developer.apple.com/search/?q=%22apns-priority%22 APNS_PRIORITY_IMMEDIATE = '10' APNS_PRIORITY_LOW = '5' class ComplexEncoder(json.JSONEncoder): def default(self, obj): if isinstance(obj, Decimal): return int(obj.to_integral_value()) # for most data types, this function isn't called. # the following is added for safety, but should not # be required. return json.JSONEncoder.default(self, obj) # pragma nocover class APNSException(Exception): pass class APNSClient(object): def __init__(self, cert_file, key_file, topic, alt=False, use_sandbox=False, max_connections=APNS_MAX_CONNECTIONS, logger=None, metrics=None, load_connections=True, max_retry=2): """Create the APNS client connector. The cert_file and key_file can be derived from the exported `.p12` **Apple Push Services: *bundleID* ** key contained in the **Keychain Access** application. To extract the proper PEM formatted data, you can use the following commands: ``` openssl pkcs12 -in file.p12 -out apns_cert.pem -clcerts -nokeys openssl pkcs12 -in file.p12 -out apns_key.pem -nocerts -nodes ``` The *topic* is the Bundle ID of the bridge recipient iOS application. Since the cert needs to be tied directly to an application, the topic is usually similar to "com.example.MyApplication". :param cert_file: Path to the PEM formatted APNs certification file. :type cert_file: str :param key_file: Path to the PEM formatted APNs key file. :type key_file: str :param topic: The *Bundle ID* that identifies the assoc. iOS app. :type topic: str :param alt: Use the alternate APNs publication port (if 443 is blocked) :type alt: bool :param use_sandbox: Use the development sandbox :type use_sandbox: bool :param max_connections: Max number of pooled connections to use :type max_connections: int :param logger: Status logger :type logger: logger :param metrics: Metric recorder :type metrics: autopush.metrics.IMetric :param load_connections: used for testing :type load_connections: bool :param max_retry: Number of HTTP2 transmit attempts :type max_retry: int """ self.server = SANDBOX if use_sandbox else SERVER self.port = 2197 if alt else 443 self.log = logger self.metrics = metrics self.topic = topic self._max_connections = max_connections self._max_retry = max_retry self.connections = deque(maxlen=max_connections) if load_connections: self.ssl_context = hyper.tls.init_context(cert=(cert_file, key_file)) self.connections.extendleft((HTTP20Connection( self.server, self.port, ssl_context=self.ssl_context, force_proto='h2') for x in range(0, max_connections))) if self.log: self.log.debug("Starting APNS connection") def send(self, router_token, payload, apns_id, priority=True, topic=None, exp=None): """Send the dict of values to the remote bridge This sends the raw data to the remote bridge application using the APNS2 HTTP2 API. :param router_token: APNs provided hex token identifying recipient :type router_token: str :param payload: Data to send to recipient :type payload: dict :param priority: True is high priority, false is low priority :type priority: bool :param topic: BundleID for the recipient application (overides default) :type topic: str :param exp: Message expiration timestamp :type exp: timestamp """ body = json.dumps(payload, cls=ComplexEncoder) priority = APNS_PRIORITY_IMMEDIATE if priority else APNS_PRIORITY_LOW # NOTE: Hyper requires that all header values be strings. 'Priority' # is a integer string, which may be "simplified" and cause an error. # The added str() function safeguards against that. headers = { 'apns-id': apns_id, 'apns-priority': str(priority), 'apns-topic': topic or self.topic, } if exp: headers['apns-expiration'] = str(exp) url = '/3/device/' + router_token attempt = 0 while True: try: connection = self._get_connection() # request auto-opens closed connections, so if a connection # has timed out or failed for other reasons, it's automatically # re-established. stream_id = connection.request( 'POST', url=url, body=body, headers=headers) # get_response() may return an AttributeError. Not really sure # how it happens, but the connected socket may get set to None. # We'll treat that as a premature socket closure. response = connection.get_response(stream_id) if response.status != 200: reason = json.loads( response.read().decode('utf-8'))['reason'] raise RouterException( "APNS Transmit Error {}:{}".format(response.status, reason), status_code=response.status, response_body="APNS could not process " "your message {}".format(reason), log_exception=True, reason=reason ) break except (HTTP20Error, IOError): connection.close() attempt += 1 if attempt < self._max_retry: continue raise finally: # Returning a closed connection to the pool is ok. # hyper will reconnect on .request() self._return_connection(connection) def _get_connection(self): try: connection = self.connections.pop() return connection except IndexError: raise RouterException( "Too many APNS requests, increase pool from {}".format( self._max_connections ), status_code=503, response_body="APNS busy, please retry") def _return_connection(self, connection): self.connections.appendleft(connection)
mpl-2.0
mozilla-services/autopush
autopush/config.py
1
16793
"""Autopush Config Object and Setup""" import json import socket from argparse import Namespace # noqa from hashlib import sha256 from typing import ( # noqa Any, Dict, List, Optional, Type, Union ) from attr import ( attrs, attrib, Factory ) from cryptography.fernet import Fernet, MultiFernet from cryptography.hazmat.primitives import constant_time import autopush.db as db from autopush.exceptions import ( InvalidConfig, InvalidTokenException, VapidAuthException ) from autopush.ssl import AutopushSSLContextFactory from autopush.types import JSONDict # noqa from autopush.utils import ( CLIENT_SHA256_RE, canonical_url, get_amid, resolve_ip, repad, base64url_decode, parse_auth_header, ) from autopush.crypto_key import CryptoKey, CryptoKeyException def _init_crypto_key(ck): # type: (Optional[Union[str, List[str]]]) -> List[str] """Provide a default or ensure the provided's a list""" # if CRYPTO_KEY is not set by docker, it may pass an empty string, # which is converted into an Array element and prevents the config # file value from being read if ck is None or ck == ['']: return [Fernet.generate_key()] return ck if isinstance(ck, list) else [ck] def _nested(cls, **kwargs): # type: (Type, **Any) -> Any """Defines an attr cls nested within another attr. This attribute constructs the nested attr from a dict argument (representing its kwargs) unless already an instance of cls. """ def converter(arg): return arg if isinstance(arg, cls) else cls(**arg) return attrib(converter=converter, **kwargs) @attrs class SSLConfig(object): """AutopushSSLContextFactory configuration""" key = attrib(default=None) # type: Optional[str] cert = attrib(default=None) # type: Optional[str] dh_param = attrib(default=None) # type: Optional[str] def cf(self, **kwargs): # type: (**Any) -> Optional[AutopushSSLContextFactory] """Build our AutopushSSLContextFactory (if configured)""" if not self.key: return None return AutopushSSLContextFactory( self.key, self.cert, dh_file=self.dh_param, **kwargs ) @attrs class DDBTableConfig(object): """A DynamoDB Table's configuration""" tablename = attrib() # type: str read_throughput = attrib(default=5) # type: int write_throughput = attrib(default=5) # type: int @attrs class AutopushConfig(object): """Main Autopush Settings Object""" debug = attrib(default=False) # type: bool fernet = attrib(init=False) # type: MultiFernet _crypto_key = attrib( converter=_init_crypto_key, default=None) # type: List[str] bear_hash_key = attrib(default=Factory(list)) # type: List[str] human_logs = attrib(default=True) # type: bool hostname = attrib(default=None) # type: Optional[str] port = attrib(default=None) # type: Optional[int] _resolve_hostname = attrib(default=False) # type: bool router_scheme = attrib(default=None) # type: Optional[str] router_hostname = attrib(default=None) # type: Optional[str] router_port = attrib(default=None) # type: Optional[int] endpoint_scheme = attrib(default=None) # type: Optional[str] endpoint_hostname = attrib(default=None) # type: Optional[str] endpoint_port = attrib(default=None) # type: Optional[int] proxy_protocol_port = attrib(default=None) # type: Optional[int] memusage_port = attrib(default=None) # type: Optional[int] statsd_host = attrib(default="localhost") # type: str statsd_port = attrib(default=8125) # type: int megaphone_api_url = attrib(default=None) # type: Optional[str] megaphone_api_token = attrib(default=None) # type: Optional[str] megaphone_poll_interval = attrib(default=30) # type: int datadog_api_key = attrib(default=None) # type: Optional[str] datadog_app_key = attrib(default=None) # type: Optional[str] datadog_flush_interval = attrib(default=None) # type: Optional[int] router_table = _nested( DDBTableConfig, default=dict(tablename="router") ) # type: DDBTableConfig message_table = _nested( DDBTableConfig, default=dict(tablename="message") ) # type: DDBTableConfig preflight_uaid = attrib( default="deadbeef00000000deadbeef00000000") # type: str ssl = _nested(SSLConfig, default=Factory(SSLConfig)) # type: SSLConfig router_ssl = _nested( SSLConfig, default=Factory(SSLConfig)) # type: SSLConfig client_certs = attrib(default=None) # type: Optional[Dict[str, str]] router_url = attrib(init=False) # type: str endpoint_url = attrib(init=False) # type: str ws_url = attrib(init=False) # type: str router_conf = attrib(default=Factory(dict)) # type: JSONDict # twisted Agent's connectTimeout connect_timeout = attrib(default=0.5) # type: float max_data = attrib(default=4096) # type: int env = attrib(default='development') # type: str ami_id = attrib(default=None) # type: Optional[str] cors = attrib(default=False) # type: bool hello_timeout = attrib(default=0) # type: int # Force timeout in idle seconds msg_limit = attrib(default=100) # type: int auto_ping_interval = attrib(default=None) # type: Optional[int] auto_ping_timeout = attrib(default=None) # type: Optional[int] max_connections = attrib(default=None) # type: Optional[int] close_handshake_timeout = attrib(default=None) # type: Optional[int] # Generate messages per legacy rules, only used for testing to # generate legacy data. _notification_legacy = attrib(default=False) # type: bool # Use the cryptography library use_cryptography = attrib(default=False) # type: bool # Strict-Transport-Security max age (Default 1 year in secs) sts_max_age = attrib(default=31536000) # type: int # Don't cache ssl.wrap_socket's SSLContexts no_sslcontext_cache = attrib(default=False) # type: bool # DynamoDB endpoint override aws_ddb_endpoint = attrib(default=None) # type: str allow_table_rotation = attrib(default=True) # type: bool def __attrs_post_init__(self): """Initialize the Settings object""" # Setup hosts/ports/urls if not self.hostname: self.hostname = socket.gethostname() if self._resolve_hostname: self.hostname = resolve_ip(self.hostname) if not self.endpoint_hostname: self.endpoint_hostname = self.hostname if not self.router_hostname: self.router_hostname = self.hostname self.router_url = canonical_url( self.router_scheme or 'http', self.router_hostname, self.router_port ) self.endpoint_url = canonical_url( self.endpoint_scheme or 'http', self.endpoint_hostname, self.endpoint_port ) # not accurate under autoendpoint (like router_url) self.ws_url = "{}://{}:{}/".format( 'wss' if self.ssl.key else 'ws', self.hostname, self.port ) self.fernet = MultiFernet([Fernet(key) for key in self._crypto_key]) @property def enable_tls_auth(self): """Whether TLS authentication w/ client certs is enabled""" return self.client_certs is not None @classmethod def from_argparse(cls, ns, **kwargs): # type: (Namespace, **Any) -> AutopushConfig """Create an instance from argparse/additional kwargs""" router_conf = {} if ns.key_hash: db.key_hash = ns.key_hash if ns.apns_creds: # if you have the critical elements for each external # router, create it try: router_conf["apns"] = json.loads(ns.apns_creds) except (ValueError, TypeError): raise InvalidConfig( "Invalid JSON specified for APNS config options") if ns.senderid_list: # Create a common gcmclient try: sender_ids = json.loads(ns.senderid_list) except (ValueError, TypeError): raise InvalidConfig("Invalid JSON specified for senderid_list") try: # This is an init check to verify that things are # configured correctly. Otherwise errors may creep in # later that go unaccounted. sender_ids[sender_ids.keys()[0]] except (IndexError, TypeError): raise InvalidConfig("No GCM SenderIDs specified or found.") router_conf["gcm"] = {"ttl": ns.gcm_ttl, "dryrun": ns.gcm_dryrun, "max_data": ns.max_data, "collapsekey": ns.gcm_collapsekey, "senderIDs": sender_ids, "endpoint": ns.gcm_endpoint} client_certs = None # endpoint only if getattr(ns, 'client_certs', None): try: client_certs_arg = json.loads(ns.client_certs) except (ValueError, TypeError): raise InvalidConfig("Invalid JSON specified for client_certs") if client_certs_arg: if not ns.ssl_key: raise InvalidConfig("client_certs specified without SSL " "enabled (no ssl_key specified)") client_certs = {} for name, sigs in client_certs_arg.iteritems(): if not isinstance(sigs, list): raise InvalidConfig( "Invalid JSON specified for client_certs") for sig in sigs: sig = sig.upper() if (not name or not CLIENT_SHA256_RE.match(sig) or sig in client_certs): raise InvalidConfig( "Invalid client_certs argument") client_certs[sig] = name if ns.fcm_creds: try: router_conf["fcm"] = { "version": ns.fcm_version, "ttl": ns.fcm_ttl, "dryrun": ns.fcm_dryrun, "max_data": ns.max_data, "collapsekey": ns.fcm_collapsekey, "creds": json.loads(ns.fcm_creds) } if not router_conf["fcm"]["creds"]: raise InvalidConfig( "Empty credentials for FCM config options" ) for creds in router_conf["fcm"]["creds"].values(): if "auth" not in creds: raise InvalidConfig( "Missing auth for FCM config options" ) except (ValueError, TypeError): raise InvalidConfig( "Invalid JSON specified for FCM config options" ) if ns.adm_creds: # Create a common admclient try: router_conf["adm"] = json.loads(ns.adm_creds) except (ValueError, TypeError): raise InvalidConfig( "Invalid JSON specified for ADM config options") ami_id = None # Not a fan of double negatives, but this makes more # understandable args if not ns.no_aws: ami_id = get_amid() or "Unknown" allow_table_rotation = not ns.no_table_rotation return cls( crypto_key=ns.crypto_key, datadog_api_key=ns.datadog_api_key, datadog_app_key=ns.datadog_app_key, datadog_flush_interval=ns.datadog_flush_interval, hostname=ns.hostname, statsd_host=ns.statsd_host, statsd_port=ns.statsd_port, router_conf=router_conf, resolve_hostname=ns.resolve_hostname, ami_id=ami_id, client_certs=client_certs, msg_limit=ns.msg_limit, connect_timeout=ns.connection_timeout, memusage_port=ns.memusage_port, use_cryptography=ns.use_cryptography, no_sslcontext_cache=ns._no_sslcontext_cache, router_table=dict( tablename=ns.router_tablename, read_throughput=ns.router_read_throughput, write_throughput=ns.router_write_throughput ), message_table=dict( tablename=ns.message_tablename, read_throughput=ns.message_read_throughput, write_throughput=ns.message_write_throughput ), ssl=dict( key=ns.ssl_key, cert=ns.ssl_cert, dh_param=ns.ssl_dh_param ), sts_max_age=ns.sts_max_age, allow_table_rotation=allow_table_rotation, **kwargs ) def make_endpoint(self, uaid, chid, key=None): """Create an v1 or v2 WebPush endpoint from the identifiers. Both endpoints use bytes instead of hex to reduce ID length. v1 is the uaid + chid v2 is the uaid + chid + sha256(key).bytes :param uaid: User Agent Identifier :param chid: Channel or Subscription ID :param key: Optional Base64 URL-encoded application server key :returns: Push endpoint """ root = self.endpoint_url + '/wpush/' base = (uaid.replace('-', '').decode("hex") + chid.replace('-', '').decode("hex")) if key is None: return root + 'v1/' + self.fernet.encrypt(base).strip('=') raw_key = base64url_decode(key.encode('utf8')) ep = self.fernet.encrypt(base + sha256(raw_key).digest()).strip('=') return root + 'v2/' + ep def parse_endpoint(self, metrics, token, version="v1", ckey_header=None, auth_header=None): """Parse an endpoint into component elements of UAID, CHID and optional key hash if v2 :param token: The obscured subscription data. :param version: This is the API version of the token. :param ckey_header: the Crypto-Key header bearing the public key (from Crypto-Key: p256ecdsa=) :param auth_header: The Authorization header bearing the VAPID info :raises ValueError: In the case of a malformed endpoint. :returns: a dict containing (uaid=UAID, chid=CHID, public_key=KEY) """ token = self.fernet.decrypt(repad(token).encode('utf8')) public_key = None if ckey_header: try: crypto_key = CryptoKey(ckey_header) except CryptoKeyException: raise InvalidTokenException("Invalid key data") public_key = crypto_key.get_label('p256ecdsa') if auth_header: vapid_auth = parse_auth_header(auth_header) if not vapid_auth: raise VapidAuthException("Invalid Auth token") metrics.increment("notification.auth", tags="vapid:{version},scheme:{scheme}".format( **vapid_auth ).split(",")) # pull the public key from the VAPID auth header if needed try: if vapid_auth['version'] != 1: public_key = vapid_auth['k'] except KeyError: raise VapidAuthException("Missing Public Key") if version == 'v1' and len(token) != 32: raise InvalidTokenException("Corrupted push token") if version == 'v2': if not auth_header: raise VapidAuthException("Missing Authorization Header") if len(token) != 64: raise InvalidTokenException("Corrupted push token") if not public_key: raise VapidAuthException("Invalid key data") try: decoded_key = base64url_decode(public_key) except TypeError: raise VapidAuthException("Invalid key data") if not constant_time.bytes_eq(sha256(decoded_key).digest(), token[32:]): raise VapidAuthException("Key mismatch") return dict(uaid=token[:16].encode('hex'), chid=token[16:32].encode('hex'), version=version, public_key=public_key)
mpl-2.0
mozilla-services/autopush
autopush/tests/test_metrics.py
1
3703
import unittest import twisted.internet.base import pytest from mock import Mock, patch, call from autopush.metrics import ( IMetrics, DatadogMetrics, TwistedMetrics, SinkMetrics, periodic_reporter, ) class IMetricsTestCase(unittest.TestCase): def test_default(self): im = IMetrics() im.start() with pytest.raises(NotImplementedError): im.increment("test") with pytest.raises(NotImplementedError): im.gauge("test", 10) with pytest.raises(NotImplementedError): im.timing("test", 10) class SinkMetricsTestCase(unittest.TestCase): def test_passing(self): sm = SinkMetrics() sm.start() assert sm.increment("test") is None assert sm.gauge("test", 10) is None assert sm.timing("test", 10) is None class TwistedMetricsTestCase(unittest.TestCase): @patch("autopush.metrics.reactor") def test_basic(self, mock_reactor): twisted.internet.base.DelayedCall.debug = True m = TwistedMetrics('127.0.0.1') m.start() assert len(mock_reactor.mock_calls) > 0 m._metric = Mock() m.increment("test", 5) m._metric.increment.assert_called_with("test", 5) m.gauge("connection_count", 200) m._metric.gauge.assert_called_with("connection_count", 200) m.timing("lifespan", 113) m._metric.timing.assert_called_with("lifespan", 113) @patch("autopush.metrics.reactor") def test_tags(self, mock_reactor): twisted.internet.base.DelayedCall.debug = True m = TwistedMetrics('127.0.0.1') m.start() assert len(mock_reactor.mock_calls) > 0 m._metric = Mock() m.increment("test", 5, tags=["foo:bar"]) m._metric.increment.assert_called_with("test", 5, tags=["foo:bar"]) m.gauge("connection_count", 200, tags=["foo:bar", "baz:quux"]) m._metric.gauge.assert_called_with("connection_count", 200, tags=["foo:bar", "baz:quux"]) class DatadogMetricsTestCase(unittest.TestCase): @patch("autopush.metrics.datadog") def test_basic(self, mock_dog): hostname = "localhost" m = DatadogMetrics("someapikey", "someappkey", namespace="testpush", hostname="localhost") assert len(mock_dog.mock_calls) > 0 m._client = Mock() m.start() m._client.start.assert_called_with(flush_interval=10, roll_up_interval=10) m.increment("test", 5) m._client.increment.assert_called_with("testpush.test", 5, host=hostname) m.gauge("connection_count", 200) m._client.gauge.assert_called_with("testpush.connection_count", 200, host=hostname) m.timing("lifespan", 113) m._client.timing.assert_called_with("testpush.lifespan", value=113, host=hostname) class PeriodicReporterTestCase(unittest.TestCase): def test_periodic_reporter(self): metrics = Mock(spec=SinkMetrics) periodic_reporter(metrics) periodic_reporter(metrics, prefix='foo') metrics.gauge.assert_has_calls([ call('twisted.threadpool.idleWorkerCount', 0), call('twisted.threadpool.busyWorkerCount', 0), call('twisted.threadpool.backloggedWorkCount', 0), call('foo.twisted.threadpool.idleWorkerCount', 0), call('foo.twisted.threadpool.busyWorkerCount', 0), call('foo.twisted.threadpool.backloggedWorkCount', 0), ])
mpl-2.0
mozilla-services/autopush
autopush/tests/test_z_main.py
1
16788
"""Test main instantiation This is named test_z_main.py to run it last. Due to issues in this test, the testing environment is unclean and no further tests can be run reliably. """ import unittest import datetime import json from mock import Mock, patch import pytest from twisted.internet.defer import Deferred from twisted.trial import unittest as trialtest import hyper import oauth2client import hyper.tls import autopush.db from autopush.config import AutopushConfig from autopush.db import ( DatabaseManager, get_rotating_message_tablename, make_rotating_tablename, ) from autopush.exceptions import InvalidConfig from autopush.http import skip_request_logging from autopush.main import ( ConnectionApplication, EndpointApplication, ) from autopush.tests.support import test_db from autopush.utils import resolve_ip import autopush.tests connection_main = ConnectionApplication.main endpoint_main = EndpointApplication.main class ConfigTestCase(unittest.TestCase): def test_resolve_host(self): ip = resolve_ip("example.com") conf = AutopushConfig( hostname="example.com", resolve_hostname=True) assert conf.hostname == ip @patch("autopush.utils.socket") def test_resolve_host_no_interface(self, mock_socket): mock_socket.getaddrinfo.return_value = "" ip = resolve_ip("example.com") assert ip == "example.com" def test_new_month(self): today = datetime.date.today() next_month = today.month + 1 next_year = today.year if next_month > 12: # pragma: nocover next_month = 1 next_year += 1 tomorrow = datetime.datetime(year=next_year, month=next_month, day=1) db = test_db() db._tomorrow = Mock() db._tomorrow.return_value = tomorrow db.create_initial_message_tables() assert len(db.message_tables) == 3 class ConfigAsyncTestCase(trialtest.TestCase): def test_update_rotating_tables(self): from autopush.db import get_month conf = AutopushConfig( hostname="example.com", resolve_hostname=True) db = DatabaseManager.from_config( conf, resource=autopush.tests.boto_resource) db.create_initial_message_tables() # Erase the tables it has on init, and move current month back one last_month = get_month(-1) db.current_month = last_month.month db.message_tables = [make_rotating_tablename("message", delta=-1), make_rotating_tablename("message", delta=0)] # Create the next month's table, just in case today is the day before # a new month, in which case the lack of keys will cause an error in # update_rotating_tables next_month = get_month(1) assert next_month.month not in db.message_tables # Get the deferred back e = Deferred() d = db.update_rotating_tables() def check_tables(result): assert db.current_month == get_month().month assert len(db.message_tables) == 2 d.addCallback(check_tables) d.addBoth(lambda x: e.callback(True)) return e def test_update_rotating_tables_month_end(self): """Test that rotating adds next months table This test is intended to ensure that if the next day is a new month, then update_rotating_tables realizes this and add's the new table to the message_tables. A pre-requisite is that today cannot be the last day of the current month. Therefore, we first sub in _tomorrow to ensure it always appears as next month, and then remove the new table create_initial_tables made so we can observe update_rotating_tables add the new one. Note that sorting message table keys to find the last month does *not work* since the month digit is not zero-padded. """ today = datetime.date.today() next_month = today.month + 1 next_year = today.year if next_month > 12: # pragma: nocover next_month = 1 next_year += 1 tomorrow = datetime.datetime(year=next_year, month=next_month, day=1) conf = AutopushConfig( hostname="example.com", resolve_hostname=True) db = DatabaseManager.from_config( conf, resource=autopush.tests.boto_resource) db._tomorrow = Mock(return_value=tomorrow) db.create_initial_message_tables() # We should have 3 tables, one for next/this/last month assert len(db.message_tables) == 3 # Grab next month's table name and remove it next_month = get_rotating_message_tablename( conf.message_table.tablename, delta=1, boto_resource=db.resource ) db.message_tables.pop(db.message_tables.index(next_month)) # Get the deferred back d = db.update_rotating_tables() def check_tables(result): assert len(db.message_tables) == 3 assert next_month in db.message_tables d.addCallback(check_tables) return d def test_update_not_needed(self): conf = AutopushConfig( hostname="google.com", resolve_hostname=True) db = DatabaseManager.from_config( conf, resource=autopush.tests.boto_resource) db.create_initial_message_tables() # Erase the tables it has on init, and move current month back one db.message_tables = [] # Get the deferred back e = Deferred() d = db.update_rotating_tables() def check_tables(result): assert len(db.message_tables) == 1 d.addCallback(check_tables) d.addBoth(lambda x: e.callback(True)) return e def test_no_rotation(self): today = datetime.date.today() next_month = today.month + 1 next_year = today.year if next_month > 12: # pragma: nocover next_month = 1 next_year += 1 tomorrow = datetime.datetime(year=next_year, month=next_month, day=1) conf = AutopushConfig( hostname="example.com", resolve_hostname=True, allow_table_rotation=False ) resource = autopush.tests.boto_resource db = DatabaseManager.from_config( conf, resource=resource) db._tomorrow = Mock(return_value=tomorrow) db.create_initial_message_tables() assert len(db.message_tables) == 1 assert db.message_tables[0] == resource.get_latest_message_tablename( prefix=conf.message_table.tablename ) def check_tables(result): assert len(db.message_tables) == 1 assert db.message_tables[0] == \ resource.get_latest_message_tablename( prefix=conf.message_table.tablename ) dd = db.update_rotating_tables() dd.addCallback(check_tables) return dd class ConnectionMainTestCase(unittest.TestCase): def setUp(self): patchers = [ "autopush.main.TimerService.startService", "autopush.main.reactor", ] self.mocks = {} for name in patchers: patcher = patch(name) self.mocks[name] = patcher.start() def tearDown(self): for mock in self.mocks.values(): mock.stop() def test_basic(self): connection_main([], False, resource=autopush.tests.boto_resource) def test_ssl(self): connection_main([ "--ssl_dh_param=keys/dhparam.pem", "--ssl_cert=keys/server.crt", "--ssl_key=keys/server.key", "--router_ssl_cert=keys/server.crt", "--router_ssl_key=keys/server.key", ], False, resource=autopush.tests.boto_resource) def test_memusage(self): connection_main([ "--memusage_port=8083", ], False, resource=autopush.tests.boto_resource) def test_skip_logging(self): # Should skip setting up logging on the handler mock_handler = Mock() skip_request_logging(mock_handler) assert len(mock_handler.mock_calls) == 0 class EndpointMainTestCase(unittest.TestCase): class TestArg(AutopushConfig): # important stuff apns_creds = json.dumps({"firefox": {"cert": "cert.file", "key": "key.file"}}) gcm_endpoint = "gcm-http.googleapis.com/gcm/send" # less important stuff gcm_ttl = 999 gcm_dryrun = False gcm_collapsekey = "collapse" max_data = 4096 # filler crypto_key = 'AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA=' datadog_api_key = "datadog_api_key" datadog_app_key = "datadog_app_key" datadog_flush_interval = "datadog_flush_interval" hostname = "hostname" statsd_host = "statsd_host" statsd_port = 8125 router_tablename = "none" router_read_throughput = 0 router_write_throughput = 0 resolve_hostname = False message_tablename = "None" message_read_throughput = 0 message_write_throughput = 0 senderid_list = '{"12345":{"auth":"abcd"}}' key_hash = "supersikkret" no_aws = True fcm_enabled = True fcm_ttl = 999 fcm_dryrun = False fcm_collapsekey = "collapse" fcm_creds = json.dumps({"12345": {"auth": "abcd"}}) fcm_version = 0 ssl_key = "keys/server.crt" ssl_cert = "keys/server.key" ssl_dh_param = None msg_limit = 1000 _client_certs = dict(partner1=["1A:"*31 + "F9"], partner2=["2B:"*31 + "E8", "3C:"*31 + "D7"]) client_certs = json.dumps(_client_certs) connection_timeout = 1 proxy_protocol_port = None memusage_port = None disable_simplepush = True use_cryptography = False sts_max_age = 1234 _no_sslcontext_cache = False aws_ddb_endpoint = None no_table_rotation = False adm_creds = json.dumps({ "dev": { "app_id": "amzn1.application.StringOfStuff", "client_id": "amzn1.application-oa2-client.ev4nM0reStuff", "client_secret": "deadbeef0000decafbad1111" } }) def setUp(self): patchers = [ "autopush.db.preflight_check", "autopush.main.TimerService.startService", "autopush.main.reactor", ] self.mocks = {} for name in patchers: patcher = patch(name) self.mocks[name] = patcher.start() def tearDown(self): for mock in self.mocks.values(): mock.stop() autopush.db.key_hash = "" def test_basic(self): endpoint_main( [], False, resource=autopush.tests.boto_resource ) def test_ssl(self): endpoint_main([ "--ssl_dh_param=keys/dhparam.pem", "--ssl_cert=keys/server.crt", "--ssl_key=keys/server.key", ], False, resource=autopush.tests.boto_resource) def test_bad_senderidlist(self): returncode = endpoint_main([ "--gcm_endpoint='gcm-http.googleapis.com/gcm/send'", "--senderid_list='[Invalid'" ], False) assert returncode not in (None, 0) def test_bad_apnsconf(self): returncode = endpoint_main([ "--apns_creds='[Invalid'" ], False) assert returncode not in (None, 0) def test_client_certs(self): cert = self.TestArg._client_certs['partner1'][0] returncode = endpoint_main([ "--ssl_cert=keys/server.crt", "--ssl_key=keys/server.key", '--client_certs={"foo": ["%s"]}' % cert ], False, resource=autopush.tests.boto_resource) assert not returncode def test_proxy_protocol_port(self): endpoint_main([ "--proxy_protocol_port=8081", ], False, resource=autopush.tests.boto_resource) def test_memusage(self): endpoint_main([ "--memusage_port=8083", ], False, resource=autopush.tests.boto_resource) def test_client_certs_parse(self): conf = AutopushConfig.from_argparse(self.TestArg) assert conf.client_certs["1A:"*31 + "F9"] == 'partner1' assert conf.client_certs["2B:"*31 + "E8"] == 'partner2' assert conf.client_certs["3C:"*31 + "D7"] == 'partner2' def test_bad_client_certs(self): cert = self.TestArg._client_certs['partner1'][0] ssl_opts = ["--ssl_cert=keys/server.crt", "--ssl_key=keys/server.key"] assert endpoint_main( ssl_opts + ["--client_certs='[Invalid'"], False) == 1 assert endpoint_main( ssl_opts + ['--client_certs={"": ["%s"]}' % cert], False) == 1 assert endpoint_main( ssl_opts + ['--client_certs={"quux": [""]}'], False) == 1 assert endpoint_main( ssl_opts + ['--client_certs={"foo": "%s"}' % cert], False) == 1 assert endpoint_main( ['--client_certs={"foo": ["%s"]}' % cert], False) == 1 @patch('autopush.router.apns2.HTTP20Connection', spec=hyper.HTTP20Connection) @patch('hyper.tls', spec=hyper.tls) @patch('autopush.router.fcmv1client.ServiceAccountCredentials', spec=oauth2client.service_account.ServiceAccountCredentials) def test_conf(self, *args): self.TestArg.fcm_service_cred_path = "some/file.json" self.TestArg.fcm_project_id = "fir_testbridge" conf = AutopushConfig.from_argparse(self.TestArg) app = EndpointApplication(conf, resource=autopush.tests.boto_resource) # verify that the hostname is what we said. assert conf.hostname == self.TestArg.hostname assert app.routers["gcm"].router_conf['collapsekey'] == "collapse" assert app.routers["apns"].router_conf['firefox']['cert'] == \ "cert.file" assert app.routers["apns"].router_conf['firefox']['key'] == "key.file" assert app.routers["adm"].router_conf['dev']['app_id'] == \ "amzn1.application.StringOfStuff" assert app.routers["adm"].router_conf['dev']['client_id'] == \ "amzn1.application-oa2-client.ev4nM0reStuff" assert app.routers["adm"].router_conf['dev']['client_secret'] == \ "deadbeef0000decafbad1111" conf = AutopushConfig.from_argparse(self.TestArg) assert conf.router_conf['fcm']['version'] == 0 app = EndpointApplication(conf, resource=autopush.tests.boto_resource) assert app.routers["fcm"].router_conf["version"] == 0 def test_bad_senders(self): old_list = self.TestArg.senderid_list self.TestArg.senderid_list = "{}" with pytest.raises(InvalidConfig): AutopushConfig.from_argparse(self.TestArg) self.TestArg.senderid_list = old_list def test_bad_fcm_senders(self): old_list = self.TestArg.fcm_creds self.TestArg.fcm_creds = json.dumps({"12345": {"foo": "abcd"}}) with pytest.raises(InvalidConfig): AutopushConfig.from_argparse(self.TestArg) self.TestArg.fcm_creds = "{}" with pytest.raises(InvalidConfig): AutopushConfig.from_argparse(self.TestArg) self.TestArg.fcm_creds = old_list def test_gcm_start(self): endpoint_main([ "--gcm_endpoint='gcm-http.googleapis.com/gcm/send'", """--senderid_list={"123":{"auth":"abcd"}}""", ], False, resource=autopush.tests.boto_resource) @patch("requests.get") def test_aws_ami_id(self, request_mock): class MockReply: content = "ami_123" request_mock.return_value = MockReply self.TestArg.no_aws = False conf = AutopushConfig.from_argparse(self.TestArg) assert conf.ami_id == "ami_123" def test_no_sslcontext_cache(self): conf = AutopushConfig.from_argparse(self.TestArg) assert not conf.no_sslcontext_cache self.TestArg._no_sslcontext_cache = True conf = AutopushConfig.from_argparse(self.TestArg) assert conf.no_sslcontext_cache
mpl-2.0
unitedstates/congress
congress/tasks/bill_info.py
1
50773
from congress.tasks import utils import logging import re import json from lxml import etree import copy import datetime def create_govtrack_xml(bill, options): govtrack_type_codes = {'hr': 'h', 's': 's', 'hres': 'hr', 'sres': 'sr', 'hjres': 'hj', 'sjres': 'sj', 'hconres': 'hc', 'sconres': 'sc'} root = etree.Element("bill") root.set("session", bill['congress']) root.set("type", govtrack_type_codes[bill['bill_type']]) root.set("number", bill['number']) root.set("updated", utils.format_datetime(bill['updated_at'])) def make_node(parent, tag, text, **attrs): if options.get("govtrack", False): # Rewrite bioguide_id attributes as just id with GovTrack person IDs. attrs2 = {} for k, v in attrs.items(): if v: if k == "bioguide_id": # remap "bioguide_id" attributes to govtrack "id" k = "id" v = str(utils.translate_legislator_id('bioguide', v, 'govtrack')) if k == "thomas_id": # remap "thomas_id" attributes to govtrack "id" k = "id" v = str(utils.translate_legislator_id('thomas', v, 'govtrack')) attrs2[k] = v attrs = attrs2 return utils.make_node(parent, tag, text, **attrs) # for American Memory Century of Lawmaking bills... for source in bill.get("sources", []): n = make_node(root, "source", "") for k, v in sorted(source.items()): if k == "source": n.text = v elif k == "source_url": n.set("url", v) else: n.set(k, str(v)) if "original_bill_number" in bill: make_node(root, "bill-number", bill["original_bill_number"]) make_node(root, "state", bill['status'], datetime=bill['status_at']) old_status = make_node(root, "status", None) make_node(old_status, "introduced" if bill['status'] in ("INTRODUCED", "REFERRED") else "unknown", None, datetime=bill['status_at']) # dummy for the sake of comparison make_node(root, "introduced", None, datetime=bill['introduced_at']) titles = make_node(root, "titles", None) for title in bill['titles']: n = make_node(titles, "title", title['title']) n.set("type", title['type']) if title['as']: n.set("as", title['as']) if title['is_for_portion']: n.set("partial", "1") def get_legislator_id_attr(p): if "bioguide_id" in p: return { "bioguide_id": p["bioguide_id"] } if "thomas_id" in p: return { "thomas_id": p["thomas_id"] } return { } if bill['sponsor']: # TODO: Sponsored by committee? make_node(root, "sponsor", None, **get_legislator_id_attr(bill['sponsor'])) else: make_node(root, "sponsor", None) cosponsors = make_node(root, "cosponsors", None) for cosp in bill['cosponsors']: n = make_node(cosponsors, "cosponsor", None, **get_legislator_id_attr(cosp)) if cosp["sponsored_at"]: n.set("joined", cosp["sponsored_at"]) if cosp["withdrawn_at"]: n.set("withdrawn", cosp["withdrawn_at"]) actions = make_node(root, "actions", None) for action in bill['actions']: a = make_node(actions, action['type'] if action['type'] in ("vote", "vote-aux", "calendar", "topresident", "signed", "enacted", "vetoed") else "action", None, datetime=action['acted_at']) if action.get("status"): a.set("state", action["status"]) if action['type'] in ('vote', 'vote-aux'): a.clear() # re-insert date between some of these attributes a.set("how", action["how"]) a.set("type", action["vote_type"]) if action.get("roll") != None: a.set("roll", action["roll"]) a.set("datetime", utils.format_datetime(action['acted_at'])) a.set("where", action["where"]) a.set("result", action["result"]) if action.get("suspension"): a.set("suspension", "1") if action.get("status"): a.set("state", action["status"]) if action['type'] == 'calendar' and "calendar" in action: a.set("calendar", action["calendar"]) if action["under"]: a.set("under", action["under"]) if action["number"]: a.set("number", action["number"]) if action['type'] == 'enacted': a.clear() # re-insert date between some of these attributes a.set("number", "%s-%s" % (bill['congress'], action["number"])) a.set("type", action["law"]) a.set("datetime", utils.format_datetime(action['acted_at'])) if action.get("status"): a.set("state", action["status"]) if action['type'] == 'vetoed': if action.get("pocket"): a.set("pocket", "1") if action.get('text'): make_node(a, "text", action['text']) if action.get('in_committee'): make_node(a, "committee", None, name=action['in_committee']) for cr in action['references']: make_node(a, "reference", None, ref=cr['reference'], label=cr['type']) committees = make_node(root, "committees", None) for cmt in bill['committees']: make_node(committees, "committee", None, code=(cmt["committee_id"] + cmt["subcommittee_id"]) if cmt.get("subcommittee_id", None) else cmt["committee_id"], name=cmt["committee"], subcommittee=cmt.get("subcommittee").replace("Subcommittee on ", "") if cmt.get("subcommittee") else "", activity=", ".join(c.title() for c in cmt["activity"])) relatedbills = make_node(root, "relatedbills", None) for rb in bill['related_bills']: if rb['type'] == "bill": rb_bill_type, rb_number, rb_congress = utils.split_bill_id(rb['bill_id']) make_node(relatedbills, "bill", None, session=rb_congress, type=govtrack_type_codes[rb_bill_type], number=rb_number, relation="unknown" if rb['reason'] == "related" else rb['reason']) subjects = make_node(root, "subjects", None) if bill['subjects_top_term']: make_node(subjects, "term", None, name=bill['subjects_top_term']) for s in bill['subjects']: if s != bill['subjects_top_term']: make_node(subjects, "term", None, name=s) amendments = make_node(root, "amendments", None) for amd in bill['amendments']: make_node(amendments, "amendment", None, number=amd["chamber"] + str(amd["number"])) if bill.get('summary'): make_node(root, "summary", bill['summary']['text'], date=bill['summary']['date'], status=bill['summary']['as']) if bill.get('committee_reports'): committee_reports = make_node(root, "committee-reports", None) for report in bill.get('committee_reports', []): make_node(committee_reports, "report", report) return etree.tostring(root, pretty_print=True) def sponsor_for(sponsor_dict): if sponsor_dict is None: # TODO: This can hopefully be removed. In testing s414-113 # was missing sponsor data. But all bills have a sponsor? return None # TODO: Don't do regex matching here. Find another way. m = re.match(r'(?P<title>(Rep\.|Sen\.|Del\.|Resident Commissioner)) (?P<name>.*?) +\[(?P<party>[DRIL])-(?P<state>[A-Z][A-Z])(-(?P<district>\d{1,2}|At Large|None))?\]$', sponsor_dict['fullName']) if not m: raise ValueError(sponsor_dict) return { 'title': m.group("title"), 'name': m.group("name"), # the firstName, middleName, lastName fields have inconsistent capitalization - some are all uppercase 'state': sponsor_dict["state"], 'district': sponsor_dict.get("district"), # missing for senators #'party': m.group('party'), 'bioguide_id': sponsor_dict['bioguideId'], 'type': 'person' } def summary_for(summaries): # Some bills are missing the summaries entirely? if summaries is None: return None # Take the most recent summary, by looking at the lexicographically last updateDate. summaries = summaries['item'] summary = sorted(summaries, key = lambda s: s['updateDate'])[-1] # Build dict. return { "date": summary['updateDate'], "as": summary['name'], "text": strip_tags(summary['text']), } def strip_tags(text): # Preserve paragraph breaks. Convert closing p tags (and surrounding whitespace) into two newlines. Strip trailing whitespace text = re.sub("\s*</\s*p\s*>\s*", "\n\n", text).strip() # naive stripping of tags, should work okay in this limited context text = re.sub("<[^>]+>", "", text) # compress and strip whitespace artifacts, except for the paragraph breaks text = re.sub("[ \t\r\f\v]{2,}", " ", text).strip() # Replace HTML entities with characters. text = utils.unescape(text) return text def committees_for(committee_list): if committee_list is None: return [] committee_list = committee_list['item'] activity_text_map = { "Referred to": ["referral"], "Hearings by": ["hearings"], "Markup by": ["markup"], "Reported by": ["reporting"], "Discharged from": ["discharged"], "Reported original measure": ["origin", "reporting"], } def fix_subcommittee_name(name): return re.sub("(.*) Subcommittee$", lambda m : "Subcommittee on " + m.group(1), name) def get_activitiy_list(item): if not item['activities']: return [] return sum([activity_text_map.get(i['name'], [i['name']]) for i in item['activities']['item']], []) def fixup_committee_name(name): # Preserve backwards compatiblity. if name == "House House Administration": return "House Administration" return name def build_dict(item): committee_dict = { 'activity': get_activitiy_list(item), 'committee': fixup_committee_name(item['chamber'] + ' ' + re.sub(" Committee$", "", item['name'])), 'committee_id': item['systemCode'][0:-2].upper(), } subcommittees_list = [] if 'subcommittees' in item and item['subcommittees'] is not None: for subcommittee in item['subcommittees']['item']: subcommittee_dict = copy.deepcopy(committee_dict) subcommittee_dict.update({ 'subcommittee': fix_subcommittee_name(subcommittee['name']), 'subcommittee_id': subcommittee['systemCode'][-2:], 'activity': get_activitiy_list(subcommittee), }) subcommittees_list.append(subcommittee_dict) return [committee_dict] + subcommittees_list return sum([build_dict(committee) for committee in committee_list], []) def titles_for(title_list): def build_dict(item): full_type = item['titleType'] is_for_portion = False # "Official Titles as Introduced", "Short Titles on Conference report" splits = re.split(" as | on ", full_type, 1) if len(splits) == 2: title_type, state = splits if state.endswith(" for portions of this bill"): is_for_portion = True state = state.replace(" for portions of this bill" ,"") state = state.replace(":", "").lower() else: title_type, state = full_type, None if "Popular Title" in title_type: title_type = "popular" elif "Short Title" in title_type: title_type = "short" elif "Official Title" in title_type: title_type = "official" elif "Display Title" in title_type: title_type = "display" elif title_type == "Non-bill-report": # TODO: What kind of title is this? Maybe assign # a better title_type code once we know. title_type = "nonbillreport" else: raise Exception("Unknown title type: " + title_type) return { 'title': item['title'], 'is_for_portion': is_for_portion, 'as': state, 'type': title_type } titles = [build_dict(title) for title in title_list] # THOMAS used to give us the titles in a particular order: # short as introduced # short as introduced (for portion) # short as some later stage # short as some later stage (for portion) # official as introduced # official as some later stage # The "as" stages (introduced, etc.) were in the order in which actions # actually occurred. This was handy because to get the current title for # a bill, you need to know which action type was most recent. The new # order is reverse-chronological, so we have to turn the order around # for backwards compatibility. Rather than do a simple .reverse(), I'm # adding an explicit sort order here which gets very close to the THOMAS # order. # Unfortunately this can no longer be relied on because the new bulk # data has the "as" stages sometimes in the wrong order: The "reported to # senate" status for House bills seems to be consistently out of place. titles_copy = list(titles) # clone before beginning sort def first_index_of(**kwargs): for i, title in enumerate(titles_copy): for k, v in kwargs.items(): k = k.replace("_", "") if title.get(k) != v: break else: # break not called --- all match return i titles.sort(key = lambda title: ( # keep the same 'short', 'official', 'display' order intact first_index_of(type=title['type']), # within each of those categories, reverse the 'as' order -first_index_of(type=title['type'], _as=title.get('as')), # put titles for portions last, within the type/as category title['is_for_portion'], # and within that, just sort alphabetically, case-insensitively (which is # what it appears THOMAS used to do) title['title'].lower(), )) return titles # the most current title of a given type is the first one in the last 'as' subgroup # of the titles for the whole bill (that is, if there's no title for the whole bill # in the last 'as' subgroup, use the previous 'as' subgroup and so on) --- we think # this logic matches THOMAS/Congress.gov. def current_title_for(titles, title_type): current_title = None current_as = -1 # not None, cause for popular titles, None is a valid 'as' for title in titles: if title['type'] != title_type or title['is_for_portion'] == True: continue if title['as'] == current_as: continue # right type, new 'as', store first one current_title = title['title'] current_as = title['as'] return current_title def actions_for(action_list, bill_id, title): # The bulk XML data has action history information from multiple sources. For # major actions, the Library of Congress (code 9) action item often duplicates # the information of a House/Senate action item. We have to skip one so that we # don't tag multiple history items with the same parsed action info, which # would imply the action (like a vote) ocurred multiple times. THOMAS appears # to have suppressed the Library of Congress action lines in certain cases # to avoid duplication - they were not in our older data files. # # Also, there are some ghost action items with totally empty text. Remove those. # TODO: When removed from upstream data, we can remove that check. closure = { "prev": None, } def keep_action(item, closure): if item['text'] in (None, ""): return False keep = True if closure['prev']: if item['sourceSystem']['code'] == "9": # Date must match previous action.. # If both this and previous have a time, the times must match. # The text must approximately match. Sometimes the LOC text has a prefix # and different whitespace. And they may drop references -- so we'll # use our action_for helper function to drop references from both # prior to the string comparison. if item['actionDate'] == closure["prev"]["actionDate"] \ and (item.get('actionTime') == closure["prev"].get("actionTime") or not item.get('actionTime') or not closure["prev"].get("actionTime")) \ and action_for(item)['text'].replace(" ", "").endswith(action_for(closure["prev"])['text'].replace(" ", "")): keep = False closure['prev'] = item return keep action_list = [item for item in action_list if keep_action(item, closure)] # Turn the actions into dicts. The actions are in reverse-chronological # order in the bulk data XML. Process them in chronological order so that # our bill status logic sees the actions in the right order. def build_dict(item, closure): action_dict = action_for(item) extra_action_info, new_status = parse_bill_action(action_dict, closure['prev_status'], bill_id, title) # only change/reflect status change if there was one if new_status: action_dict['status'] = new_status closure['prev_status'] = new_status # add additional parsed fields if extra_action_info: action_dict.update(extra_action_info) return action_dict closure = { "prev_status": "INTRODUCED", } return [build_dict(action, closure) for action in reversed(action_list)] # clean text, pull out the action type, any other associated metadata with an action def action_for(item): # acted_at if not item.get('actionTime'): acted_at = item.get('actionDate', '') else: # Although we get the action date & time in an ISO-ish format (split # across two fields), and although we know it's in local time at the # U.S. Capitol (i.e. U.S. Eastern), we don't know the UTC offset which # is a part of how we used to serialize the time. So parse and then # use pytz (via format_datetime) to re-serialize. acted_at = utils.format_datetime(datetime.datetime.strptime(item.get('actionDate', '') + " " + item['actionTime'], "%Y-%m-%d %H:%M:%S")) # text & references # (amendment actions don't always have text?) text = item['text'] if item['text'] is not None else '' # strip out links text = re.sub(r"</?[Aa]( \S.*?)?>", "", text) # remove and extract references references = [] match = re.search("\s*\(([^)]+)\)\s*$", text) if match: # remove the matched section text = text[0:match.start()] + text[match.end():] types = match.group(1) # fix use of comma or colon instead of a semi colon between reference types # have seen some accidental capitalization combined with accidental comma, thus the 'T' # e.g. "text of Title VII as reported in House: CR H3075-3077, Text omission from Title VII:" (hr5384-109) types = re.sub("[,:] ([a-zT])", r"; \1", types) # fix "CR:" types = re.sub("CR:", "CR", types) # fix a missing semicolon altogether between references # e.g. sres107-112, "consideration: CR S1877-1878 text as" types = re.sub("(\d+) +([a-z])", r"\1; \2", types) for reference in re.split("; ?", types): if ": " not in reference: type, reference = None, reference else: type, reference = reference.split(": ", 1) references.append({'type': type, 'reference': reference}) # extract committee IDs if item.get('committee'): # Data format through Dec. 13, 2019 had only one <committee/> (though node could be empty). committee_nodes = [item['committee']] elif item.get('committees'): # Starting on Dec. 13, 2019, and with a slow rollout, multiple committees could be specified. # Thankfully our JSON output format allowed it already. committee_nodes = item['committees'].get("item", []) else: # <committee/> or <committees/>, whichever was present, was empty committee_nodes = [] # form dict action_dict = { 'acted_at': acted_at, 'action_code': item.get('actionCode', ''), 'committees': [committee_item['systemCode'][0:-2].upper() for committee_item in committee_nodes] if committee_nodes else None, # if empty, store None 'references': references, 'type': 'action', # replaced by parse_bill_action if a regex matches 'text': text, } if not action_dict["committees"]: # remove if empty - not present in how we used to generate the file del action_dict["committees"] # sometimes there are links (one case is for bills passed by a rule in a resolution, the link will point to the resolution) if (item.get("links") or {}).get("link") is not None: action_dict["links"] = item["links"]["link"] return action_dict def cosponsors_for(cosponsors_list): if cosponsors_list is None: return [] cosponsors_list = cosponsors_list['item'] def build_dict(item): cosponsor_dict = sponsor_for(item) del cosponsor_dict["type"] # always 'person' cosponsor_dict.update({ 'sponsored_at': item['sponsorshipDate'], 'withdrawn_at': item['sponsorshipWithdrawnDate'], 'original_cosponsor': item['isOriginalCosponsor'] == 'True' }) return cosponsor_dict cosponsors = [build_dict(cosponsor) for cosponsor in cosponsors_list] # TODO: Can remove. Sort like the old THOMAS order to make diffs easier. cosponsors.sort(key = lambda c: c['name'].lower()) return cosponsors def related_bills_for(related_bills_list): if related_bills_list is None: return [] related_bills_list = related_bills_list['item'] def build_dict(item): return { 'reason': item['relationshipDetails']['item'][0]['type'].replace('bill', '').strip().lower(), 'bill_id': '{0}{1}-{2}'.format(item['type'].replace('.', '').lower(), item['number'], item['congress']), 'type': 'bill', 'identified_by': item['relationshipDetails']['item'][0]['identifiedBy'] } # Are these THOMAS related bill relation texts gone from the bulk data? reasons = ( ("Identical bill identified by (CRS|House|Senate)", "identical"), ("Companion bill", "identical"), ("Related bill (as )?identified by (CRS|the House Clerk's office|House committee|Senate)", "related"), ("passed in (House|Senate) in lieu of .*", "supersedes"), ("Rule related to .* in (House|Senate)", "rule"), ("This bill has text inserted from .*", "includes"), ("Text from this bill was inserted in .*", "included-in"), ("Bill related to rule .* in House", "ruled-by"), ("This bill caused other related action on .*", "caused-action"), ("Other related action happened to this bill because of .*", "action-caused-by"), ("Bill that causes .* to be laid on table in House", "caused-action"), ("Bill laid on table by virtue of .* passage in House", "action-caused-by"), ("Bill that caused the virtual passage of .* in House", "caused-action"), ("Bill passed by virtue of .* passage in House", "caused-action-by"), ("Bill on wich enrollment has been corrected by virtue of .* passage in House", "caused-action"), ) return [build_dict(related_bill) for related_bill in related_bills_list] # get the public or private law number from any enacted action def slip_law_from(actions): for action in actions: if action["type"] == "enacted": return { 'law_type': action["law"], 'congress': action["congress"], 'number': action["number"] } # find the latest status change in a set of processed actions def latest_status(actions, introduced_at): status, status_date = "INTRODUCED", introduced_at for action in actions: if action.get('status', None): status = action['status'] status_date = action['acted_at'] return status, status_date # look at the final set of processed actions and pull out the major historical events def history_from_actions(actions): history = {} activation = activation_from(actions) if activation: history['active'] = True history['active_at'] = activation['acted_at'] else: history['active'] = False house_vote = None for action in actions: if (action['type'] == 'vote') and (action['where'] == 'h') and (action['vote_type'] != "override"): house_vote = action if house_vote: history['house_passage_result'] = house_vote['result'] history['house_passage_result_at'] = house_vote['acted_at'] senate_vote = None for action in actions: if (action['type'] == 'vote') and (action['where'] == 's') and (action['vote_type'] != "override"): senate_vote = action if senate_vote: history['senate_passage_result'] = senate_vote['result'] history['senate_passage_result_at'] = senate_vote['acted_at'] senate_vote = None for action in actions: if (action['type'] == 'vote-aux') and (action['vote_type'] == 'cloture') and (action['where'] == 's') and (action['vote_type'] != "override"): senate_vote = action if senate_vote: history['senate_cloture_result'] = senate_vote['result'] history['senate_cloture_result_at'] = senate_vote['acted_at'] vetoed = None for action in actions: if action['type'] == 'vetoed': vetoed = action if vetoed: history['vetoed'] = True history['vetoed_at'] = vetoed['acted_at'] else: history['vetoed'] = False house_override_vote = None for action in actions: if (action['type'] == 'vote') and (action['where'] == 'h') and (action['vote_type'] == "override"): house_override_vote = action if house_override_vote: history['house_override_result'] = house_override_vote['result'] history['house_override_result_at'] = house_override_vote['acted_at'] senate_override_vote = None for action in actions: if (action['type'] == 'vote') and (action['where'] == 's') and (action['vote_type'] == "override"): senate_override_vote = action if senate_override_vote: history['senate_override_result'] = senate_override_vote['result'] history['senate_override_result_at'] = senate_override_vote['acted_at'] enacted = None for action in actions: if action['type'] == 'enacted': enacted = action if enacted: history['enacted'] = True history['enacted_at'] = action['acted_at'] else: history['enacted'] = False topresident = None for action in actions: if action['type'] == 'topresident': topresident = action if topresident and (not history['vetoed']) and (not history['enacted']): history['awaiting_signature'] = True history['awaiting_signature_since'] = action['acted_at'] else: history['awaiting_signature'] = False return history # find the first action beyond the standard actions every bill gets. # - if the bill's first action is "referral" then the first action not those # most common # e.g. hr3590-111 (active), s1-113 (inactive) # - if the bill's first action is "action", then the next action, if one is present # resolutions # e.g. sres5-113 (active), sres4-113 (inactive) # - if the bill's first action is anything else (e.g. "vote"), then that first action # bills that skip committee # e.g. s227-113 (active) def activation_from(actions): # there's NOT always at least one :( # as of 2013-06-10, hr2272-113 has no actions at all if len(actions) == 0: return None first = actions[0] if first['type'] in ["referral", "calendar", "action"]: for action in actions[1:]: if (action['type'] != "referral") and (action['type'] != "calendar") and ("Sponsor introductory remarks" not in action['text']): return action return None else: return first def parse_bill_action(action_dict, prev_status, bill_id, title): """Parse a THOMAS bill action line. Returns attributes to be set in the XML file on the action line.""" bill_type, number, congress = utils.split_bill_id(bill_id) line = action_dict['text'] status = None action = { "type": "action" } # If a line starts with an amendment number, this action is on the amendment and cannot # be parsed yet. m = re.search(r"^(H|S)\.Amdt\.(\d+)", line, re.I) if m != None: # Process actions specific to amendments separately. return None, None # Otherwise, parse the action line for key actions. # VOTES # A House Vote. line = re.sub(", the Passed", ", Passed", line) # 106 h4733 and others m = re.search("(" + "|".join([ "On passage", "Passed House", "Two-thirds of the Members present having voted in the affirmative the bill is passed,?", "On motion to suspend the rules and pass the (?:bill|resolution)", "On agreeing to the (?:resolution|conference report)", "On motion to suspend the rules and agree to the (?:resolution|conference report)", "House Agreed to Senate Amendments.*?", "On motion (?:that )?the House (?:suspend the rules and )?(?:agree(?: with an amendment)? to|concur in) the Senate amendments?(?: to the House amendments?| to the Senate amendments?)*", ]) + ")" + "(, the objections of the President to the contrary notwithstanding.?)?" + "(, as amended| \(Amended\))?" + "\.? (Passed|Failed|Agreed to|Rejected)?" # hr1625-115 has a stray period here + " ?(by voice vote|without objection|by (the Yeas and Nays?|Yea-Nay Vote|recorded vote)" + "(:? \(2/3 required\))?: (\d+ ?- ?\d+(, \d+ Present)? [ \)]*)?\((Roll no\.|Record Vote No:) \d+\))", line, re.I) if m != None: motion, is_override, as_amended, pass_fail, how = m.group(1), m.group(2), m.group(3), m.group(4), m.group(5) # print(line) # print(m.groups()) if re.search(r"Passed House|House Agreed to", motion, re.I): pass_fail = 'pass' elif re.search("(ayes|yeas) had prevailed", line, re.I): pass_fail = 'pass' elif re.search(r"Pass|Agreed", pass_fail, re.I): pass_fail = 'pass' else: pass_fail = 'fail' if "Two-thirds of the Members present" in motion: is_override = True if is_override: vote_type = "override" elif re.search(r"(agree (with an amendment )?to|concur in) the Senate amendment", line, re.I): vote_type = "pingpong" elif re.search("conference report", line, re.I): vote_type = "conference" elif bill_type[0] == "h": vote_type = "vote" else: vote_type = "vote2" roll = None m = re.search(r"\((Roll no\.|Record Vote No:) (\d+)\)", how, re.I) if m != None: how = "roll" # normalize the ugly how roll = m.group(2) suspension = None if roll and "On motion to suspend the rules" in motion: suspension = True # alternate form of as amended, e.g. hr3979-113 if "the House agree with an amendment" in motion: as_amended = True action["type"] = "vote" action["vote_type"] = vote_type action["how"] = how action['where'] = "h" action['result'] = pass_fail if roll: action["roll"] = roll action["suspension"] = suspension # correct upstream data error if bill_id == "s2012-114" and "Roll no. 250" in line: as_amended = True if bill_id == "s2943-114" and "On passage Passed without objection" in line: as_amended = True # get the new status of the bill after this vote new_status = new_status_after_vote(vote_type, pass_fail == "pass", "h", bill_type, suspension, as_amended, title, prev_status) if new_status: status = new_status # Passed House, not necessarily by an actual vote (think "deem") m = re.search(r"Passed House pursuant to|House agreed to Senate amendment (with amendment )?pursuant to|Pursuant to the provisions of [HSCONJRES\. ]+ \d+, [HSCONJRES\. ]+ \d+ is considered passed House", line, re.I) if m != None: vote_type = "vote" if (bill_type[0] == "h") else "vote2" if "agreed to Senate amendment" in line: vote_type = "pingpong" pass_fail = "pass" as_amended = ("with amendment" in line) or ("as amended" in line) action["type"] = "vote" action["vote_type"] = vote_type action["how"] = "by special rule" action["where"] = "h" action["result"] = pass_fail # It's always pursuant to another bill, and a bill number is given in the action line, which we parse out # into the bill_ids field of the action. It's also represented # structurally in the links->link elements of the original XML which we just put in "links". # get the new status of the bill after this vote new_status = new_status_after_vote(vote_type, pass_fail == "pass", "h", bill_type, False, as_amended, title, prev_status) if new_status: status = new_status # House motions to table adversely dispose of a pending matter, if agreed to. An agreed-to "motion to table the measure", # which is very infrequent, kills the legislation. If not agreed to, nothing changes. So this regex only captures # agreed-to motions to table. m = re.search("On motion to table the measure Agreed to" + " ?(by voice vote|without objection|by (the Yeas and Nays|Yea-Nay Vote|recorded vote)" + ": (\d+ - \d+(, \d+ Present)? [ \)]*)?\((Roll no\.|Record Vote No:) \d+\))", line, re.I) if m != None: how = m.group(1) pass_fail = 'fail' # In order to classify this as resulting in the same thing as regular failed vote on passage, new_status_after_vote # needs to know if this was a vote in the originating chamber or not. if prev_status == "INTRODUCED" or bill_id.startswith("hres"): vote_type = "vote" elif False: vote_type = "vote2" else: raise Exception("Need to classify %s as being in the originating chamber or not." % prev_status) roll = None m = re.search(r"\((Roll no\.|Record Vote No:) (\d+)\)", how, re.I) if m != None: how = "roll" # normalize the ugly how roll = m.group(2) action["type"] = "vote" action["vote_type"] = vote_type action["how"] = how action['where'] = "h" action['result'] = pass_fail if roll: action["roll"] = roll # get the new status of the bill after this vote new_status = new_status_after_vote(vote_type, pass_fail == "pass", "h", bill_type, False, False, title, prev_status) if new_status: status = new_status # A Senate Vote # (There are some annoying weird cases of double spaces which are taken care of # at the end.) m = re.search("(" + "|".join([ "Passed Senate", "Failed of passage in Senate", "Disagreed to in Senate", "Resolution agreed to in Senate", "Senate (?:agreed to|concurred in) (?:the )?(?:conference report|House amendment(?: to the Senate amendments?| to the House amendments?)*)", "Senate receded from its amendment and concurred", # hr1-115 r"Cloture \S*\s?on the motion to proceed .*?not invoked in Senate", r"Cloture(?: motion)? on the motion to proceed to the (?:bill|measure) invoked in Senate", "Cloture invoked in Senate", "Cloture on (?:the motion to (?:proceed to |concur in )(?:the House amendment (?:to the Senate amendment )?to )?)(?:the bill|H.R. .*) (?:not )?invoked in Senate", "(?:Introduced|Received|Submitted) in the Senate, (?:read twice, |considered, |read the third time, )+and (?:passed|agreed to)", ]) + ")" + "(,?.*,?) " + "(without objection|by Unanimous Consent|by Voice Vote|(?:by )?Yea-Nay( Vote)?\. \d+\s*-\s*\d+\. Record Vote (No|Number): \d+)", line.replace(" ", " "), re.I) if m != None: motion, extra, how = m.group(1), m.group(2), m.group(3) roll = None # put disagreed check first, cause "agreed" is contained inside it if re.search("disagreed|not invoked", motion, re.I): pass_fail = "fail" elif re.search("passed|agreed|concurred|invoked", motion, re.I): pass_fail = "pass" else: pass_fail = "fail" voteaction_type = "vote" if re.search("over veto", extra, re.I): vote_type = "override" elif re.search("conference report", motion, re.I): vote_type = "conference" elif re.search("cloture", motion, re.I): vote_type = "cloture" voteaction_type = "vote-aux" # because it is not a vote on passage elif re.search("Senate agreed to (the )?House amendment|Senate concurred in (the )?House amendment", motion, re.I): vote_type = "pingpong" elif bill_type[0] == "s": vote_type = "vote" else: vote_type = "vote2" m = re.search(r"Record Vote (No|Number): (\d+)", how, re.I) if m != None: roll = m.group(2) how = "roll" as_amended = False if re.search(r"with amendments|with an amendment", extra, re.I): as_amended = True action["type"] = voteaction_type action["vote_type"] = vote_type action["how"] = how action["result"] = pass_fail action["where"] = "s" if roll: action["roll"] = roll # get the new status of the bill after this vote new_status = new_status_after_vote(vote_type, pass_fail == "pass", "s", bill_type, False, as_amended, title, prev_status) if new_status: status = new_status # OLD-STYLE VOTES (93rd Congress-ish) m = re.search(r"Measure passed (House|Senate)(, amended(?: \(.*?\)|, with an amendment to the title)?)?(?:,? in lieu[^,]*)?(?:, roll call #(\d+) \(\d+-\d+\))?", line, re.I) if m != None: chamber = m.group(1)[0].lower() # 'h' or 's' as_amended = m.group(2) roll_num = m.group(3) # GovTrack legacy scraper missed these: if chamber == 's' and (as_amended or roll_num or "lieu" in line): return action, status pass_fail = "pass" vote_type = "vote" if bill_type[0] == chamber else "vote2" action["type"] = "vote" action["vote_type"] = vote_type action["how"] = "(method not recorded)" if not roll_num else "roll" if roll_num: action["roll"] = roll_num action["result"] = pass_fail action["where"] = chamber new_status = new_status_after_vote(vote_type, pass_fail == "pass", chamber, bill_type, False, as_amended, title, prev_status) if new_status: status = new_status m = re.search(r"(House|Senate) agreed to (?:House|Senate) amendments?( with an amendment)?( under Suspension of the Rules)?(?:, roll call #(\d+) \(\d+-\d+\))?\.", line, re.I) if m != None: chamber = m.group(1)[0].lower() # 'h' or 's' as_amended = m.group(2) suspension = m.group(3) roll_num = m.group(4) # GovTrack legacy scraper missed these: if (chamber == 'h' and not roll_num) or (chamber == 's' and rull_num): return action, status # REMOVE ME pass_fail = "pass" vote_type = "pingpong" action["type"] = "vote" action["vote_type"] = vote_type action["how"] = "(method not recorded)" if not roll_num else "roll" if roll_num: action["roll"] = roll_num action["result"] = pass_fail action["where"] = chamber action["suspension"] = (suspension != None) new_status = new_status_after_vote(vote_type, pass_fail == "pass", chamber, bill_type, False, as_amended, title, prev_status) if new_status: status = new_status # PSUDO-REPORTING (because GovTrack did this, but should be changed) # TODO: Make a new status for this as pre-reported. m = re.search(r"Placed on (the )?([\w ]+) Calendar( under ([\w ]+))?[,\.] Calendar No\. (\d+)\.|Committee Agreed to Seek Consideration Under Suspension of the Rules|Ordered to be Reported", line, re.I) if m != None: # TODO: This makes no sense. if prev_status in ("INTRODUCED", "REFERRED"): status = "REPORTED" action["type"] = "calendar" # TODO: Useless. But good for GovTrack compatibility. if m.group(2): # not 'Ordered to be Reported' action["calendar"] = m.group(2) action["under"] = m.group(4) action["number"] = m.group(5) # COMMITTEE ACTIONS # reported m = re.search(r"Committee on (.*)\. Reported by", line, re.I) if m != None: action["type"] = "reported" action["committee"] = m.group(1) if prev_status in ("INTRODUCED", "REFERRED"): status = "REPORTED" m = re.search(r"Reported to Senate from the (.*?)( \(without written report\))?\.", line, re.I) if m != None: # 93rd Congress action["type"] = "reported" action["committee"] = m.group(1) if prev_status in ("INTRODUCED", "REFERRED"): status = "REPORTED" # hearings held by a committee m = re.search(r"(Committee on .*?)\. Hearings held", line, re.I) if m != None: action["committee"] = m.group(1) action["type"] = "hearings" m = re.search(r"Committee on (.*)\. Discharged (by Unanimous Consent)?", line, re.I) if m != None: action["committee"] = m.group(1) action["type"] = "discharged" if prev_status in ("INTRODUCED", "REFERRED"): status = "REPORTED" m = re.search("Cleared for White House|Presented to President", line, re.I) if m != None: action["type"] = "topresident" m = re.search("Signed by President", line, re.I) if m != None: action["type"] = "signed" status = "ENACTED:SIGNED" m = re.search("Pocket Vetoed by President", line, re.I) if m != None: action["type"] = "vetoed" action["pocket"] = "1" status = "VETOED:POCKET" # need to put this in an else, or this regex will match the pocket veto and override it else: m = re.search("Vetoed by President", line, re.I) if m != None: action["type"] = "vetoed" status = "PROV_KILL:VETO" m = re.search("Sent to Archivist of the United States unsigned", line, re.I) if m != None: status = "ENACTED:TENDAYRULE" m = re.search("^(?:Became )?(Public|Private) Law(?: No:)? ([\d\-]+)\.", line, re.I) if m != None: action["law"] = m.group(1).lower() pieces = m.group(2).split("-") action["congress"] = pieces[0] action["number"] = pieces[1] action["type"] = "enacted" if prev_status in ("ENACTED:SIGNED", "ENACTED:VETO_OVERRIDE", "ENACTED:TENDAYRULE"): pass # this is a final administrative step elif prev_status == "PROV_KILL:VETO" or prev_status.startswith("VETOED:"): # somehow missed the override steps status = "ENACTED:VETO_OVERRIDE" elif bill_id in ("s2641-93", "hr1589-94", "s2527-100", "hr1677-101", "hr2978-101", "hr2126-104", "s1322-104"): status = "ENACTED:TENDAYRULE" else: raise Exception("Missing Signed by President action? If this is a case of the 10-day rule, hard code the bill id %s here." % bill_id) # Check for referral type m = re.search(r"Referred to (?:the )?(House|Senate)?\s?(?:Committee|Subcommittee)?", line, re.I) if m != None: action["type"] = "referral" if prev_status == "INTRODUCED": status = "REFERRED" # sweep the action line for bill IDs of related bills bill_ids = utils.extract_bills(line, congress) bill_ids = [b for b in bill_ids if b != bill_id] if bill_ids and (len(bill_ids) > 0): action['bill_ids'] = bill_ids return action, status def new_status_after_vote(vote_type, passed, chamber, bill_type, suspension, amended, title, prev_status): if vote_type == "vote": # vote in originating chamber if passed: if bill_type in ("hres", "sres"): return 'PASSED:SIMPLERES' # end of life for a simple resolution if chamber == "h": return 'PASS_OVER:HOUSE' # passed by originating chamber, now in second chamber else: return 'PASS_OVER:SENATE' # passed by originating chamber, now in second chamber if suspension: return 'PROV_KILL:SUSPENSIONFAILED' # provisionally killed by failure to pass under suspension of the rules if chamber == "h": return 'FAIL:ORIGINATING:HOUSE' # outright failure else: return 'FAIL:ORIGINATING:SENATE' # outright failure if vote_type in ("vote2", "pingpong"): # vote in second chamber or subsequent pingpong votes if passed: if amended: # mesure is passed but not in identical form if chamber == "h": return 'PASS_BACK:HOUSE' # passed both chambers, but House sends it back to Senate else: return 'PASS_BACK:SENATE' # passed both chambers, but Senate sends it back to House else: # bills and joint resolutions not constitutional amendments, not amended from Senate version if bill_type in ("hjres", "sjres") and title.startswith("Proposing an amendment to the Constitution of the United States"): return 'PASSED:CONSTAMEND' # joint resolution that looks like an amendment to the constitution if bill_type in ("hconres", "sconres"): return 'PASSED:CONCURRENTRES' # end of life for concurrent resolutions return 'PASSED:BILL' # passed by second chamber, now on to president if vote_type == "pingpong": # chamber failed to accept the other chamber's changes, but it can vote again return 'PROV_KILL:PINGPONGFAIL' if suspension: return 'PROV_KILL:SUSPENSIONFAILED' # provisionally killed by failure to pass under suspension of the rules if chamber == "h": return 'FAIL:SECOND:HOUSE' # outright failure else: return 'FAIL:SECOND:SENATE' # outright failure if vote_type == "cloture": if not passed: return "PROV_KILL:CLOTUREFAILED" else: return None if vote_type == "override": if not passed: if bill_type[0] == chamber: if chamber == "h": return 'VETOED:OVERRIDE_FAIL_ORIGINATING:HOUSE' else: return 'VETOED:OVERRIDE_FAIL_ORIGINATING:SENATE' else: if chamber == "h": return 'VETOED:OVERRIDE_FAIL_SECOND:HOUSE' else: return 'VETOED:OVERRIDE_FAIL_SECOND:SENATE' else: if bill_type[0] == chamber: if chamber == "h": return 'VETOED:OVERRIDE_PASS_OVER:HOUSE' else: return 'VETOED:OVERRIDE_PASS_OVER:SENATE' else: # The override passed both chambers -- the veto is overridden. return "ENACTED:VETO_OVERRIDE" if vote_type == "conference": # This is tricky to integrate into status because we have to wait for both # chambers to pass the conference report. if passed: if prev_status.startswith("CONFERENCE:PASSED:"): if bill_type in ("hjres", "sjres") and title.startswith("Proposing an amendment to the Constitution of the United States"): return 'PASSED:CONSTAMEND' # joint resolution that looks like an amendment to the constitution if bill_type in ("hconres", "sconres"): return 'PASSED:CONCURRENTRES' # end of life for concurrent resolutions return 'PASSED:BILL' else: if chamber == "h": return 'CONFERENCE:PASSED:HOUSE' else: return 'CONFERENCE:PASSED:SENATE' return None def amendments_for(amendment_list): if amendment_list is None: return [] amendment_list = amendment_list['amendment'] def build_dict(item): # Malformed XML containing duplicate elements causes attributes to parse as a list for attr in ['type', 'number', 'congress']: if type(item[attr]) is list: item[attr] = item[attr][0] return { 'amendment_id': "{0}{1}-{2}".format(item['type'].lower(), item['number'], item['congress']), 'amendment_type': item['type'].lower(), 'chamber': item['type'][0].lower(), 'number': item['number'] } return [build_dict(amendment) for amendment in amendment_list] def committee_reports_for(committeeReports): ret = [] for report in (committeeReports or {}).get("committeeReport", []): ret.append( report["citation"] ) return ret
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/mock_helpers.py
1
1924
from unittest.mock import Mock class DynamicObject: pass class MockFilter(object): # This is the ONE parameter constructor def __init__(self): self._count = 0 self._first = DynamicObject() def first(self): # This is the another method that's just coming along for the ride. return self._first def count(self): # This is the needed Count method return self._count class MockQuery(object): # This is the ONE parameter constructor def __init__(self): self._filter = MockFilter() self._filter_by = MockFilter() def filter(self, place_holder): # This is used to mimic the query.filter() call return self._filter def filter_by(self, **kwargs): # This is used to mimic the query.filter_by() call return self._filter_by class MockSession(object): def __init__(self): self._query = MockQuery() self.dirty = [] def flush(self): pass def query(self, place_holder): # This is used to mimic the session.query call return self._query def mock_response( status=200, content="CONTENT", json_data=None, raise_for_status=None, url=None): """ since we typically test a bunch of different requests calls for a service, we are going to do a lot of mock responses, so its usually a good idea to have a helper function that builds these things """ mock_resp = Mock() # mock raise_for_status call w/optional error mock_resp.raise_for_status = Mock() if raise_for_status: mock_resp.raise_for_status.side_effect = raise_for_status # set status code and content mock_resp.status_code = status mock_resp.content = content mock_resp.url = url # add json data if provided if json_data: mock_resp.json = Mock( return_value=json_data ) return mock_resp
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_c23_award_financial_4.py
1
7997
from random import choice from string import ascii_uppercase, ascii_lowercase, digits from tests.unit.dataactcore.factories.staging import AwardFinancialFactory, AwardFinancialAssistanceFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'c23_award_financial_4' def test_column_headers(database): expected_subset = {'source_row_number', 'source_value_uri', 'source_value_transaction_obligated_amou_sum', 'target_value_federal_action_obligation_sum', 'target_value_original_loan_subsidy_cost_sum', 'difference', 'uniqueid_URI'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Test for each unique URI in File C, the sum of each TransactionObligatedAmount should match (but with opposite signs) the sum of the FederalActionObligation or OriginalLoanSubsidyCost amounts reported in D2. This rule does not apply if the ATA field is populated and is different from the Agency ID. """ # Create a 12 character random uri uri_1 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_2 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_3 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_4 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) # Simple sum af_1_row_1 = AwardFinancialFactory(transaction_obligated_amou=1100, uri=uri_1, allocation_transfer_agency=None) af_1_row_2 = AwardFinancialFactory(transaction_obligated_amou=11, uri=uri_1.lower(), allocation_transfer_agency=None) # Non-ignored rows with a matching ATA/AID af_2_row_1 = AwardFinancialFactory(transaction_obligated_amou=9900, uri=uri_2, allocation_transfer_agency=None) af_2_row_2 = AwardFinancialFactory(transaction_obligated_amou=99, uri=uri_2.lower(), allocation_transfer_agency='good', agency_identifier='good') # Ignored row with non-matching ATA/AID af_3 = AwardFinancialFactory(transaction_obligated_amou=8888, uri=uri_3, allocation_transfer_agency='good', agency_identifier='bad') # No TOA in File C, ignored af_4 = AwardFinancialFactory(transaction_obligated_amou=None, uri=uri_4.lower(), allocation_transfer_agency='good', agency_identifier='good') # Correct sum afa_1_row_1 = AwardFinancialAssistanceFactory(uri=uri_1, federal_action_obligation=-1100, original_loan_subsidy_cost=None, record_type='1') afa_1_row_2 = AwardFinancialAssistanceFactory(uri=uri_1, federal_action_obligation=-10, original_loan_subsidy_cost=None, record_type='1') # original loan subsidy cost used in this row because assistance type is '08' afa_1_row_3 = AwardFinancialAssistanceFactory(uri=uri_1, original_loan_subsidy_cost=-1, assistance_type='08', federal_action_obligation=None, record_type='1') # federal action obligation used in this row (it's 0), because assistance type is not 07 and 08 afa_1_row_4 = AwardFinancialAssistanceFactory(uri=uri_1, original_loan_subsidy_cost=-2222, assistance_type='09', federal_action_obligation=None, record_type='1') # Ignored because record type isn't 1 afa_1_row_5 = AwardFinancialAssistanceFactory(uri=uri_1, federal_action_obligation=-1100, original_loan_subsidy_cost=None, record_type='2') # Uri 2 Test for non-ignored ATA afa_2 = AwardFinancialAssistanceFactory(uri=uri_2, federal_action_obligation=-9999, original_loan_subsidy_cost=None, record_type='1') # Uri 3 test for ignoring a non-matching ATA/AID afa_3 = AwardFinancialAssistanceFactory(uri=uri_3, federal_action_obligation=-9999, record_type='1') # This one matches but will be ignored afa_4 = AwardFinancialAssistanceFactory(uri=uri_4, federal_action_obligation=-9999) errors = number_of_errors(_FILE, database, models=[af_1_row_1, af_1_row_2, af_2_row_1, af_2_row_2, af_3, af_4, afa_1_row_1, afa_1_row_2, afa_1_row_3, afa_1_row_4, afa_1_row_5, afa_2, afa_3, afa_4]) assert errors == 0 def test_failure(database): """ Test failure for each unique URI in File C, the sum of each TransactionObligatedAmount should match (but with opposite signs) the sum of the FederalActionObligation or OriginalLoanSubsidyCost amounts reported in D2. This rule does not apply if the ATA field is populated and is different from the Agency ID. """ # Create a 12 character random uri uri_1 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_2 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_3 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) uri_4 = ''.join(choice(ascii_uppercase + ascii_lowercase + digits) for _ in range(12)) # Simple addition that doesn't add up right af_1_row_1 = AwardFinancialFactory(transaction_obligated_amou=1100, uri=uri_1, allocation_transfer_agency=None) af_1_row_2 = AwardFinancialFactory(transaction_obligated_amou=11, uri=uri_1.lower(), allocation_transfer_agency=None) # Incorrect addition based on assistance type in AFA af_2 = AwardFinancialFactory(transaction_obligated_amou=9999, uri=uri_2, allocation_transfer_agency=None) # Don't ignore when ATA and AID match af_3 = AwardFinancialFactory(transaction_obligated_amou=1100, uri=uri_3, allocation_transfer_agency='good', agency_identifier='good') # Not ignored with TOA of 0 af_4 = AwardFinancialFactory(transaction_obligated_amou=0, uri=uri_4, allocation_transfer_agency='good', agency_identifier='good') # Sum of this uri doesn't add up to af uri sum afa_1_row_1 = AwardFinancialAssistanceFactory(uri=uri_1, federal_action_obligation=-1100, original_loan_subsidy_cost=None, record_type='1') afa_1_row_2 = AwardFinancialAssistanceFactory(uri=uri_1, federal_action_obligation=-10, original_loan_subsidy_cost=None, record_type='1') # Both of these rows use the column that isn't filled in for summing so neither results in the correct number afa_2_row_1 = AwardFinancialAssistanceFactory(uri=uri_2, federal_action_obligation=-9999, original_loan_subsidy_cost=None, record_type='1') afa_2_row_2 = AwardFinancialAssistanceFactory(uri=uri_2, federal_action_obligation=None, original_loan_subsidy_cost=-1000, assistance_type='07', record_type='1') # This shouldn't be ignored afa_3 = AwardFinancialAssistanceFactory(uri=uri_3, federal_action_obligation=0, original_loan_subsidy_cost=None, record_type='1') # This shouldn't be ignored afa_4 = AwardFinancialAssistanceFactory(uri=uri_4, federal_action_obligation=1, original_loan_subsidy_cost=None, record_type='1') errors = number_of_errors(_FILE, database, models=[af_1_row_1, af_1_row_2, af_2, af_3, af_4, afa_1_row_1, afa_1_row_2, afa_2_row_1, afa_2_row_2, afa_3, afa_4]) assert errors == 4
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/c42d328ef2fa_add_frec_code_to_submission_model.py
1
1050
"""add frec_code to submission model Revision ID: c42d328ef2fa Revises: 4d8408c33fee Create Date: 2017-07-10 13:16:56.855163 """ # revision identifiers, used by Alembic. revision = 'c42d328ef2fa' down_revision = '4d8408c33fee' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.add_column('submission', sa.Column('frec_code', sa.Text(), nullable=True)) op.add_column('frec', sa.Column('cgac_code', sa.Text(), nullable=True)) ### end Alembic commands ### def downgrade_data_broker(): op.execute("DELETE FROM submission " "WHERE cgac_code IS NULL") ### commands auto generated by Alembic - please adjust! ### op.drop_column('frec', 'cgac_code') op.drop_column('submission', 'frec_code') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/4a1988f74a78_add_detached_d2_submission_models.py
1
7389
"""add detached D2 submission models Revision ID: 4a1988f74a78 Revises: 4bf29ae16467 Create Date: 2017-01-20 11:40:50.782401 """ # revision identifiers, used by Alembic. revision = '4a1988f74a78' down_revision = '4bf29ae16467' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.create_table('detached_award_financial_assistance', sa.Column('created_at', sa.DateTime(), nullable=True), sa.Column('updated_at', sa.DateTime(), nullable=True), sa.Column('detached_award_financial_assistance_id', sa.Integer(), nullable=False), sa.Column('submission_id', sa.Integer(), nullable=False), sa.Column('job_id', sa.Integer(), nullable=False), sa.Column('row_number', sa.Integer(), nullable=False), sa.Column('action_date', sa.Text(), nullable=True), sa.Column('action_type', sa.Text(), nullable=True), sa.Column('assistance_type', sa.Text(), nullable=True), sa.Column('award_description', sa.Text(), nullable=True), sa.Column('awardee_or_recipient_legal', sa.Text(), nullable=True), sa.Column('awardee_or_recipient_uniqu', sa.Text(), nullable=True), sa.Column('awarding_agency_code', sa.Text(), nullable=True), sa.Column('awarding_agency_name', sa.Text(), nullable=True), sa.Column('awarding_office_code', sa.Text(), nullable=True), sa.Column('awarding_office_name', sa.Text(), nullable=True), sa.Column('awarding_sub_tier_agency_c', sa.Text(), nullable=True), sa.Column('awarding_sub_tier_agency_n', sa.Text(), nullable=True), sa.Column('award_modification_amendme', sa.Text(), nullable=True), sa.Column('business_funds_indicator', sa.Text(), nullable=True), sa.Column('business_types', sa.Text(), nullable=True), sa.Column('cfda_number', sa.Text(), nullable=True), sa.Column('cfda_title', sa.Text(), nullable=True), sa.Column('correction_late_delete_ind', sa.Text(), nullable=True), sa.Column('face_value_loan_guarantee', sa.Numeric(), nullable=True), sa.Column('fain', sa.Text(), nullable=True), sa.Column('federal_action_obligation', sa.Numeric(), nullable=True), sa.Column('fiscal_year_and_quarter_co', sa.Text(), nullable=True), sa.Column('funding_agency_code', sa.Text(), nullable=True), sa.Column('funding_agency_name', sa.Text(), nullable=True), sa.Column('funding_office_name', sa.Text(), nullable=True), sa.Column('funding_office_code', sa.Text(), nullable=True), sa.Column('funding_sub_tier_agency_co', sa.Text(), nullable=True), sa.Column('funding_sub_tier_agency_na', sa.Text(), nullable=True), sa.Column('legal_entity_address_line1', sa.Text(), nullable=True), sa.Column('legal_entity_address_line2', sa.Text(), nullable=True), sa.Column('legal_entity_address_line3', sa.Text(), nullable=True), sa.Column('legal_entity_city_code', sa.Text(), nullable=True), sa.Column('legal_entity_city_name', sa.Text(), nullable=True), sa.Column('legal_entity_congressional', sa.Text(), nullable=True), sa.Column('legal_entity_country_code', sa.Text(), nullable=True), sa.Column('legal_entity_county_code', sa.Text(), nullable=True), sa.Column('legal_entity_county_name', sa.Text(), nullable=True), sa.Column('legal_entity_foreign_city', sa.Text(), nullable=True), sa.Column('legal_entity_foreign_posta', sa.Text(), nullable=True), sa.Column('legal_entity_foreign_provi', sa.Text(), nullable=True), sa.Column('legal_entity_state_code', sa.Text(), nullable=True), sa.Column('legal_entity_state_name', sa.Text(), nullable=True), sa.Column('legal_entity_zip5', sa.Text(), nullable=True), sa.Column('legal_entity_zip_last4', sa.Text(), nullable=True), sa.Column('non_federal_funding_amount', sa.Numeric(), nullable=True), sa.Column('original_loan_subsidy_cost', sa.Numeric(), nullable=True), sa.Column('period_of_performance_curr', sa.Text(), nullable=True), sa.Column('period_of_performance_star', sa.Text(), nullable=True), sa.Column('place_of_performance_city', sa.Text(), nullable=True), sa.Column('place_of_performance_code', sa.Text(), nullable=True), sa.Column('place_of_performance_congr', sa.Text(), nullable=True), sa.Column('place_of_perform_country_c', sa.Text(), nullable=True), sa.Column('place_of_perform_county_na', sa.Text(), nullable=True), sa.Column('place_of_performance_forei', sa.Text(), nullable=True), sa.Column('place_of_perform_state_nam', sa.Text(), nullable=True), sa.Column('place_of_performance_zip4a', sa.Text(), nullable=True), sa.Column('record_type', sa.Integer(), nullable=True), sa.Column('sai_number', sa.Text(), nullable=True), sa.Column('total_funding_amount', sa.Numeric(), nullable=True), sa.Column('uri', sa.Text(), nullable=True), sa.Column('is_valid', sa.Boolean(), server_default='False', nullable=False), sa.PrimaryKeyConstraint('detached_award_financial_assistance_id') ) op.create_index(op.f('ix_detached_award_financial_assistance_fain'), 'detached_award_financial_assistance', ['fain'], unique=False) op.create_index(op.f('ix_detached_award_financial_assistance_job_id'), 'detached_award_financial_assistance', ['job_id'], unique=False) op.create_index(op.f('ix_detached_award_financial_assistance_submission_id'), 'detached_award_financial_assistance', ['submission_id'], unique=False) op.create_index(op.f('ix_detached_award_financial_assistance_uri'), 'detached_award_financial_assistance', ['uri'], unique=False) op.create_table('submission_sub_tier_affiliation', sa.Column('created_at', sa.DateTime(), nullable=True), sa.Column('updated_at', sa.DateTime(), nullable=True), sa.Column('submission_sub_tier_affiliation_id', sa.Integer(), nullable=False), sa.Column('submission_id', sa.Integer(), nullable=True), sa.Column('sub_tier_agency_id', sa.Integer(), nullable=True), sa.ForeignKeyConstraint(['sub_tier_agency_id'], ['sub_tier_agency.sub_tier_agency_id'], name='fk_sub_tier_submission_affiliation_agency_id'), sa.ForeignKeyConstraint(['submission_id'], ['submission.submission_id'], name='fk_submission_sub_tier_affiliation_id'), sa.PrimaryKeyConstraint('submission_sub_tier_affiliation_id') ) op.add_column('submission', sa.Column('d2_submission', sa.Boolean(), server_default='False', nullable=False)) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_column('submission', 'd2_submission') op.drop_table('submission_sub_tier_affiliation') op.drop_index(op.f('ix_detached_award_financial_assistance_uri'), table_name='detached_award_financial_assistance') op.drop_index(op.f('ix_detached_award_financial_assistance_submission_id'), table_name='detached_award_financial_assistance') op.drop_index(op.f('ix_detached_award_financial_assistance_job_id'), table_name='detached_award_financial_assistance') op.drop_index(op.f('ix_detached_award_financial_assistance_fain'), table_name='detached_award_financial_assistance') op.drop_table('detached_award_financial_assistance') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/2ae156c8f46d_update_d1_and_d2_for_daims_v1_1.py
1
4727
"""update d1 and d2 for daims v1.1 Revision ID: 2ae156c8f46d Revises: 4b1ee78268fb Create Date: 2017-08-28 15:16:00.926683 """ # revision identifiers, used by Alembic. revision = '2ae156c8f46d' down_revision = '4b1ee78268fb' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.add_column('award_procurement', sa.Column('award_or_idv_flag', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('place_of_perform_country_n', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('place_of_perform_county_na', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('place_of_perform_state_nam', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('referenced_idv_agency_name', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('referenced_idv_type', sa.Text(), nullable=True)) op.add_column('award_procurement', sa.Column('referenced_multi_or_single', sa.Text(), nullable=True)) op.add_column('detached_award_procurement', sa.Column('place_of_perform_country_n', sa.Text(), nullable=True)) op.add_column('detached_award_procurement', sa.Column('place_of_perform_state_nam', sa.Text(), nullable=True)) op.add_column('detached_award_procurement', sa.Column('referenced_idv_agency_name', sa.Text(), nullable=True)) op.add_column('detached_award_procurement', sa.Column('referenced_multi_or_single', sa.Text(), nullable=True)) op.add_column('detached_award_procurement', sa.Column('award_or_idv_flag', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('legal_entity_country_name', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('place_of_perform_country_n', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('place_of_perform_county_co', sa.Text(), nullable=True)) op.add_column('detached_award_financial_assistance', sa.Column('legal_entity_country_name', sa.Text(), nullable=True)) op.add_column('detached_award_financial_assistance', sa.Column('place_of_perform_country_n', sa.Text(), nullable=True)) op.add_column('detached_award_financial_assistance', sa.Column('place_of_perform_county_co', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('legal_entity_country_name', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('place_of_perform_country_n', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('place_of_perform_county_co', sa.Text(), nullable=True)) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_column('published_award_financial_assistance', 'place_of_perform_county_co') op.drop_column('published_award_financial_assistance', 'place_of_perform_country_n') op.drop_column('published_award_financial_assistance', 'legal_entity_country_name') op.drop_column('detached_award_financial_assistance', 'place_of_perform_county_co') op.drop_column('detached_award_financial_assistance', 'place_of_perform_country_n') op.drop_column('detached_award_financial_assistance', 'legal_entity_country_name') op.drop_column('award_financial_assistance', 'place_of_perform_county_co') op.drop_column('award_financial_assistance', 'place_of_perform_country_n') op.drop_column('award_financial_assistance', 'legal_entity_country_name') op.drop_column('detached_award_procurement', 'referenced_multi_or_single') op.drop_column('detached_award_procurement', 'referenced_idv_agency_name') op.drop_column('detached_award_procurement', 'place_of_perform_state_nam') op.drop_column('detached_award_procurement', 'place_of_perform_country_n') op.drop_column('detached_award_procurement', 'award_or_idv_flag') op.drop_column('award_procurement', 'referenced_multi_or_single') op.drop_column('award_procurement', 'referenced_idv_type') op.drop_column('award_procurement', 'referenced_idv_agency_name') op.drop_column('award_procurement', 'place_of_perform_state_nam') op.drop_column('award_procurement', 'place_of_perform_county_na') op.drop_column('award_procurement', 'place_of_perform_country_n') op.drop_column('award_procurement', 'award_or_idv_flag') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs37_1.py
1
3684
from tests.unit.dataactcore.factories.staging import FABSFactory from dataactcore.models.domainModels import CFDAProgram from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs37_1' def test_column_headers(database): expected_subset = {'row_number', 'cfda_number', 'action_date', 'action_type', 'correction_delete_indicatr', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Test valid. For new (ActionType = A) or mixed aggregate (ActionType = E) assistance awards specifically, the CFDA_Number must be active as of the ActionDate. This does not apply to correction records (those with CorrectionDeleteIndicator = C and delete records). """ cfda = CFDAProgram(program_number=12.340, published_date='20130427', archived_date='') fabs_1 = FABSFactory(cfda_number='12.340', action_date='20140111', action_type='a', correction_delete_indicatr='B') fabs_2 = FABSFactory(cfda_number='12.340', action_date='20140111', action_type='E', correction_delete_indicatr=None) fabs_3 = FABSFactory(cfda_number='12.340', action_date='20130427', action_type='a', correction_delete_indicatr='B') fabs_4 = FABSFactory(cfda_number='12.340', action_date='20110111', action_type='B', correction_delete_indicatr='B') fabs_5 = FABSFactory(cfda_number='12.340', action_date='20110111', action_type='A', correction_delete_indicatr='C') # Ignore correction delete indicator of D fabs_6 = FABSFactory(cfda_number='12.340', action_date='20110111', action_type='e', correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4, fabs_5, fabs_6, cfda]) assert errors == 0 cfda = CFDAProgram(program_number=12.350, published_date='20130427', archived_date='20150427') fabs_1 = FABSFactory(cfda_number='12.350', action_date='20140111', action_type='E', correction_delete_indicatr='B') fabs_2 = FABSFactory(cfda_number='12.350', action_date='20140111', action_type='a', correction_delete_indicatr=None) fabs_3 = FABSFactory(cfda_number='12.350', action_date='20130427', action_type='A', correction_delete_indicatr='B') fabs_4 = FABSFactory(cfda_number='12.350', action_date='20110111', action_type='B', correction_delete_indicatr='B') fabs_5 = FABSFactory(cfda_number='12.350', action_date='20110111', action_type='e', correction_delete_indicatr='C') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4, fabs_5, cfda]) assert errors == 0 def test_failure(database): """ Test invalid. For new (ActionType = A) or mixed aggregate (ActionType = E) assistance awards specifically, the CFDA_Number must be active as of the ActionDate. This does not apply to correction records (those with CorrectionDeleteIndicator = C and delete records). """ cfda = CFDAProgram(program_number=12.340, published_date='20130427', archived_date='') fabs_1 = FABSFactory(cfda_number='12.340', action_date='20120111', action_type='e', correction_delete_indicatr='B') fabs_2 = FABSFactory(cfda_number='12.340', action_date='20120111', action_type='A', correction_delete_indicatr=None) fabs_3 = FABSFactory(cfda_number='12.340', action_date='20120111', action_type='a', correction_delete_indicatr='B') fabs_4 = FABSFactory(cfda_number='12.340', action_date='20120111', action_type='E', correction_delete_indicatr=None) errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4, cfda]) assert errors == 4
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_load_usps_files.py
1
5408
import urllib.parse import pytest import json import re from datetime import datetime from unittest.mock import MagicMock from dataactcore.models.domainModels import ExternalDataLoadDate from dataactcore.models.lookups import EXTERNAL_DATA_TYPE_DICT from dataactvalidator.scripts.load_usps_files import get_payload_string, check_response_status, get_login_tokens, \ get_file_info from tests.unit.mock_helpers import MockSession def mocked_requests_post(url): """ Mocks post request to USPS service. Returns a MagicMock in the structure of a request object. """ if url == 'http://working_url.com': return MagicMock(status_code=200, text=json.dumps({'key': 'value', 'response': 'success'})) return MagicMock(status_code=500, text=json.dumps({'error': 'message', 'messages': 'failure'})) def test_get_pay_load_string(): """ Tests get_payload_string() to ensure payload is properly formatted""" test_dict = {"a": "123", "b": "456"} result = get_payload_string(test_dict) assert re.match(r'^obj={\"[a-b]\":\"\d{3}\",$\n\"[a-b]\":\"\d{3}\"}', urllib.parse.unquote(result), re.M) def test_check_response_status_success(): """ Tests check_response_status() function when a POST request is successful (status 200) """ test_response = check_response_status(mocked_requests_post('http://working_url.com')) assert test_response == {'key': 'value', 'response': 'success'} def test_check_response_failed(): """ Tests check_response_status() function when a POST request is a failure (status 500) """ with pytest.raises(SystemExit) as pytest_wrapped_error: check_response_status(mocked_requests_post('http://not_working_url.com')) assert pytest_wrapped_error.type == SystemExit assert pytest_wrapped_error.value.code == 1 def test_get_login_tokens(): """ Tests the get_login_tokens() function: Creates dictionary of login keys to use to make additional request calls to USPS """ test_data = {'logonkey': '1234', 'tokenkey': 'abc123', 'message': 'Success'} assert get_login_tokens(test_data) == {'logonkey': '1234', 'tokenkey': 'abc123'} def test_get_file_info(): """ Tests the get_file_info() function: returns the fileid for the latest version of the zip4 file to be downloaded """ # Create test external data load date last_load_date = datetime.strptime('1700-01-01', '%Y-%m-%d') test_external_data_load_date = ExternalDataLoadDate(external_data_load_date_id=-1, last_load_date_start=last_load_date, last_load_date_end=last_load_date, external_data_type_id=EXTERNAL_DATA_TYPE_DICT['usps_download']) mock_session = MockSession() mock_session.query('').filter_by()._first = test_external_data_load_date test_data = {"response": "success", "fileList": [{"fileid": "12345", "status": "N", "filepath": "file/path/", "filename": "anotherfile.tar", "fulfilled": "2012-08-15"}, {"fileid": "23456", "status": "N", "filepath": "file/path/", "filename": "somefile.tar", "fulfilled": "2012-10-15"}, {"fileid": "23456", "status": "N", "filepath": "file/path/", "filename": "laterfile.tar", "fulfilled": "2012-09-15"}]} expected_date = datetime.strptime('2012-10-15', '%Y-%m-%d').date() assert get_file_info(mock_session, test_data) == ('23456', expected_date, test_external_data_load_date) def _assert_system_exit(expected_code, f, *args): with pytest.raises(SystemExit) as cm: f(*args) if isinstance(cm.exception, int): assert cm.exception == expected_code else: assert cm.exception.code == expected_code def test_exit_code_3(): # Create test external data load date last_load_date = datetime.strptime('2013-01-01', '%Y-%m-%d') test_external_data_load_date = ExternalDataLoadDate(external_data_load_date_id=-1, last_load_date_start=last_load_date, last_load_date_end=last_load_date, external_data_type_id=EXTERNAL_DATA_TYPE_DICT['usps_download']) mock_session = MockSession() mock_session.query('').filter_by()._first = test_external_data_load_date test_data = {"response": "success", "fileList": [{"fileid": "12345", "status": "N", "filepath": "file/path/", "filename": "anotherfile.tar", "fulfilled": "2012-08-15"}, {"fileid": "23456", "status": "N", "filepath": "file/path/", "filename": "somefile.tar", "fulfilled": "2012-10-15"}, {"fileid": "23456", "status": "N", "filepath": "file/path/", "filename": "laterfile.tar", "fulfilled": "2012-09-15"}]} _assert_system_exit(3, get_file_info, mock_session, test_data)
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/4be5e411246b_adding_principle_place_street_to_.py
1
1093
"""Adding principle place street to subaward Revision ID: 4be5e411246b Revises: 87d7a9b0ea7b Create Date: 2019-08-07 15:13:50.092991 """ # revision identifiers, used by Alembic. revision = '4be5e411246b' down_revision = '87d7a9b0ea7b' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa from sqlalchemy.dialects import postgresql def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('subaward', sa.Column('place_of_perform_street', sa.Text(), nullable=True)) op.add_column('subaward', sa.Column('sub_place_of_perform_street', sa.Text(), nullable=True)) # ### end Alembic commands ### def downgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('subaward', 'sub_place_of_perform_street') op.drop_column('subaward', 'place_of_perform_street') # ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactbroker/routes/file_routes.py
1
14201
from flask import request from webargs import fields as webargs_fields, validate as webargs_validate from webargs.flaskparser import use_kwargs from dataactbroker.handlers.fileHandler import ( FileHandler, get_status, list_submissions as list_submissions_handler, list_published_files as list_published_files_handler, get_upload_file_url, get_detached_upload_file_url, get_submission_comments, submission_report_url, update_submission_comments, list_history, file_history_url, get_comments_file, get_submission_zip) from dataactbroker.handlers.submission_handler import ( delete_all_submission_data, get_submission_stats, list_banners, check_current_submission_page, publish_dabs_submission, certify_dabs_submission, publish_and_certify_dabs_submission, get_published_submission_ids, get_submission_metadata, get_submission_data, get_revalidation_threshold, get_latest_publication_period, revert_to_published) from dataactbroker.decorators import convert_to_submission_id from dataactbroker.permissions import (requires_login, requires_submission_perms, requires_agency_perms, requires_sub_agency_perms) from dataactcore.interfaces.function_bag import get_fabs_meta from dataactcore.models.lookups import FILE_TYPE_DICT, FILE_TYPE_DICT_LETTER from dataactcore.utils.jsonResponse import JsonResponse from dataactcore.utils.requestDictionary import RequestDictionary from dataactcore.utils.statusCode import StatusCode # Add the file submission route def add_file_routes(app, is_local, server_path): """ Create routes related to file submission for flask app """ # Keys for the post route will correspond to the four types of files @app.route("/v1/upload_dabs_files/", methods=["POST"]) @requires_agency_perms('writer') def upload_dabs_files(): if "multipart/form-data" not in request.headers['Content-Type']: return JsonResponse.error(ValueError("Request must be a multipart/form-data type"), StatusCode.CLIENT_ERROR) file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return file_manager.validate_upload_dabs_files() @app.route("/v1/check_status/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') @use_kwargs({'type': webargs_fields.String(missing='')}) def check_status(submission, **kwargs): type = kwargs.get('type') return get_status(submission, type) @app.route("/v1/submission_metadata/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') def submission_metadata(submission): return JsonResponse.create(StatusCode.OK, get_submission_metadata(submission)) @app.route("/v1/submission_data/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') @use_kwargs({'type': webargs_fields.String(missing='')}) def submission_data(submission, **kwargs): type = kwargs.get('type') return get_submission_data(submission, type) @app.route("/v1/revalidation_threshold/", methods=["GET"]) @requires_login def revalidation_threshold(): return JsonResponse.create(StatusCode.OK, get_revalidation_threshold()) @app.route("/v1/latest_publication_period/", methods=["GET"]) @requires_login def latest_publication_period(): return JsonResponse.create(StatusCode.OK, get_latest_publication_period()) @app.route("/v1/list_banners/", methods=["GET"]) @use_kwargs({'login': webargs_fields.Boolean(missing=False)}) def get_banner_list(login): return list_banners(login) @app.route("/v1/list_submissions/", methods=["POST"]) @requires_login @use_kwargs({ 'page': webargs_fields.Int(missing=1), 'limit': webargs_fields.Int(missing=5), 'published': webargs_fields.String( required=True, validate=webargs_validate.OneOf(('mixed', 'true', 'false'))), 'sort': webargs_fields.String(missing='modified'), 'order': webargs_fields.String(missing='desc'), 'fabs': webargs_fields.Bool(missing=False), 'filters': webargs_fields.Dict(keys=webargs_fields.String(), missing={}) }) def list_submissions(published, **kwargs): """ List submission IDs associated with the current user """ page = kwargs.get('page') limit = kwargs.get('limit') sort = kwargs.get('sort') order = kwargs.get('order') fabs = kwargs.get('fabs') filters = kwargs.get('filters') return list_submissions_handler(page, limit, published, sort, order, fabs, filters) @app.route("/v1/list_latest_published_files/", methods=["GET"]) @requires_login @use_kwargs({ 'type': webargs_fields.String( required=True, validate=webargs_validate.OneOf(('fabs', 'dabs'))), 'agency': webargs_fields.String(), 'year': webargs_fields.Int(), 'period': webargs_fields.Int() }) def list_latest_published_files(**kwargs): """ List submission IDs associated with the current user """ sub_type = kwargs.get('type') agency = kwargs.get('agency') year = kwargs.get('year') period = kwargs.get('period') return list_published_files_handler(sub_type, agency, year, period) @app.route("/v1/list_history/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') def submission_list_history(submission): """ List all publish and certify history for a specific submission """ return list_history(submission) @app.route("/v1/get_certified_file/", methods=["GET"]) @use_kwargs({ 'submission_id': webargs_fields.Int(required=True), 'published_files_history_id': webargs_fields.Int(required=True), 'is_warning': webargs_fields.Bool(missing=False) }) @requires_submission_perms('reader') def get_certified_file(submission, published_files_history_id, **kwargs): """ Get the signed URL for the specified file history """ is_warning = kwargs.get('is_warning') return file_history_url(published_files_history_id, is_warning, is_local, submission) @app.route("/v1/get_submitted_published_file/", methods=["GET"]) @use_kwargs({ 'published_files_history_id': webargs_fields.Int(required=True) }) @requires_login def get_submitted_published_file(published_files_history_id): """ Get the signed URL for the specified submitted and published file """ return file_history_url(published_files_history_id, False, is_local) @app.route("/v1/check_current_page/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') def check_current_page(submission): return check_current_submission_page(submission) @app.route("/v1/get_fabs_meta/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') def get_fabs_metadata(submission): """ Return metadata of FABS submission """ return JsonResponse.create(StatusCode.OK, get_fabs_meta(submission.submission_id)) @app.route("/v1/upload_fabs_file/", methods=["POST"]) @requires_sub_agency_perms('editfabs') def upload_fabs_file(): if "multipart/form-data" not in request.headers['Content-Type']: return JsonResponse.error(ValueError("Request must be a multipart/form-data type"), StatusCode.CLIENT_ERROR) params = RequestDictionary.derive(request) fabs = params.get('_files', {}).get('fabs', None) file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return file_manager.upload_fabs_file(fabs) @app.route("/v1/publish_fabs_file/", methods=["POST"]) @convert_to_submission_id @requires_submission_perms('fabs', check_owner=False) def publish_fabs_file(submission): file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return file_manager.publish_fabs_submission(submission) @app.route("/v1/get_obligations/", methods=["GET"]) @convert_to_submission_id @requires_submission_perms('reader') def get_obligations(submission): return JsonResponse.create(StatusCode.OK, get_submission_stats(submission.submission_id)) @app.route("/v1/get_submission_comments/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') def get_sub_comments(submission): return get_submission_comments(submission) @app.route("/v1/update_submission_comments/", methods=['POST']) @convert_to_submission_id @requires_submission_perms('writer') def update_sub_comments(submission): return update_submission_comments(submission, request.json, is_local) @app.route("/v1/get_comments_file/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') def get_submission_comments_file(submission): return get_comments_file(submission, is_local) @app.route("/v1/get_submission_zip/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') @use_kwargs({ 'publish_history_id': webargs_fields.Int(), 'certify_history_id': webargs_fields.Int() }) def get_sub_zip(submission, **kwargs): publish_history_id = kwargs.get('publish_history_id') certify_history_id = kwargs.get('certify_history_id') return get_submission_zip(submission, publish_history_id, certify_history_id, is_local) @app.route("/v1/report_url/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') @use_kwargs({ 'file_type': webargs_fields.String( required=True, validate=webargs_validate.OneOf(FILE_TYPE_DICT.keys() - {'executive_compensation', 'sub_award'}) ), 'warning': webargs_fields.Bool(), 'cross_type': webargs_fields.String(validate=webargs_validate.OneOf(['program_activity', 'award_financial', 'award_procurement', 'award'])) }) def post_submission_report_url(submission, file_type, **kwargs): warning = kwargs.get('warning') cross_type = kwargs.get('cross_type') return submission_report_url(submission, bool(warning), file_type, cross_type) @app.route("/v1/get_file_url/", methods=['GET']) @convert_to_submission_id @requires_submission_perms('reader') @use_kwargs({ 'file_type': webargs_fields.String( required=True, validate=webargs_validate.OneOf(FILE_TYPE_DICT_LETTER.values()) ) }) def get_file_url(submission, file_type): return get_upload_file_url(submission, file_type) @app.route("/v1/get_detached_file_url/", methods=['GET']) @requires_login @use_kwargs({ 'job_id': webargs_fields.Int(required=True) }) def get_detached_file_url(job_id): return get_detached_upload_file_url(job_id) @app.route("/v1/delete_submission/", methods=['POST']) @convert_to_submission_id @requires_submission_perms('writer', check_fabs='editfabs') def delete_submission(submission): """ Deletes all data associated with the specified submission NOTE: THERE IS NO WAY TO UNDO THIS """ return delete_all_submission_data(submission) @app.route("/v1/published_submissions/", methods=["GET"]) @requires_login @use_kwargs({'reporting_fiscal_year': webargs_fields.String(required=True), 'reporting_fiscal_period': webargs_fields.String(required=True), 'cgac_code': webargs_fields.String(), 'frec_code': webargs_fields.String(), 'is_quarter': webargs_fields.Bool()}) def get_published_submissions(reporting_fiscal_year, reporting_fiscal_period, **kwargs): """ Check if cgac (or frec) code, year, and quarter already has a published submission """ cgac_code = kwargs.get('cgac_code') frec_code = kwargs.get('frec_code') is_quarter = kwargs.get('is_quarter') return get_published_submission_ids(cgac_code, frec_code, reporting_fiscal_year, reporting_fiscal_period, is_quarter) @app.route('/v1/publish_dabs_submission/', methods=['POST']) @convert_to_submission_id @requires_submission_perms('submitter', check_owner=False) def publish_dabs_sub(submission): file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return publish_dabs_submission(submission, file_manager) @app.route("/v1/certify_dabs_submission/", methods=['POST']) @convert_to_submission_id @requires_submission_perms('submitter', check_owner=False) def certify_dabs_sub(submission): return certify_dabs_submission(submission) @app.route('/v1/publish_and_certify_dabs_submission/', methods=['POST']) @convert_to_submission_id @requires_submission_perms('submitter', check_owner=False) def publish_and_certify_dabs_sub(submission): file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return publish_and_certify_dabs_submission(submission, file_manager) @app.route("/v1/restart_validation/", methods=['POST']) @convert_to_submission_id @requires_submission_perms('writer', check_fabs='editfabs') @use_kwargs({'is_fabs': webargs_fields.Bool(missing=False)}) def restart_validation(submission, **kwargs): is_fabs = kwargs.get('is_fabs') return FileHandler.restart_validation(submission, is_fabs) @app.route("/v1/revert_submission/", methods=['POST']) @convert_to_submission_id @requires_submission_perms('submitter') def revert_submission(submission): """ Revert an updated DABS submission to the state it was when it was last published """ file_manager = FileHandler(request, is_local=is_local, server_path=server_path) return revert_to_published(submission, file_manager)
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/3b9dffe063a6_create_zips_grouped_table_and_add_.py
1
2745
"""Create zips_grouped table and add multicolumn index to zips table Revision ID: 3b9dffe063a6 Revises: be4dcb9eede6 Create Date: 2020-08-03 11:48:55.068356 """ # revision identifiers, used by Alembic. revision = '3b9dffe063a6' down_revision = 'be4dcb9eede6' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('zips_grouped', sa.Column('created_at', sa.DateTime(), nullable=True), sa.Column('updated_at', sa.DateTime(), nullable=True), sa.Column('zips_grouped_id', sa.Integer(), nullable=False), sa.Column('zip5', sa.Text(), nullable=True), sa.Column('state_abbreviation', sa.Text(), nullable=True), sa.Column('county_number', sa.Text(), nullable=True), sa.Column('congressional_district_no', sa.Text(), nullable=True), sa.PrimaryKeyConstraint('zips_grouped_id') ) op.create_index(op.f('ix_zips_grouped_zip5'), 'zips_grouped', ['zip5'], unique=False) op.create_index('ix_zips_zip5_state_abbreviation_county_number', 'zips', ['zip5', 'state_abbreviation', 'county_number'], unique=False) # Populating the new table op.execute(""" INSERT INTO zips_grouped (zip5, state_abbreviation, county_number) SELECT zip5, state_abbreviation, county_number FROM zips GROUP BY zip5, state_abbreviation, county_number; WITH district_counts AS ( SELECT zip5, COUNT(DISTINCT zips.congressional_district_no) AS cd_count FROM zips GROUP BY zip5) UPDATE zips_grouped SET created_at = NOW(), updated_at = NOW(), congressional_district_no = CASE WHEN cd_count <> 1 THEN '90' END FROM district_counts AS dc WHERE dc.zip5 = zips_grouped.zip5; UPDATE zips_grouped SET congressional_district_no = zips.congressional_district_no FROM zips WHERE zips_grouped.congressional_district_no IS NULL AND zips.zip5 = zips_grouped.zip5; """) # ### end Alembic commands ### def downgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.drop_index('ix_zips_zip5_state_abbreviation_county_number', table_name='zips') op.drop_index(op.f('ix_zips_grouped_zip5'), table_name='zips_grouped') op.drop_table('zips_grouped') # ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs32_2.py
1
1706
from tests.unit.dataactcore.factories.staging import FABSFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs32_2' def test_column_headers(database): expected_subset = {'row_number', 'period_of_performance_star', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ When provided, PeriodOfPerformanceStartDate must be a valid date between 19991001 and 20991231. (i.e., a date between 10/01/1999 and 12/31/2099) """ fabs_1 = FABSFactory(period_of_performance_star='20120725', correction_delete_indicatr='c') fabs_2 = FABSFactory(period_of_performance_star=None, correction_delete_indicatr=None) fabs_3 = FABSFactory(period_of_performance_star='5', correction_delete_indicatr='') fabs_4 = FABSFactory(period_of_performance_star='', correction_delete_indicatr='C') # Ignore correction delete indicator of D fabs_5 = FABSFactory(period_of_performance_star='19990131', correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4, fabs_5]) assert errors == 0 def test_failure(database): """ When provided, PeriodOfPerformanceStartDate must be a valid date between 19991001 and 20991231. (i.e., a date between 10/01/1999 and 12/31/2099) """ fabs_1 = FABSFactory(period_of_performance_star='19990131', correction_delete_indicatr='') fabs_2 = FABSFactory(period_of_performance_star='21000101', correction_delete_indicatr='c') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2]) assert errors == 2
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/1e0b1d3e3cca_d_model_to_text.py
2
3725
"""d-model-to-text Revision ID: 1e0b1d3e3cca Revises: 1ae491ca0925 Create Date: 2016-09-08 09:34:56.153584 """ # revision identifiers, used by Alembic. revision = '1e0b1d3e3cca' down_revision = '1ae491ca0925' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.alter_column('award_financial_assistance', 'face_value_loan_guarantee', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_financial_assistance', 'federal_action_obligation', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_financial_assistance', 'non_federal_funding_amount', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_financial_assistance', 'original_loan_subsidy_cost', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_financial_assistance', 'record_type', existing_type=sa.INTEGER(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_financial_assistance', 'total_funding_amount', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_procurement', 'current_total_value_award', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) op.alter_column('award_procurement', 'potential_total_value_awar', existing_type=sa.NUMERIC(), type_=sa.Text(), existing_nullable=True) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.alter_column('award_procurement', 'potential_total_value_awar', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_procurement', 'current_total_value_award', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_financial_assistance', 'total_funding_amount', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_financial_assistance', 'record_type', existing_type=sa.Text(), type_=sa.INTEGER(), existing_nullable=True) op.alter_column('award_financial_assistance', 'original_loan_subsidy_cost', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_financial_assistance', 'non_federal_funding_amount', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_financial_assistance', 'federal_action_obligation', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) op.alter_column('award_financial_assistance', 'face_value_loan_guarantee', existing_type=sa.Text(), type_=sa.NUMERIC(), existing_nullable=True) ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs13_1.py
1
1673
from tests.unit.dataactcore.factories.staging import FABSFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs13_1' def test_column_headers(database): expected_subset = {'row_number', 'record_type', 'legal_entity_zip5', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Test LegalEntityZIP5 must be blank for aggregate records (i.e., when RecordType = 1) record types 2 and 3 don't affect success. """ fabs = FABSFactory(record_type=2, legal_entity_zip5='12345', correction_delete_indicatr='') fabs_2 = FABSFactory(record_type=2, legal_entity_zip5=None, correction_delete_indicatr=None) fabs_3 = FABSFactory(record_type=3, legal_entity_zip5='12345', correction_delete_indicatr=None) fabs_null = FABSFactory(record_type=1, legal_entity_zip5=None, correction_delete_indicatr='c') fabs_null_2 = FABSFactory(record_type=1, legal_entity_zip5='', correction_delete_indicatr='C') # Ignore correction delete indicator of D fabs_4 = FABSFactory(record_type=1, legal_entity_zip5='Test', correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[fabs, fabs_2, fabs_null, fabs_null_2, fabs_3, fabs_4]) assert errors == 0 def test_failure(database): """ Test failure when LegalEntityZIP5 isn't blank for aggregate records (i.e., when RecordType = 1) """ fabs = FABSFactory(record_type=1, legal_entity_zip5='Test', correction_delete_indicatr='') errors = number_of_errors(_FILE, database, models=[fabs]) assert errors == 1
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactbroker/helpers/filters_helper.py
1
6379
from sqlalchemy import or_, and_ from flask import g from dataactcore.models.lookups import FILE_TYPE_DICT_LETTER_ID, RULE_SEVERITY_DICT from dataactcore.models.domainModels import CGAC, FREC from dataactcore.models.errorModels import PublishedErrorMetadata, ErrorMetadata from dataactcore.models.jobModels import Submission from dataactcore.models.validationModels import RuleSql, RuleSetting from dataactcore.utils.responseException import ResponseException from dataactcore.utils.statusCode import StatusCode def agency_filter(sess, query, cgac_model, frec_model, agency_list): """ Given the provided query, add a filter by agencies provided the agency list. Note that this does not include the additional permissions filter listed in this file. Arguments: sess: the database connection query: the sqlalchemy query to apply the filters to cgac_model: the model to apply the cgacs filter frec_model: the model to apply the frecs filter agency_list: list of strings representing the agency codes to filter with Raises: ResponseException: if any of the strings in the agency_list are invalid Returns: the same queryset provided with agency filters included """ agency_filters = [] cgac_codes = [cgac_code for cgac_code in agency_list if isinstance(cgac_code, str) and len(cgac_code) == 3] frec_codes = [frec_code for frec_code in agency_list if isinstance(frec_code, str) and len(frec_code) == 4] if len(cgac_codes) + len(frec_codes) != len(agency_list): raise ResponseException('All codes in the agency_codes filter must be valid agency codes', StatusCode.CLIENT_ERROR) # If the number of CGACs or FRECs returned from a query using the codes doesn't match the length of # each list (ignoring duplicates) then something included wasn't a valid agency cgac_list = set(cgac_codes) frec_list = set(frec_codes) if (cgac_list and sess.query(CGAC).filter(CGAC.cgac_code.in_(cgac_list)).count() != len(cgac_list)) or \ (frec_list and sess.query(FREC).filter(FREC.frec_code.in_(frec_list)).count() != len(frec_list)): raise ResponseException("All codes in the agency_codes filter must be valid agency codes", StatusCode.CLIENT_ERROR) if len(cgac_list) > 0: agency_filters.append(cgac_model.cgac_code.in_(cgac_list)) if len(frec_list) > 0: agency_filters.append(frec_model.frec_code.in_(frec_list)) return query.filter(or_(*agency_filters)) def permissions_filter(query): """ Given the provided query, add a filter to only include agencies the user has access to. Arguments: query: the sqlalchemy query to apply the filters to Returns: the same queryset provided with permissions filter included """ if not g.user.website_admin: affiliation_filters = [] cgac_codes = [aff.cgac.cgac_code for aff in g.user.affiliations if aff.cgac] frec_codes = [aff.frec.frec_code for aff in g.user.affiliations if aff.frec] affiliation_filters.append(Submission.user_id == g.user.user_id) if cgac_codes: affiliation_filters.append(Submission.cgac_code.in_(cgac_codes)) if frec_codes: affiliation_filters.append(Submission.frec_code.in_(frec_codes)) query = query.filter(or_(*affiliation_filters)) return query def file_filter(query, file_model, files): """ Given the provided query, add a filter by files provided the files list. Arguments: query: the sqlalchemy query to apply the filters to file_model: the model to apply the file filter files: list of files representing the agency codes to filter with Returns: the same queryset provided with file filters included """ model_file_type_id = { PublishedErrorMetadata: 'file_type_id', ErrorMetadata: 'file_type_id', RuleSql: 'file_id', RuleSetting: 'file_id' } if file_model not in model_file_type_id: valid_file_models = [model_file_type.__name__ for model_file_type in model_file_type_id.keys()] error_message = 'Invalid file model. Use one of the following instead: {}.' raise ResponseException(error_message.format(', '.join(sorted(valid_file_models)))) file_type_filters = [] if files: for file_type in files: file_id = getattr(file_model, model_file_type_id[file_model]) target_file_id = getattr(file_model, 'target_{}'.format(model_file_type_id[file_model])) if file_type in ['A', 'B', 'C']: file_type_filters.append(and_(file_id == FILE_TYPE_DICT_LETTER_ID[file_type], target_file_id.is_(None))) else: file_types = file_type.split('-')[1] # Append both orders of the source/target files to the list file_type_filters.append(and_(file_id == FILE_TYPE_DICT_LETTER_ID[file_types[:1]], target_file_id == FILE_TYPE_DICT_LETTER_ID[file_types[1:]])) file_type_filters.append(and_(file_id == FILE_TYPE_DICT_LETTER_ID[file_types[1:]], target_file_id == FILE_TYPE_DICT_LETTER_ID[file_types[:1]])) return query.filter(or_(*file_type_filters)) def rule_severity_filter(query, error_level, error_model=ErrorMetadata): """ Given the provided query, add a filter by files provided the files list. Arguments: query: the sqlalchemy query to apply the filters to error_level: the error level to filter on (could be 'error' or 'warning') error_model: the model to apply the filter to (must have a severity_id field) Returns: the same queryset provided with rule severity filter included """ # If the error level isn't "mixed" add a filter on which severity to pull if error_level == 'error': query = query.filter(error_model.severity_id == RULE_SEVERITY_DICT['fatal']) elif error_level == 'warning': query = query.filter(error_model.severity_id == RULE_SEVERITY_DICT['warning']) return query
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_a22_appropriations.py
1
2209
from dataactcore.models.stagingModels import Appropriation from dataactcore.models.domainModels import SF133 from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'a22_appropriations' def test_column_headers(database): expected_subset = {'uniqueid_TAS', 'row_number', 'obligations_incurred_total_cpe', 'expected_value_GTAS SF133 Line 2190', 'difference'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Tests that SF 133 amount sum for line 2190 matches Appropriation obligations_incurred_total_cpe for the specified fiscal year and period """ tas = 'tas_one_line' sf = SF133(line=2190, tas=tas, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000') ap = Appropriation(job_id=1, row_number=1, tas=tas, obligations_incurred_total_cpe=1) assert number_of_errors(_FILE, database, models=[sf, ap]) == 0 # Test with split SF133 lines tas = 'tas_two_lines' sf_1 = SF133(line=2190, tas=tas, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000', disaster_emergency_fund_code='n') sf_2 = SF133(line=2190, tas=tas, period=1, fiscal_year=2016, amount=4, agency_identifier='sys', main_account_code='000', sub_account_code='000', disaster_emergency_fund_code='o') ap = Appropriation(job_id=1, row_number=1, tas=tas, obligations_incurred_total_cpe=5) assert number_of_errors(_FILE, database, models=[sf_1, sf_2, ap]) == 0 def test_failure(database): """ Tests that SF 133 amount sum for line 2190 does not match Appropriation obligations_incurred_total_cpe for the specified fiscal year and period """ tas = 'fail_tas' sf = SF133(line=2190, tas=tas, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000') ap = Appropriation(job_id=1, row_number=1, tas=tas, obligations_incurred_total_cpe=0) assert number_of_errors(_FILE, database, models=[sf, ap]) == 1
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_b6_object_class_program_activity_1.py
1
1389
from tests.unit.dataactcore.factories.staging import ObjectClassProgramActivityFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'b6_object_class_program_activity_1' def test_column_headers(database): expected_subset = {'row_number', 'gross_outlays_undelivered_fyb', 'ussgl480200_undelivered_or_fyb', 'difference', 'uniqueid_TAS', 'uniqueid_DisasterEmergencyFundCode', 'uniqueid_ProgramActivityCode', 'uniqueid_ProgramActivityName', 'uniqueid_ObjectClass', 'uniqueid_ByDirectReimbursableFundingSource'} actual = set(query_columns(_FILE, database)) assert (actual & expected_subset) == expected_subset def test_success(database): """ Test Object Class Program Activity gross_outlays_undelivered_fyb equals ussgl480200_undelivered_or_fyb """ op = ObjectClassProgramActivityFactory(gross_outlays_undelivered_fyb=1, ussgl480200_undelivered_or_fyb=1) assert number_of_errors(_FILE, database, models=[op]) == 0 def test_failure(database): """ Test Object Class Program Activity gross_outlays_undelivered_fyb doesnt' equal ussgl480200_undelivered_or_fyb """ op = ObjectClassProgramActivityFactory(gross_outlays_undelivered_fyb=1, ussgl480200_undelivered_or_fyb=0) assert number_of_errors(_FILE, database, models=[op]) == 1
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/a98cd1871ea9_remove_user_status_id_from_user.py
1
1073
"""remove user_status_id from user Revision ID: a98cd1871ea9 Revises: e97127c44797 Create Date: 2016-12-22 11:59:35.173573 """ # revision identifiers, used by Alembic. revision = 'a98cd1871ea9' down_revision = 'e97127c44797' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_constraint('users_user_status_id_fkey', 'users', type_='foreignkey') op.drop_column('users', 'user_status_id') ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('user_status_id', sa.INTEGER(), autoincrement=False, nullable=True)) op.create_foreign_key('users_user_status_id_fkey', 'users', 'user_status', ['user_status_id'], ['user_status_id']) ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_script_read_zips.py
1
6618
from io import StringIO from dataactvalidator.scripts.read_zips import update_state_congr_table_census, group_zips from dataactcore.models.domainModels import StateCongressional, Zips, ZipsGrouped def test_parse_census_district_file(database): census_file_mock = StringIO("""state_code,congressional_district_no,census_year\nA,1,20""") sess = database.session update_state_congr_table_census(census_file_mock, sess) state_congressional = sess.query(StateCongressional).one_or_none() assert state_congressional.state_code == 'A' assert state_congressional.congressional_district_no == '01' assert state_congressional.census_year == 20 def test_group_zips(database): """ Testing the grouping of zips. """ sess = database.session # Only difference is the zip_last4, these will be merged together zip_same1 = Zips(zip5='12345', zip_last4='6789', state_abbreviation='VA', county_number='000', congressional_district_no='01') zip_same2 = Zips(zip5='12345', zip_last4='6780', state_abbreviation='VA', county_number='000', congressional_district_no='01') # Different states, same everything else zip_state1 = Zips(zip5='54321', zip_last4='6789', state_abbreviation='VA', county_number='000', congressional_district_no='01') zip_state2 = Zips(zip5='54321', zip_last4='6780', state_abbreviation='WA', county_number='000', congressional_district_no='01') # Different county codes, same everything else zip_county1 = Zips(zip5='11111', zip_last4='1111', state_abbreviation='VA', county_number='000', congressional_district_no='01') zip_county2 = Zips(zip5='11111', zip_last4='1112', state_abbreviation='VA', county_number='001', congressional_district_no='01') # Everything matches except for congressional district zip_cd1 = Zips(zip5='22222', zip_last4='2222', state_abbreviation='VA', county_number='000', congressional_district_no='01') zip_cd2 = Zips(zip5='22222', zip_last4='2223', state_abbreviation='VA', county_number='000', congressional_district_no='02') # Different states, different congressional district zip_state_cd1 = Zips(zip5='33333', zip_last4='3333', state_abbreviation='VA', county_number='000', congressional_district_no='01') zip_state_cd2 = Zips(zip5='33333', zip_last4='3334', state_abbreviation='WA', county_number='000', congressional_district_no='02') # Null congressional district zip_null_cd = Zips(zip5='44444', zip_last4='4444', state_abbreviation='WA', county_number='000', congressional_district_no=None) sess.add_all([zip_same1, zip_same2, zip_state1, zip_state2, zip_county1, zip_county2, zip_cd1, zip_cd2, zip_state_cd1, zip_state_cd2, zip_null_cd]) sess.commit() # Creating the temp tables to use for testing sess.execute(""" CREATE TABLE temp_zips AS SELECT * FROM zips; CREATE TABLE temp_zips_grouped (LIKE zips_grouped INCLUDING ALL); """) sess.commit() group_zips(sess) # Moving into zips_grouped for easier parsing sess.execute(""" INSERT INTO zips_grouped SELECT * FROM temp_zips_grouped """) sess.commit() # Combined first set of zips zips = sess.query(ZipsGrouped).filter_by(zip5=zip_same1.zip5).all() assert len(zips) == 1 assert zips[0].zip5 == zip_same1.zip5 assert zips[0].state_abbreviation == zip_same1.state_abbreviation assert zips[0].county_number == zip_same1.county_number assert zips[0].congressional_district_no == zip_same1.congressional_district_no # Different states, same everything else zips = sess.query(ZipsGrouped).filter_by(zip5=zip_state1.zip5).order_by(ZipsGrouped.state_abbreviation).all() assert len(zips) == 2 assert zips[0].zip5 == zip_state1.zip5 assert zips[0].state_abbreviation == zip_state1.state_abbreviation assert zips[0].county_number == zip_state1.county_number assert zips[0].congressional_district_no == zip_state1.congressional_district_no assert zips[1].zip5 == zip_state2.zip5 assert zips[1].state_abbreviation == zip_state2.state_abbreviation assert zips[1].county_number == zip_state2.county_number assert zips[1].congressional_district_no == zip_state2.congressional_district_no # Different counties, same everything else zips = sess.query(ZipsGrouped).filter_by(zip5=zip_county1.zip5).order_by(ZipsGrouped.county_number).all() assert len(zips) == 2 assert zips[0].zip5 == zip_county1.zip5 assert zips[0].state_abbreviation == zip_county1.state_abbreviation assert zips[0].county_number == zip_county1.county_number assert zips[0].congressional_district_no == zip_county1.congressional_district_no assert zips[1].zip5 == zip_county2.zip5 assert zips[1].state_abbreviation == zip_county2.state_abbreviation assert zips[1].county_number == zip_county2.county_number assert zips[1].congressional_district_no == zip_county2.congressional_district_no # Different congressional districts zips = sess.query(ZipsGrouped).filter_by(zip5=zip_cd1.zip5).all() assert len(zips) == 1 assert zips[0].zip5 == zip_cd1.zip5 assert zips[0].state_abbreviation == zip_cd1.state_abbreviation assert zips[0].county_number == zip_cd1.county_number assert zips[0].congressional_district_no == '90' # Different states, different congressional districts zips = sess.query(ZipsGrouped).filter_by(zip5=zip_state_cd1.zip5).order_by(ZipsGrouped.state_abbreviation).all() assert len(zips) == 2 assert zips[0].zip5 == zip_state_cd1.zip5 assert zips[0].state_abbreviation == zip_state_cd1.state_abbreviation assert zips[0].county_number == zip_state_cd1.county_number assert zips[0].congressional_district_no == '90' assert zips[1].zip5 == zip_state_cd2.zip5 assert zips[1].state_abbreviation == zip_state_cd2.state_abbreviation assert zips[1].county_number == zip_state_cd2.county_number assert zips[1].congressional_district_no == '90' # Null congressional district zips = sess.query(ZipsGrouped).filter_by(zip5=zip_null_cd.zip5).all() assert len(zips) == 1 assert zips[0].zip5 == zip_null_cd.zip5 assert zips[0].state_abbreviation == zip_null_cd.state_abbreviation assert zips[0].county_number == zip_null_cd.county_number assert zips[0].congressional_district_no == '90'
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/c4d42e86c655_add_descriptions_to_published_.py
1
3105
"""Add descriptions to (Published)AwardFinancialAssistance Revision ID: c4d42e86c655 Revises: 668d9fa93acb Create Date: 2018-04-04 11:00:18.103961 """ # revision identifiers, used by Alembic. revision = 'c4d42e86c655' down_revision = '668d9fa93acb' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.add_column('award_financial_assistance', sa.Column('action_type_description', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('assistance_type_desc', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('business_funds_ind_desc', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('business_types_desc', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('correction_delete_ind_desc', sa.Text(), nullable=True)) op.add_column('award_financial_assistance', sa.Column('record_type_description', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('action_type_description', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('assistance_type_desc', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('business_funds_ind_desc', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('business_types_desc', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('correction_delete_ind_desc', sa.Text(), nullable=True)) op.add_column('published_award_financial_assistance', sa.Column('record_type_description', sa.Text(), nullable=True)) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_column('published_award_financial_assistance', 'record_type_description') op.drop_column('published_award_financial_assistance', 'correction_delete_ind_desc') op.drop_column('published_award_financial_assistance', 'business_types_desc') op.drop_column('published_award_financial_assistance', 'business_funds_ind_desc') op.drop_column('published_award_financial_assistance', 'assistance_type_desc') op.drop_column('published_award_financial_assistance', 'action_type_description') op.drop_column('award_financial_assistance', 'record_type_description') op.drop_column('award_financial_assistance', 'correction_delete_ind_desc') op.drop_column('award_financial_assistance', 'business_types_desc') op.drop_column('award_financial_assistance', 'business_funds_ind_desc') op.drop_column('award_financial_assistance', 'assistance_type_desc') op.drop_column('award_financial_assistance', 'action_type_description') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/scripts/backfill_ppop_scope_fabs.py
1
2049
import logging from dataactcore.interfaces.db import GlobalDB from dataactcore.broker_logging import configure_logging from dataactvalidator.health_check import create_app logger = logging.getLogger(__name__) BACKFILL_FABS_PPOP_SCOPE_SQL_1 = """ UPDATE published_fabs SET place_of_performance_scope = CASE WHEN place_of_performance_code ~ '^00\*{5}$' THEN 'Multi-state' WHEN place_of_performance_code ~ '^[a-zA-Z]{2}\*{5}$' THEN 'State-wide' WHEN place_of_performance_code ~ '^[a-zA-Z]{2}\*\*\d{3}$' THEN 'County-wide' WHEN UPPER(place_of_performance_code) = '00FORGN' THEN 'Foreign' WHEN place_of_performance_code ~ '^[a-zA-Z]{2}\d{4}[\dRr]$' THEN 'City-wide' END WHERE (place_of_performance_zip4a IS NULL AND place_of_performance_scope IS NULL); """ BACKFILL_FABS_PPOP_SCOPE_SQL_2 = """ UPDATE published_fabs SET place_of_performance_scope = CASE WHEN LOWER(place_of_performance_zip4a) = 'city-wide' THEN 'City-wide' WHEN place_of_performance_zip4a ~ '^\d{5}(-?\d{4})?$' THEN 'Single ZIP Code' END WHERE (place_of_performance_code ~ '^[a-zA-Z]{2}\d{4}[\dRr]$' AND place_of_performance_scope IS NULL); """ if __name__ == '__main__': configure_logging() with create_app().app_context(): sess = GlobalDB.db().session affected = 0 logger.info('Backfilling empty place_of_performance_scope values in the fabs table (part i).') executed = sess.execute(BACKFILL_FABS_PPOP_SCOPE_SQL_1) affected += executed.rowcount sess.commit() logger.info('Backfilling empty place_of_performance_scope values in the fabs table (part ii).') executed = sess.execute(BACKFILL_FABS_PPOP_SCOPE_SQL_2) affected += executed.rowcount sess.commit() logger.info('Backfill completed, {} rows affected\n'.format(affected)) sess.close()
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs34_2.py
1
1873
from tests.unit.dataactcore.factories.staging import FABSFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs34_2' def test_column_headers(database): expected_subset = {'row_number', 'period_of_performance_star', 'period_of_performance_curr', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ While they are optional fields, if either PeriodOfPerformanceCurrentEndDate or PeriodOfPerformanceStartDate is provided, both fields must be provided. """ fabs_1 = FABSFactory(period_of_performance_star=None, period_of_performance_curr=None, correction_delete_indicatr='') fabs_2 = FABSFactory(period_of_performance_star='20120724', period_of_performance_curr='20120724', correction_delete_indicatr='c') # Ignore correction delete indicator of D fabs_3 = FABSFactory(period_of_performance_star=None, period_of_performance_curr='20120724', correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3]) assert errors == 0 def test_failure(database): """ While they are optional fields, if either PeriodOfPerformanceCurrentEndDate or PeriodOfPerformanceStartDate is provided, both fields must be provided. """ fabs_1 = FABSFactory(period_of_performance_star=None, period_of_performance_curr='20120724', correction_delete_indicatr='') fabs_2 = FABSFactory(period_of_performance_star='20120725', period_of_performance_curr=None, correction_delete_indicatr='C') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2]) assert errors == 2
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/scripts/populate_published_comments.py
1
2470
import logging from sqlalchemy import func from dataactcore.interfaces.db import GlobalDB from dataactcore.broker_logging import configure_logging from dataactcore.models.jobModels import Submission, PublishHistory, Comment, PublishedComment from dataactvalidator.health_check import create_app from dataactcore.models.userModel import User # noqa logger = logging.getLogger(__name__) if __name__ == '__main__': sess = GlobalDB.db().session configure_logging() with create_app().app_context(): logger.info('Moving published comments') # Create a CTE of the max publish history date for DABS submissions max_publish_history = sess.query(func.max(PublishHistory.updated_at).label('max_updated_at'), PublishHistory.submission_id.label('submission_id')). \ join(Submission, PublishHistory.submission_id == Submission.submission_id). \ filter(Submission.is_fabs.is_(False)). \ group_by(PublishHistory.submission_id).cte('max_publish_history') # Get all comments that were written before the latest publication for all published/updated submissions publish_history_list = sess.query(Comment). \ join(max_publish_history, max_publish_history.c.submission_id == Comment.submission_id). \ filter(Comment.updated_at < max_publish_history.c.max_updated_at).\ order_by(Comment.submission_id, Comment.file_type_id).all() # Create a list of comments and a list of all submissions involved comments_list = [] submissions_list = [] for obj in publish_history_list: tmp_obj = obj.__dict__ tmp_obj.pop('_sa_instance_state') tmp_obj.pop('created_at') tmp_obj.pop('updated_at') tmp_obj.pop('comment_id') comments_list.append(tmp_obj) if tmp_obj['submission_id'] not in submissions_list: submissions_list.append(tmp_obj['submission_id']) # Delete all comments from the submissions we're inserting for sess.query(PublishedComment).filter(PublishedComment.submission_id.in_(submissions_list)).\ delete(synchronize_session=False) # Save all the objects in the published comment table sess.bulk_save_objects([PublishedComment(**comment) for comment in comments_list]) sess.commit() logger.info('Published comments moved')
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs38_2_1.py
1
2802
from tests.unit.dataactcore.factories.domain import OfficeFactory from tests.unit.dataactcore.factories.staging import FABSFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs38_2_1' def test_column_headers(database): expected_subset = {'row_number', 'funding_office_code', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Test when provided, FundingOfficeCode must be a valid value from the Federal Hierarchy, including being designated specifically as a Funding Office in the hierarchy. """ office_1 = OfficeFactory(office_code='12345a', contract_funding_office=True, financial_assistance_funding_office=False) office_2 = OfficeFactory(office_code='12345b', contract_funding_office=False, financial_assistance_funding_office=True) fabs_1 = FABSFactory(funding_office_code='12345a', correction_delete_indicatr='') # test case insensitive fabs_2 = FABSFactory(funding_office_code='12345A', correction_delete_indicatr='c') fabs_3 = FABSFactory(funding_office_code='', correction_delete_indicatr=None) fabs_4 = FABSFactory(funding_office_code=None, correction_delete_indicatr='C') # Testing second type of funding office fabs_5 = FABSFactory(funding_office_code='12345b', correction_delete_indicatr='') # Ignore correction delete indicator of D fabs_6 = FABSFactory(funding_office_code='1234567', correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[office_1, office_2, fabs_1, fabs_2, fabs_3, fabs_4, fabs_5, fabs_6]) assert errors == 0 def test_failure(database): """ Test fail when provided, FundingOfficeCode must be a valid value from the Federal Hierarchy, including being designated specifically as a Funding Office in the hierarchy. """ office_1 = OfficeFactory(office_code='123456', contract_funding_office=True, financial_assistance_funding_office=True) office_2 = OfficeFactory(office_code='987654', contract_funding_office=False, financial_assistance_funding_office=False) fabs_1 = FABSFactory(funding_office_code='12345', correction_delete_indicatr=None) fabs_2 = FABSFactory(funding_office_code='1234567', correction_delete_indicatr='') # Test fail if funding office is false even if code matches fabs_3 = FABSFactory(funding_office_code='987654', correction_delete_indicatr='c') errors = number_of_errors(_FILE, database, models=[office_1, office_2, fabs_1, fabs_2, fabs_3]) assert errors == 3
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/models/lookups.py
1
22140
# This file defines a series of constants that represent the values used in the # broker's "helper" tables. Rather than define the values in the db setup scripts # and then make db calls to lookup the surrogate keys, we'll define everything # here, in a file that can be used by the db setup scripts *and* the application # code. from collections import namedtuple from dataactcore.models.stagingModels import (AwardFinancialAssistance, AwardFinancial, Appropriation, ObjectClassProgramActivity, AwardProcurement, FABS) LookupType = namedtuple('LookupType', ['id', 'name', 'desc']) LookupFileType = namedtuple('LookupFileType', ['id', 'name', 'desc', 'letter', 'order', 'crossfile', 'model']) FILE_STATUS = [ LookupType(1, 'complete', 'File has been processed'), LookupType(2, 'header_error', 'The file has errors in the header row'), LookupType(3, 'unknown_error', 'An unknown error has occurred with this file'), LookupType(4, 'single_row_error', 'Error occurred in job manager'), LookupType(5, 'job_error', 'File has not yet been validated'), LookupType(6, 'incomplete', 'File has not yet been validated'), LookupType(7, 'encoding_error', 'File contains invalid characters that could not be validated'), LookupType(8, 'row_count_error', 'Raw file row count does not match the number of rows validated'), LookupType(9, 'file_type_error', 'Invalid file type. Valid file types include .csv and .txt') ] FILE_STATUS_DICT = {item.name: item.id for item in FILE_STATUS} FILE_STATUS_DICT_ID = {item.id: item.name for item in FILE_STATUS} ERROR_TYPE = [ LookupType(1, 'type_error', 'The value provided was of the wrong type. Note that all type errors in a line' ' must be fixed before the rest of the validation logic is applied to that line.'), LookupType(2, 'required_error', 'This field is required for all submissions but was not provided in this row.'), LookupType(3, 'value_error', 'The value provided was invalid.'), LookupType(4, 'read_error', 'Could not parse this record correctly.'), LookupType(5, 'write_error', 'Could not write this record into the staging table.'), LookupType(6, 'rule_failed', 'A rule failed for this value.'), LookupType(7, 'length_error', 'Value was longer than allowed length.'), LookupType(8, 'field_format_error', 'Date should follow the YYYYMMDD format.'), LookupType(9, 'blank_file_error', 'File does not contain data. For files A and B, this must be addressed prior to' ' publication/certification. Blank file C does not prevent' ' publication/certification.') ] ERROR_TYPE_DICT = {item.name: item.id for item in ERROR_TYPE} JOB_STATUS = [ LookupType(1, 'waiting', 'check dependency table'), LookupType(2, 'ready', 'can be assigned'), LookupType(3, 'running', 'job is currently in progress'), LookupType(4, 'finished', 'job is complete'), LookupType(5, 'invalid', 'job is invalid'), LookupType(6, 'failed', 'job failed to complete') ] JOB_STATUS_DICT = {item.name: item.id for item in JOB_STATUS} JOB_STATUS_DICT_ID = {item.id: item.name for item in JOB_STATUS} JOB_TYPE = [ LookupType(1, 'file_upload', 'file must be uploaded to S3'), LookupType(2, 'csv_record_validation', 'do record level validation and add to staging table'), LookupType(4, 'validation', 'new information must be validated') ] JOB_TYPE_DICT = {item.name: item.id for item in JOB_TYPE} JOB_TYPE_DICT_ID = {item.id: item.name for item in JOB_TYPE} PUBLISH_STATUS = [ LookupType(1, 'unpublished', 'Has not yet been moved to data store'), LookupType(2, 'published', 'Has been moved to data store'), LookupType(3, 'updated', 'Submission was updated after being published'), LookupType(4, 'publishing', 'Submission is being published'), LookupType(5, 'reverting', 'Submission is being reverted to certified status') ] PUBLISH_STATUS_DICT = {item.name: item.id for item in PUBLISH_STATUS} PUBLISH_STATUS_DICT_ID = {item.id: item.name for item in PUBLISH_STATUS} FILE_TYPE = [ LookupFileType(1, 'appropriations', '', 'A', 1, True, Appropriation), LookupFileType(2, 'program_activity', '', 'B', 2, True, ObjectClassProgramActivity), LookupFileType(3, 'award_financial', '', 'C', 3, True, AwardFinancial), LookupFileType(4, 'award', '', 'D2', 4, True, AwardFinancialAssistance), LookupFileType(5, 'award_procurement', '', 'D1', 5, True, AwardProcurement), LookupFileType(6, 'executive_compensation', '', 'E', None, False, None), LookupFileType(7, 'sub_award', '', 'F', None, False, None), LookupFileType(8, 'fabs', '', 'FABS', None, False, FABS) ] FILE_TYPE_DICT = {item.name: item.id for item in FILE_TYPE} FILE_TYPE_DICT_ID = {item.id: item.name for item in FILE_TYPE} FILE_TYPE_DICT_LETTER = {item.id: item.letter for item in FILE_TYPE} FILE_TYPE_DICT_LETTER_ID = {item.letter: item.id for item in FILE_TYPE} FILE_TYPE_DICT_LETTER_NAME = {item.letter: item.name for item in FILE_TYPE} FILE_TYPE_DICT_NAME_LETTER = {item.name: item.letter for item in FILE_TYPE} SUBMISSION_FILENAMES = { 'A': 'SubID-{submission_id}_File-A_{FYP}_{raw_filename}_{timestamp}{ext}', 'B': 'SubID-{submission_id}_File-B_{FYP}_{raw_filename}_{timestamp}{ext}', 'C': 'SubID-{submission_id}_File-C_{FYP}_{raw_filename}_{timestamp}{ext}', 'D1': 'SubID-{submission_id}_File-D1_{FYP}_{start}_{end}_{agency_type}_{timestamp}{ext}', 'D2': 'SubID-{submission_id}_File-D2_{FYP}_{start}_{end}_{agency_type}_{timestamp}{ext}', 'E': 'SubID-{submission_id}_File-E_{FYP}_{timestamp}.csv', 'F': 'SubID-{submission_id}_File-F_{FYP}_{timestamp}.csv', 'FABS': 'SubID-{submission_id}_File-FABS_{raw_filename}_{timestamp}{ext}', 'FABS_publish': 'SubID-{submission_id}_Published-FABS_{timestamp}.csv', } DETACHED_FILENAMES = { 'A': 'File-A_{FYP}_{timestamp}.csv', 'D1': 'File-D1_{start}_{end}_{agency_type}_{timestamp}{ext}', 'D2': 'File-D2_{start}_{end}_{agency_type}_{timestamp}{ext}' } REPORT_FILENAMES = { 'PRE-DAIMS 2.0': { 'file': 'submission_{submission_id}_{file_type}_{report_type}_report.csv', 'cross-file': 'submission_{submission_id}_cross_{report_type}{file_type}_{cross_type}.csv' }, 'DAIMS 2.0': { 'file': 'submission_{submission_id}_File_{file_letter}_{file_type}_{report_type}report.csv', 'cross-file': 'submission_{submission_id}_crossfile_{report_type}' 'File_{file_letter}_to_{cross_letter}_{file_type}_{cross_type}.csv', }, 'DEV-8325': { 'file': 'SubID-{submission_id}_File-{file_letter}-{report_type}-report{FYP}.csv', 'cross-file': 'SubID-{submission_id}_File-{file_letter}-to-{cross_letter}-crossfile-{report_type}-report' '{FYP}.csv', } } PERMISSION_TYPES = [ LookupType(1, 'reader', 'This user is allowed to view any submission for their agency'), LookupType(2, 'writer', 'This user is allowed to create and edit any submission for their agency'), LookupType(3, 'submitter', 'This user is allowed to certify and submit any submission for their agency'), LookupType(4, 'editfabs', 'This user is allowed to create and edit any FABS data for their agency'), LookupType(5, 'fabs', 'This user is allowed to publish any FABS data for their agency') ] PERMISSION_TYPE_DICT = {item.name: item.id for item in PERMISSION_TYPES[:3]} ALL_PERMISSION_TYPES_DICT = {item.name: item.id for item in PERMISSION_TYPES} PERMISSION_TYPE_DICT_ID = {item.id: item.name for item in PERMISSION_TYPES} PERMISSION_SHORT_DICT = {item.name[0]: item.id for item in PERMISSION_TYPES} # These are split into DABS and FABS permissions but having DABS permissions gives read-access to FABS submissions DABS_PERMISSION_ID_LIST = [item.id for item in PERMISSION_TYPES[:3]] FABS_PERMISSION_ID_LIST = [item.id for item in PERMISSION_TYPES[3:]] # These are split into groups between DABS and FABS (not to be confused with just DABS writer/submitter) WRITER_ID_LIST = [item.id for item in PERMISSION_TYPES[1:]] SUBMITTER_ID_LIST = [item.id for item in [PERMISSION_TYPES[2], PERMISSION_TYPES[4]]] FIELD_TYPE = [ LookupType(1, 'INT', 'integer type'), LookupType(2, 'DECIMAL', 'decimal type '), LookupType(3, 'BOOLEAN', 'yes/no'), LookupType(4, 'STRING', 'string type'), LookupType(5, 'LONG', 'long integer'), LookupType(6, 'DATE', 'date type') ] FIELD_TYPE_DICT = {item.name: item.id for item in FIELD_TYPE} FIELD_TYPE_DICT_ID = {item.id: item.name for item in FIELD_TYPE} RULE_SEVERITY = [ LookupType(1, 'warning', 'warning'), LookupType(2, 'fatal', 'fatal error') ] RULE_SEVERITY_DICT = {item.name: item.id for item in RULE_SEVERITY} RULE_IMPACT = [ LookupType(1, 'low', 'low'), LookupType(2, 'medium', 'medium'), LookupType(3, 'high', 'high') ] RULE_IMPACT_DICT = {item.name: item.id for item in RULE_IMPACT} RULE_IMPACT_DICT_ID = {item.id: item.name for item in RULE_IMPACT} SUBMISSION_TYPE = [ LookupType(1, 'all', 'Warning for all pages'), LookupType(2, 'dabs', 'Warning for DABS pages'), LookupType(3, 'fabs', 'Warning for FABS pages'), LookupType(4, 'login', 'Warning for the login page') ] SUBMISSION_TYPE_DICT = {item.name: item.id for item in SUBMISSION_TYPE} ACTION_TYPE = [ LookupType(1, 'A', 'New'), LookupType(2, 'B', 'Continuation'), LookupType(3, 'C', 'Revision'), LookupType(4, 'D', 'Adjustment to Completed Project'), LookupType(4, 'E', 'Aggregate Mixed') ] ACTION_TYPE_DICT = {item.name: item.desc for item in ACTION_TYPE} ASSISTANCE_TYPE = [ LookupType(1, '02', 'block grant (A)'), LookupType(2, '03', 'formula grant (A)'), LookupType(3, '04', 'project grant (B)'), LookupType(4, '05', 'cooperative agreement (B)'), LookupType(5, '06', 'direct payment for specified use, as a subsidy or other non-reimbursable direct financial aid ' '(C)'), LookupType(6, '07', 'direct loan (E)'), LookupType(7, '08', 'guaranteed/insured loan (F)'), LookupType(8, '09', 'insurance (G)'), LookupType(9, '10', 'direct payment with unrestricted use (retirement, pension, veterans benefits, etc.) (D)'), LookupType(10, '11', 'other reimbursable, contingent, intangible, or indirect financial assistance'), ] ASSISTANCE_TYPE_DICT = {item.name: item.desc for item in ASSISTANCE_TYPE} CORRECTION_DELETE_IND = [ LookupType(1, 'C', 'Correct an Existing Record'), LookupType(2, 'D', 'Delete an Existing Record') ] CORRECTION_DELETE_IND_DICT = {item.name: item.desc for item in CORRECTION_DELETE_IND} RECORD_TYPE = [ LookupType(1, 1, 'Aggregate Record'), LookupType(2, 2, 'Non-Aggregate Record'), LookupType(3, 3, 'Non-Aggregate Record to an Individual Recipient (PII-Redacted)') ] RECORD_TYPE_DICT = {item.name: item.desc for item in RECORD_TYPE} BUSINESS_TYPE = [ LookupType(1, 'A', 'State Government'), LookupType(2, 'B', 'County Government'), LookupType(3, 'C', 'City or Township Government'), LookupType(4, 'D', 'Special District Government'), LookupType(5, 'E', 'Regional Organization'), LookupType(6, 'F', 'U.S. Territory or Possession'), LookupType(7, 'G', 'Independent School District'), LookupType(8, 'H', 'Public/State Controlled Institution of Higher Education'), LookupType(9, 'I', 'Indian/Native American Tribal Government (Federally-Recognized)'), LookupType(10, 'J', 'Indian/Native American Tribal Government (Other than Federally-Recognized)'), LookupType(11, 'K', 'Indian/Native American Tribal Designated Organization'), LookupType(12, 'L', 'Public/Indian Housing Authority'), LookupType(13, 'M', 'Nonprofit with 501C3 IRS Status (Other than an Institution of Higher Education)'), LookupType(14, 'N', 'Nonprofit without 501C3 IRS Status (Other than an Institution of Higher Education)'), LookupType(15, 'O', 'Private Institution of Higher Education'), LookupType(16, 'P', 'Individual'), LookupType(17, 'Q', 'For-Profit Organization (Other than Small Business)'), LookupType(18, 'R', 'Small Business'), LookupType(19, 'S', 'Hispanic-serving Institution'), LookupType(20, 'T', 'Historically Black College or University (HBCU)'), LookupType(21, 'U', 'Tribally Controlled College or University (TCCU)'), LookupType(22, 'V', 'Alaska Native and Native Hawaiian Serving Institutions'), LookupType(23, 'W', 'Non-domestic (non-U.S.) Entity'), LookupType(24, 'X', 'Other') ] BUSINESS_TYPE_DICT = {item.name: item.desc for item in BUSINESS_TYPE} SAM_BUSINESS_TYPE = [ LookupType(1, 'A6', 'SBA Certified 8(a), Program Participant'), LookupType(2, 'JT', 'SBA Certified 8(a), Joint Venture'), LookupType(3, 'XX', 'SBA Certified HUBZone Small Business Concern'), LookupType(4, 'A7', 'AbilityOne Non Profit Agency'), LookupType(5, '2R', 'U.S. Federal Government'), LookupType(6, '2F', 'U.S. State Government'), LookupType(7, '12', 'U.S. Local Government'), LookupType(8, '3I', 'Tribal Government'), LookupType(9, 'CY', 'Foreign Government'), LookupType(10, '20', 'Foreign Owned'), LookupType(11, '1D', 'Small Agricultural Cooperative'), LookupType(12, 'LJ', 'Limited Liability Company'), LookupType(13, 'XS', 'Subchapter S Corporation'), LookupType(14, 'MF', 'Manufacturer of Goods'), LookupType(15, '2X', 'For Profit Organization'), LookupType(16, 'A8', 'Non-Profit Organization'), LookupType(17, '2U', 'Other Not For Profit Organization'), LookupType(18, 'HK', 'Community Development Corporation Owned Firm'), LookupType(19, 'A3', 'Labor Surplus Area Firm'), LookupType(20, 'A5', 'Veteran Owned Business'), LookupType(21, 'QF', 'Service Disabled Veteran Owned Business'), LookupType(22, 'A2', 'Woman Owned Business'), LookupType(23, '23', 'Minority Owned Business'), LookupType(24, 'FR', 'Asian-Pacific American Owned'), LookupType(25, 'QZ', 'Subcontinent Asian (Asian-Indian), American Owned'), LookupType(26, 'OY', 'Black American Owned'), LookupType(27, 'PI', 'Hispanic American Owned'), LookupType(28, 'NB', 'Native American Owned'), LookupType(29, 'ZZ', 'Other'), LookupType(30, '8W', 'Woman Owned Small Business'), LookupType(31, '27', 'Self Certified Small Disadvantaged Business'), LookupType(32, 'JX', 'Self Certified HUBZone Joint Venture'), LookupType(33, '8E', 'Economically Disadvantaged Women-Owned Small Business'), LookupType(34, '8C', 'Joint Venture Women-Owned Small Business'), LookupType(35, '8D', 'Economically Disadvantaged Joint Venture Women-Owned Small Business'), LookupType(36, 'NG', 'Federal Agency'), LookupType(37, 'QW', 'Federally Funded Research and Development Center'), LookupType(38, 'C8', 'City'), LookupType(39, 'C7', 'County'), LookupType(40, 'ZR', 'Inter-municipal'), LookupType(41, 'MG', 'Local Government Owned'), LookupType(42, 'C6', 'Municipality'), LookupType(43, 'H6', 'School District'), LookupType(44, 'TW', 'Transit Authority'), LookupType(45, 'UD', 'Council of Governments'), LookupType(46, '8B', 'Housing Authorities Public/Tribal'), LookupType(47, '86', 'Interstate Entity'), LookupType(48, 'KM', 'Planning Commission'), LookupType(49, 'T4', 'Port Authority'), LookupType(50, 'H2', 'Community Development Corporation'), LookupType(51, '6D', 'Domestic Shelter'), LookupType(52, 'M8', 'Educational Institution'), LookupType(53, 'G6', '1862 Land Grant College'), LookupType(54, 'G7', '1890 Land Grant College'), LookupType(55, 'G8', '1994 Land Grant College'), LookupType(56, 'HB', 'Historically Black College or University'), LookupType(57, '1A', 'Minority Institution'), LookupType(58, '1R', 'Private University or College'), LookupType(59, 'ZW', 'School of Forestry'), LookupType(60, 'GW', 'Hispanic Servicing Institution'), LookupType(61, 'OH', 'State Controlled Institution of Higher Learning'), LookupType(62, 'HS', 'Tribal College'), LookupType(63, 'QU', 'Veterinary College'), LookupType(64, 'G3', 'Alaskan Native Servicing Institution'), LookupType(65, 'G5', 'Native Hawaiian Servicing Institution'), LookupType(66, 'BZ', 'Foundation'), LookupType(67, '80', 'Hospital'), LookupType(68, 'FY', 'Veterinary Hospital'), LookupType(69, 'HQ', 'DOT Certified DBE'), LookupType(70, '05', 'Alaskan Native Corporation Owned Firm'), LookupType(71, 'OW', 'American Indian Owned'), LookupType(72, 'XY', 'Indian Tribe (Federally Recognized),'), LookupType(73, '8U', 'Native Hawaiian Organization Owned Firm'), LookupType(74, '1B', 'Tribally Owned Firm'), LookupType(75, 'FO', 'Township'), LookupType(76, 'TR', 'Airport Authority'), LookupType(77, 'G9', 'Other Than One of the Proceeding'), LookupType(78, '2J', 'Sole Proprietorship'), LookupType(79, '2K', 'Partnership or Limited Liability Partnership'), LookupType(80, '2L', 'Corporate Entity (Not Tax Exempt),'), LookupType(81, '8H', 'Corporate Entity (Tax Exempt),'), LookupType(82, '2A', 'U.S. Government Entity'), LookupType(83, 'X6', 'International Organization') ] SAM_BUSINESS_TYPE_DICT = {item.name: item.desc for item in SAM_BUSINESS_TYPE} BUSINESS_FUNDS_IND = [ LookupType(1, 'NON', 'Not Recovery Act'), LookupType(2, 'REC', 'Recovery Act') ] BUSINESS_FUNDS_IND_DICT = {item.name: item.desc for item in BUSINESS_FUNDS_IND} BUSINESS_CATEGORY_FIELDS = ['airport_authority', 'alaskan_native_owned_corpo', 'alaskan_native_servicing_i', 'american_indian_owned_busi', 'asian_pacific_american_own', 'black_american_owned_busin', 'c1862_land_grant_college', 'c1890_land_grant_college', 'c1994_land_grant_college', 'c8a_program_participant', 'city_local_government', 'community_developed_corpor', 'community_development_corp', 'contracting_officers_deter', 'corporate_entity_not_tax_e', 'corporate_entity_tax_exemp', 'council_of_governments', 'county_local_government', 'domestic_or_foreign_entity', 'domestic_shelter', 'dot_certified_disadvantage', 'economically_disadvantaged', 'educational_institution', 'emerging_small_business', 'federal_agency', 'federally_funded_research', 'for_profit_organization', 'foreign_government', 'foreign_owned_and_located', 'foundation', 'hispanic_american_owned_bu', 'hispanic_servicing_institu', 'historically_black_college', 'historically_underutilized', 'hospital_flag', 'housing_authorities_public', 'indian_tribe_federally_rec', 'inter_municipal_local_gove', 'international_organization', 'interstate_entity', 'joint_venture_economically', 'joint_venture_women_owned', 'labor_surplus_area_firm', 'limited_liability_corporat', 'local_government_owned', 'manufacturer_of_goods', 'minority_institution', 'minority_owned_business', 'municipality_local_governm', 'native_american_owned_busi', 'native_hawaiian_owned_busi', 'native_hawaiian_servicing', 'nonprofit_organization', 'other_minority_owned_busin', 'other_not_for_profit_organ', 'partnership_or_limited_lia', 'planning_commission', 'port_authority', 'private_university_or_coll', 'sba_certified_8_a_joint_ve', 'school_district_local_gove', 'school_of_forestry', 'self_certified_small_disad', 'service_disabled_veteran_o', 'small_agricultural_coopera', 'small_disadvantaged_busine', 'sole_proprietorship', 'state_controlled_instituti', 'subchapter_s_corporation', 'subcontinent_asian_asian_i', 'the_ability_one_program', 'township_local_government', 'transit_authority', 'tribal_college', 'tribally_owned_business', 'us_federal_government', 'us_government_entity', 'us_local_government', 'us_state_government', 'us_tribal_government', 'veteran_owned_business', 'veterinary_college', 'veterinary_hospital', 'woman_owned_business', 'women_owned_small_business'] EXTERNAL_DATA_TYPE = [ LookupType(1, 'usps_download', 'external data load type for downloading zip files'), LookupType(2, 'program_activity_upload', 'program activity file loaded into S3'), LookupType(3, 'cfda', 'GSA CFDA loaded'), LookupType(4, 'agency', 'IAE agency data loaded'), LookupType(5, 'tas', 'FRB CARS/TAS data loaded'), LookupType(6, 'city', 'USGS city data loaded'), LookupType(7, 'congressional_district', 'USPS congressional district data loaded'), LookupType(8, 'country_code', 'country code data loaded'), LookupType(9, 'county_code', 'USGS county code data loaded'), LookupType(10, 'recipient', 'SAM DUNS/UEI data loaded'), LookupType(11, 'executive_compensation', 'SAM executive compensation data loaded'), LookupType(12, 'fpds', 'FPDS data loaded'), LookupType(13, 'gtas', 'FRB gtas/sf-133 data loaded'), LookupType(14, 'object_class', 'OMB object class data loaded'), LookupType(15, 'office', 'GSA office data loaded'), LookupType(16, 'program_activity', 'OMB program activity data loaded'), LookupType(17, 'state_code', 'state code data loaded'), LookupType(18, 'subaward', 'FSRS subaward data loaded'), LookupType(19, 'zip_code', 'USPS zip code data loaded'), LookupType(20, 'defc', 'disaster emergency fund code data loaded'), LookupType(21, 'failed_tas', 'TAS failing edits data loaded') ] EXTERNAL_DATA_TYPE_DICT = {item.name: item.id for item in EXTERNAL_DATA_TYPE}
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/778cabef7323_rename_and_add_office_type_booleans_in_.py
1
1561
"""Rename and add office type booleans in office table Revision ID: 778cabef7323 Revises: 7dd3f4b007e5 Create Date: 2019-05-10 10:01:26.888903 """ # revision identifiers, used by Alembic. revision = '778cabef7323' down_revision = '5f29b283f23e' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.execute("ALTER TABLE office RENAME COLUMN funding_office TO contract_funding_office") op.execute("ALTER TABLE office RENAME COLUMN contracting_office TO contract_awards_office") op.execute("ALTER TABLE office RENAME COLUMN financial_assistance_office TO financial_assistance_awards_office") op.add_column('office', sa.Column('financial_assistance_funding_office', sa.Boolean(), server_default='False', nullable=False)) # ### end Alembic commands ### def downgrade_data_broker(): # ### commands auto generated by Alembic - please adjust! ### op.execute("ALTER TABLE office RENAME COLUMN contract_funding_office TO funding_office") op.execute("ALTER TABLE office RENAME COLUMN contract_awards_office TO contracting_office") op.execute("ALTER TABLE office RENAME COLUMN financial_assistance_awards_office TO financial_assistance_office") op.drop_column('office', 'financial_assistance_funding_office') # ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_fabs39_2.py
1
2847
from tests.unit.dataactcore.factories.staging import FABSFactory from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'fabs39_2' def test_column_headers(database): expected_subset = {'row_number', 'record_type', 'place_of_performance_code', 'place_of_perform_country_c', 'uniqueid_AssistanceTransactionUniqueKey'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ PrimaryPlaceOfPerformanceCode must be 00FORGN when PrimaryPlaceofPerformanceCountryCode is not USA, not 00FORGN otherwise for record type 1 and 2. """ fabs_1 = FABSFactory(place_of_performance_code='00FORGN', place_of_perform_country_c='UKR', record_type=1, correction_delete_indicatr='') fabs_2 = FABSFactory(place_of_performance_code='00FoRGN', place_of_perform_country_c='uKr', record_type=1, correction_delete_indicatr=None) fabs_3 = FABSFactory(place_of_performance_code='ny**987', place_of_perform_country_c='USA', record_type=2, correction_delete_indicatr='c') fabs_4 = FABSFactory(place_of_performance_code='NY**987', place_of_perform_country_c='UsA', record_type=2, correction_delete_indicatr='C') fabs_5 = FABSFactory(place_of_performance_code='NY**987', place_of_perform_country_c='UKR', record_type=3, correction_delete_indicatr='') # Ignore correction delete indicator of D fabs_6 = FABSFactory(place_of_performance_code='00FORGN', place_of_perform_country_c='USA', record_type=1, correction_delete_indicatr='d') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4, fabs_5, fabs_6]) assert errors == 0 def test_failure(database): """ Test failure for PrimaryPlaceOfPerformanceCode must be 00FORGN when PrimaryPlaceofPerformanceCountryCode is not USA, not 00FORGN otherwise for record type 1 and 2. """ fabs_1 = FABSFactory(place_of_performance_code='00FORGN', place_of_perform_country_c='USA', record_type=1, correction_delete_indicatr='') fabs_2 = FABSFactory(place_of_performance_code='00FoRGN', place_of_perform_country_c='usA', record_type=1, correction_delete_indicatr=None) fabs_3 = FABSFactory(place_of_performance_code='ny**987', place_of_perform_country_c='UKR', record_type=2, correction_delete_indicatr='c') fabs_4 = FABSFactory(place_of_performance_code='NY**987', place_of_perform_country_c='ukR', record_type=2, correction_delete_indicatr='C') errors = number_of_errors(_FILE, database, models=[fabs_1, fabs_2, fabs_3, fabs_4]) assert errors == 4
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/integration/integration_test_helper.py
1
3026
import calendar from datetime import datetime from dataactcore.models.jobModels import Submission, Job def insert_submission(sess, submission_user_id, cgac_code=None, frec_code=None, start_date=None, end_date=None, is_quarter=False, number_of_errors=0, publish_status_id=1, is_fabs=False, updated_at=datetime.utcnow(), test_submission=False, published_submission_ids=[], certified=False, reporting_fiscal_year=None, reporting_fisacal_period=None): """Insert one submission into job tracker and get submission ID back.""" publishable = True if number_of_errors == 0 else False if start_date is not None: start_date = datetime.strptime(start_date, '%m/%Y') if end_date is not None: end_date = datetime.strptime(end_date, '%m/%Y') end_date = datetime.strptime( str(end_date.year) + '/' + str(end_date.month) + '/' + str(calendar.monthrange(end_date.year, end_date.month)[1]), '%Y/%m/%d' ).date() sub = Submission(created_at=datetime.utcnow(), updated_at=updated_at, user_id=submission_user_id, cgac_code=cgac_code, frec_code=frec_code, reporting_start_date=start_date, reporting_end_date=end_date, reporting_fiscal_year=reporting_fiscal_year, reporting_fiscal_period=reporting_fisacal_period, is_quarter_format=is_quarter, number_of_errors=number_of_errors, publish_status_id=publish_status_id, publishable=publishable, is_fabs=is_fabs, test_submission=test_submission, published_submission_ids=published_submission_ids, certified=certified) sess.add(sub) sess.commit() return sub.submission_id def insert_job(sess, filetype, status, type_id, submission, job_id=None, filename=None, original_filename=None, file_size=None, num_rows=None, num_valid_rows=0, num_errors=0, updated_at=None): """Insert one job into job tracker and get ID back.""" if not updated_at: updated_at = datetime.utcnow() job = Job( created_at=datetime.utcnow(), updated_at=updated_at, file_type_id=filetype, job_status_id=status, job_type_id=type_id, submission_id=submission, filename=filename, original_filename=original_filename, file_size=file_size, number_of_rows=num_rows, number_of_rows_valid=num_valid_rows, number_of_errors=num_errors ) if job_id: job.job_id = job_id sess.add(job) sess.commit() return job def get_submission(sess, sub_id): """ Get back the requested submission """ return sess.query(Submission).filter_by(submission_id=sub_id).one_or_none()
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/361fbffcf08b_added_named_job_dep_fks.py
2
1044
"""added_named_job_dep_fks Revision ID: 361fbffcf08b Revises: 9058e0136aba Create Date: 2016-08-26 10:33:18.803848 """ # revision identifiers, used by Alembic. revision = '361fbffcf08b' down_revision = '9058e0136aba' branch_labels = None depends_on = None from alembic import op def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.create_foreign_key('fk_dep_job_id', 'job_dependency', 'job', ['job_id'], ['job_id']) op.create_foreign_key('fk_prereq_job_id', 'job_dependency', 'job', ['prerequisite_id'], ['job_id']) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_constraint('fk_prereq_job_id', 'job_dependency', type_='foreignkey') op.drop_constraint('fk_dep_job_id', 'job_dependency', type_='foreignkey') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/scripts/populate_subaward_table.py
1
6271
import os import argparse import datetime import logging import json import sys from dataactcore.interfaces.db import GlobalDB from dataactcore.config import CONFIG_BROKER from dataactcore.broker_logging import configure_logging from dataactvalidator.health_check import create_app from dataactbroker.fsrs import GRANT, PROCUREMENT RAW_SQL_DIR = os.path.join(CONFIG_BROKER['path'], 'dataactcore', 'scripts', 'raw_sql') POPULATE_PROCUREMENT_SQL = os.path.join(RAW_SQL_DIR, 'populate_subaward_table_contracts.sql') POPULATE_GRANT_SQL = os.path.join(RAW_SQL_DIR, 'populate_subaward_table_grants.sql') LINK_PROCUREMENT_SQL = os.path.join(RAW_SQL_DIR, 'link_broken_subaward_contracts.sql') LINK_GRANT_SQL = os.path.join(RAW_SQL_DIR, 'link_broken_subaward_grants.sql') logger = logging.getLogger(__name__) def extract_subaward_sql(service_type, data_change_type): """ Gather the subaward SQL requested Args: service_type: type of service to ping ('procurement_service' or 'grant_service') data_change_type: type of data change involving subawards ('populate' or 'link') Raises: Exception: service type is invalid Exception: data change type is invalid """ pop_sql_map = {PROCUREMENT: POPULATE_PROCUREMENT_SQL, GRANT: POPULATE_GRANT_SQL} link_sql_map = {PROCUREMENT: LINK_PROCUREMENT_SQL, GRANT: LINK_GRANT_SQL} if service_type not in pop_sql_map: raise Exception('Invalid service type provided: {}'.format(service_type)) type_map = {'populate': pop_sql_map, 'link': link_sql_map} if data_change_type not in type_map: raise Exception('Invalid data change type provided: {}'.format(data_change_type)) with open(type_map[data_change_type][service_type], 'r') as sql_file: sql = sql_file.read() return sql def populate_subaward_table(sess, service_type, ids=None, min_id=None): """ Populates the subaward table based on the IDS (or min id) provided Args: sess: connection to the database service_type: type of service to ping (usually 'procurement_service' or 'grant_service') ids: if provided, only update these ids min_id: if provided, update all ids past this one Raises: Exception: ids and min_id both provided or both not provided Exception: service type is invalid """ if (ids is not None and min_id is not None) or (ids is None and min_id is None): raise Exception('ids or min_id, but not both, must be provided') sql = extract_subaward_sql(service_type, 'populate') if min_id is not None: operator = '>' values = min_id - 1 else: operator = 'IN' values = '({})'.format(','.join([str(id) for id in ids])) sql = sql.format(operator, values) # run the SQL inserted = sess.execute(sql) sess.commit() inserted_count = inserted.rowcount award_type = service_type[:service_type.index('_')] logger.info('Inserted {} sub-{}s to the subaward table'.format(inserted_count, award_type)) return inserted_count def fix_broken_links(sess, service_type, min_date=None): """ Attempts to resolve any unlinked subawards given the current data Args: sess: connection to the database service_type: type of service to ping (usually 'procurement_service' or 'grant_service') Raises: Exception: service type is invalid """ award_type = service_type[:service_type.index('_')] logger.info('Attempting to fix broken sub-{} links in the subaward table'.format(award_type)) subaward_type_map = {PROCUREMENT: 'sub-contract', GRANT: 'sub-grant'} if service_type not in subaward_type_map: raise Exception('Invalid service type provided: {}'.format(service_type)) sql = extract_subaward_sql(service_type, 'link') min_date_sql = '' if min_date is None else 'AND updated_at >= \'{}\''.format(min_date) sql = sql.format(min_date_sql) # run the SQL updated = sess.execute(sql) sess.commit() updated_count = updated.rowcount logger.info('Updated {} sub-{}s in the subaward table'.format(updated_count, award_type)) return updated_count if __name__ == '__main__': now = datetime.datetime.now() configure_logging() parser = argparse.ArgumentParser(description='Pull data from FSRS Feed') method = parser.add_mutually_exclusive_group(required=True) parser.add_argument('-p', '--procurements', action='store_true', help="Load just procurement awards") parser.add_argument('-g', '--grants', action='store_true', help="Load just grant awards") method.add_argument('-m', '--min_id', type=int, nargs=1, help="Load all data from a minimum id (0 for complete" " backfill)") method.add_argument('-i', '--ids', type=int, nargs='+', help="Single or list of FSRS ids to populate the subaward table") with create_app().app_context(): logger.info("Begin backfilling Subaward table") sess = GlobalDB.db().session args = parser.parse_args() metrics_json = { 'script_name': 'populate_subaward_table.py', 'records_inserted': 0, 'start_time': str(now) } service_types = [] if not (args.procurements or args.grants): logger.error('FSRS types not provided. Please specify procurements, grants, or both.') sys.exit(1) if args.procurements: service_types.append(PROCUREMENT) if args.grants: service_types.append(GRANT) records_inserted = 0 for service_type in service_types: if args.min_id: records_inserted += populate_subaward_table(sess, service_type, min_id=args.min_id[0]) elif args.ids: records_inserted += populate_subaward_table(sess, service_type, ids=args.ids) metrics_json['records_inserted'] = records_inserted metrics_json['duration'] = str(datetime.datetime.now() - now) with open('populate_subaward_table .json', 'w+') as metrics_file: json.dump(metrics_json, metrics_file)
cc0-1.0
fedspendingtransparency/data-act-broker-backend
tests/unit/dataactvalidator/test_a7_appropriations.py
1
2660
from dataactcore.models.stagingModels import Appropriation from dataactcore.models.domainModels import SF133 from tests.unit.dataactvalidator.utils import number_of_errors, query_columns _FILE = 'a7_appropriations' _TAS = 'a7_appropriations_tas' def test_column_headers(database): expected_subset = {'uniqueid_TAS', 'row_number', 'budget_authority_unobligat_fyb', 'expected_value_GTAS SF133 Line 1000', 'difference'} actual = set(query_columns(_FILE, database)) assert expected_subset == actual def test_success(database): """ Tests that SF 133 amount for line 1000 matches Appropriation budget_authority_unobligat_fyb for the specified fiscal year and period """ tas_1 = 'tas_one_line_1' tas_2 = 'tas_one_line_2' sf_1 = SF133(line=1000, tas=tas_1, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000') sf_2 = SF133(line=1000, tas=tas_2, period=1, fiscal_year=2016, amount=0, agency_identifier='sys', main_account_code='000', sub_account_code='000') ap_1 = Appropriation(job_id=1, row_number=1, tas=tas_1, budget_authority_unobligat_fyb=1) ap_2 = Appropriation(job_id=2, row_number=1, tas=tas_2, budget_authority_unobligat_fyb=None) assert number_of_errors(_FILE, database, models=[sf_1, sf_2, ap_1, ap_2]) == 0 # Test with split SF133 lines tas = 'tas_two_lines' sf_1 = SF133(line=1000, tas=tas, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000', disaster_emergency_fund_code='n') sf_2 = SF133(line=1000, tas=tas, period=1, fiscal_year=2016, amount=4, agency_identifier='sys', main_account_code='000', sub_account_code='000', disaster_emergency_fund_code='o') ap = Appropriation(job_id=1, row_number=1, tas=tas, budget_authority_unobligat_fyb=5) assert number_of_errors(_FILE, database, models=[sf_1, sf_2, ap]) == 0 def test_failure(database): """ Tests that SF 133 amount for line 1000 does not match Appropriation budget_authority_unobligat_fyb for the specified fiscal year and period """ tas = 'fail_tas' sf = SF133(line=1000, tas=tas, period=1, fiscal_year=2016, amount=1, agency_identifier='sys', main_account_code='000', sub_account_code='000') ap_1 = Appropriation(job_id=1, row_number=1, tas=tas, budget_authority_unobligat_fyb=0) ap_2 = Appropriation(job_id=2, row_number=1, tas=tas, budget_authority_unobligat_fyb=None) assert number_of_errors(_FILE, database, models=[sf, ap_1, ap_2]) == 2
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/migrations/versions/da2e50d423ff_create_frec_table.py
1
1237
"""create FREC table Revision ID: da2e50d423ff Revises: aa10ae595d3e Create Date: 2017-07-06 10:27:04.738865 """ # revision identifiers, used by Alembic. revision = 'da2e50d423ff' down_revision = 'aa10ae595d3e' branch_labels = None depends_on = None from alembic import op import sqlalchemy as sa def upgrade(engine_name): globals()["upgrade_%s" % engine_name]() def downgrade(engine_name): globals()["downgrade_%s" % engine_name]() def upgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.create_table('frec', sa.Column('created_at', sa.DateTime(), nullable=True), sa.Column('updated_at', sa.DateTime(), nullable=True), sa.Column('frec_id', sa.Integer(), nullable=False), sa.Column('frec_code', sa.Text(), nullable=True), sa.Column('agency_name', sa.Text(), nullable=True), sa.PrimaryKeyConstraint('frec_id') ) op.create_index(op.f('ix_frec_frec_code'), 'frec', ['frec_code'], unique=True) ### end Alembic commands ### def downgrade_data_broker(): ### commands auto generated by Alembic - please adjust! ### op.drop_index(op.f('ix_frec_frec_code'), table_name='frec') op.drop_table('frec') ### end Alembic commands ###
cc0-1.0
fedspendingtransparency/data-act-broker-backend
dataactcore/scripts/generate_comments_files.py
1
4873
import logging from sqlalchemy import func from dataactcore.aws.s3Handler import S3Handler from dataactcore.config import CONFIG_BROKER from dataactcore.interfaces.db import GlobalDB from dataactcore.interfaces.function_bag import filename_fyp_sub_format from dataactcore.broker_logging import configure_logging from dataactcore.models.jobModels import (Comment, PublishedComment, FileType, CertifyHistory, PublishHistory, PublishedFilesHistory, Submission) from dataactvalidator.filestreaming.csv_selection import write_stream_query from dataactvalidator.health_check import create_app from dataactcore.models.userModel import User # noqa logger = logging.getLogger(__name__) if __name__ == '__main__': sess = GlobalDB.db().session configure_logging() with create_app().app_context(): # Unpublished comments files logger.info('Generating unpublished comments files') is_local = CONFIG_BROKER['local'] headers = ['File', 'Comment'] commented_submissions = sess.query(Comment.submission_id).distinct() for submission in commented_submissions: # Preparing for the comments files submission_id = submission.submission_id submission = sess.query(Submission).filter_by(submission_id=submission_id).one() filename = 'SubID-{}_comments_{}.csv'.format(submission_id, filename_fyp_sub_format(submission)) local_file = "".join([CONFIG_BROKER['broker_files'], filename]) file_path = local_file if is_local else '{}/{}'.format(str(submission_id), filename) unpublished_query = sess.query(FileType.name, Comment.comment).\ join(FileType, Comment.file_type_id == FileType.file_type_id).\ filter(Comment.submission_id == submission_id) # Generate the file locally, then place in S3 write_stream_query(sess, unpublished_query, local_file, file_path, is_local, header=headers) logger.info('Finished generating unpublished comments files') # Published comments files logger.info('Copying published comments files') commented_pub_submissions = sess.query(PublishedComment.submission_id).distinct() for published_submission in commented_pub_submissions: submission_id = published_submission.submission_id submission = sess.query(Submission).filter_by(submission_id=submission_id).one() filename = 'SubID-{}_comments_{}.csv'.format(str(submission_id), filename_fyp_sub_format(submission)) # See if we already have this published file in the list existing_pub_history = sess.query(PublishedFilesHistory).\ filter(PublishedFilesHistory.submission_id == submission_id, PublishedFilesHistory.filename.like('%' + filename)).one_or_none() # If we already have this file in the table, we don't need to make it again, we can skip if existing_pub_history: continue agency_code = submission.frec_code if submission.frec_code else submission.cgac_code # Get the latest publish+certify IDs max_pub_id = sess.query(func.max(PublishHistory.publish_history_id).label('pub_id')).\ filter_by(submission_id=submission_id).one() max_cert_id = sess.query(func.max(CertifyHistory.certify_history_id).label('cert_id')). \ filter_by(submission_id=submission_id).one() route_vars = [agency_code, submission.reporting_fiscal_year, submission.reporting_fiscal_period // 3, max_pub_id.pub_id] new_route = '/'.join([str(var) for var in route_vars]) + '/' if not is_local: old_path = '{}/{}'.format(str(submission_id), filename) new_path = new_route + filename # Copy the file if it's a non-local submission S3Handler().copy_file(original_bucket=CONFIG_BROKER['aws_bucket'], new_bucket=CONFIG_BROKER['certified_bucket'], original_path=old_path, new_path=new_path) else: new_path = "".join([CONFIG_BROKER['broker_files'], filename]) # add published history file_history = PublishedFilesHistory(publish_history_id=max_pub_id.pub_id, certify_history_id=max_cert_id.cert_id, submission_id=submission_id, filename=new_path, file_type_id=None, comment=None, warning_filename=None) sess.add(file_history) sess.commit() logger.info('Finished copying published comments files')
cc0-1.0