diff --git "a/3351.jsonl" "b/3351.jsonl" new file mode 100644--- /dev/null +++ "b/3351.jsonl" @@ -0,0 +1,689 @@ +{"seq_id":"434281555","text":"# from: https://roboticsbackend.com/raspberry-pi-arduino-serial-communication/\n\n# run: python3 serial3_udp_spk.py\n\n#!/usr/bin/env python3\nimport serial\nimport time\nimport socket\nimport pyttsx3\n\n#engine = pyttsx3.init()\n\nif __name__ == '__main__':\n\n UDP_IP = \"192.168.1.208\"\n UDP_PORT = 5005\n\n sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # UDP\n sock.bind((UDP_IP, UDP_PORT))\n\n print(\"Waiting to receive data via UDP ...\" )\n\n while True:\n data, addr = sock.recvfrom(1024) # buffer size is 1024 bytes\n print(\"recv:\", data)\n\n #ser = serial.Serial('/dev/ttyACM0', 9600, timeout=1)\n ser = serial.Serial('/dev/ttyACM0', 115200, timeout=1)\n ser.flush()\n\n #while True:\n #ser.write(b\"Hello from Raspberry Pi!\\n\")\n #ser.write(data.encode('utf-8'))\n ser.write(data)\n line = ser.readline().decode('utf-8').rstrip()\n print(\"pi:recv: {%s}\" % (line) )\n #time.sleep(1)\n\n #engine.say(\"yeah\")\n #engine.runAndWait()\n\n","sub_path":"sensor_set/serial/serial3_udp_spk.py","file_name":"serial3_udp_spk.py","file_ext":"py","file_size_in_byte":1025,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"96723678","text":"from abc import ABC, abstractmethod\nimport json\nimport time\nfrom importlib import import_module\n\nfrom dask_ml.model_selection import GridSearchCV\nimport dask.dataframe as dd\nimport joblib\nimport pandas as pd\n\nfrom dask_ml.model_selection import train_test_split\n# from sklearn.model_selection import GridSearchCV as SkGridSearchCV\n\n\n# Hack for DaskML bug: impossible to get best params without refit=True\ndef _get_best_params_score(grigsearch):\n df_cv_res = pd.DataFrame(grigsearch.cv_results_)\n best_params = df_cv_res.loc[df_cv_res[\"rank_test_score\"] == 1, \"params\"].iloc[0]\n mean_test_score = df_cv_res.loc[df_cv_res[\"rank_test_score\"] == 1, \"mean_test_score\"].iloc[0]\n\n return df_cv_res, best_params, mean_test_score\n\n\nclass ExperimentRunner:\n \"\"\"\n This class encapsulates data preparation and running the CV experiment.\n Assumed to be run in presence of a dask client (and possibly a cluster)\n \"\"\"\n def __init__(self, input_path, target_col, model_desc):\n self.input_path = input_path\n self.target_col = target_col\n self.model_desc = model_desc\n\n def load_data(self):\n df = dd.read_hdf(self.input_path, '/data', mode='r')\n\n self.X = df.drop(self.target_col, axis=1)\n self.y = df[self.target_col]\n\n def split_data(self):\n self.X_train, self.X_test, self.y_train, self.y_test = train_test_split(self.X, self.y,\n test_size=float(\n self.model_desc.get(\"test_size\",\n 0.5)),\n random_state=int(\n self.model_desc.get(\"seed\", 31337)))\n\n def _create_model(self, **kwargs):\n module_name, class_name = self.model_desc[\"algorithm_name\"].rsplit('.', 1)\n return getattr(import_module(module_name), class_name)(**kwargs)\n\n def run(self):\n self.load_data()\n self.split_data()\n\n # nulls = X_train.isnull().sum()\n # total_nulls = nulls.sum()\n # if total_nulls > 0:\n # with pd.option_context('display.max_rows', None, 'display.max_columns', None):\n # print(nulls[nulls > 0], \"total: \", total_nulls)\n\n reg = self._create_model()\n gs = GridSearchCV(reg, self.model_desc[\"params_grid\"], cv=self.model_desc[\"num_folds\"], n_jobs=-1, refit=False)\n\n start_time = time.monotonic()\n with joblib.parallel_backend(\"dask\", scatter=[self.X_train, self.y_train]):\n gs.fit(self.X_train, self.y_train)\n finish_time = time.monotonic()\n gs_time = finish_time - start_time\n print(\"Searching for marameters for [{}]\".format(self.model_desc[\"algorithm_name\"]))\n print(\"GridSearchCV time: {}\".format(gs_time))\n cv_res, best_params, gs_score = _get_best_params_score(gs)\n print(\"GridSearchCV score: {}\".format(gs_score))\n print(\"Best params: {}\".format(best_params))\n\n start_time = time.monotonic()\n regr_best = self._create_model(**best_params).fit(self.X_train, self.y_train)\n finish_time = time.monotonic()\n test_time = finish_time - start_time\n print(\"Final training time: {}\".format(test_time))\n\n test_score = float(regr_best.score(self.X_test, self.y_test))\n\n print(\"Test score: {}\".format(test_score))\n\n ans = {\"algorithm_name\": self.model_desc[\"algorithm_name\"],\n \"gs_time\": gs_time,\n \"gs_score\": gs_score,\n \"test_time\": test_time,\n \"test_score\": test_score}\n\n if \"output_path\" in self.model_desc:\n with open(self.model_desc[\"output_path\"], \"w\") as fp:\n json.dump(ans, fp)\n\n return ans\n\n\nclass PersistExperimentRunner(ExperimentRunner):\n \"\"\"\n Persist all data in memory for compatibility with some algorithms.\n \"\"\"\n def load_data(self):\n super().load_data()\n self.X = self.X.compute()\n self.y = self.y.compute()\n\n\nclass ArrayExperimentRunner(ExperimentRunner):\n \"\"\"\n Convert data to dask arrays. Some dask ans sklearn algorithms do not support dask DataFrame.\n \"\"\"\n def load_data(self):\n super().load_data()\n self.X = self.X.to_dask_array(lengths=True)\n self.y = self.y.to_dask_array(lengths=True)\n\n\nrunners = {\n \"dask\": ExperimentRunner,\n \"dask_array\": ArrayExperimentRunner,\n \"sklearn\": PersistExperimentRunner\n}\n\n\ndef get_experiment_runner(path, target, desc):\n m = desc[\"mode\"]\n return runners[m](input_path=path,\n target_col=target,\n model_desc=desc)\n","sub_path":"Week 2/Day 3/Submissions/Alexander Fishkov/submission/runner.py","file_name":"runner.py","file_ext":"py","file_size_in_byte":4877,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"211521671","text":"# coding: utf-8\n\nprint(\"TP4 - Ex\")\n\n# Consigne :\nprint(\"\\nDans le code suivant insérer plt.grid() avant plt.show() et \"+\n \"relancer. Ajouter ensuite plt.axhline(color='red') et relancer. Enfin \"+\n \"insérer plt.legend(['sinus'],loc='upper left')\\n\\n\")\n\n# Exercice :\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom math import pi\n\nx = np.linspace(-2*pi,2*pi,1000)\n# Ensemble de x qui vont être pris en paramètres par la fct. (comme un range,\n# mais avec des float).\n\nplt.plot(x,np.sin(x)) \n# on utilise la fonction sinus du module Numpy\n\nplt.ylabel('La fonction sinus')\nplt.xlabel(\"L'axe des abcisses\")\n# Ajoute des labels pour les axes.\n\nplt.grid()\n# Affiche la grille.\n\nplt.axhline(color='red')\n# Colore l'axe des abscisses en rouge.\n\nplt.legend(['sinus'],loc='upper left')\n# Indique en haut à gauche la légende 'sinus'.\n\nplt.show()\n# Affiche le graphe\n\n# Conclusion :\nprint(\"\\nRésultat :\\nLe module numpy permet de faire du calcul scientifique, \"+\n \"le module matplotlib.pyplot permet de faire des graphiques.\")","sub_path":"Info/TP/TP4/TP4 - Ex01.py","file_name":"TP4 - Ex01.py","file_ext":"py","file_size_in_byte":1037,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"68890087","text":"\"\"\"\nTests for magic imports.\n\"\"\"\n\nimport os\n\nimport pandas as pd\nfrom six import string_types\n\nfrom quilt.data import GroupNode, DataNode\nfrom quilt.tools import command\nfrom quilt.tools.const import PACKAGE_DIR_NAME\nfrom quilt.tools.package import Package, PackageException\nfrom quilt.tools.store import PackageStore\nfrom .utils import QuiltTestCase\n\nclass ImportTest(QuiltTestCase):\n def test_imports(self):\n mydir = os.path.dirname(__file__)\n build_path = os.path.join(mydir, './build.yml')\n command.build('foo/package', build_path)\n\n # Good imports\n\n from quilt.data.foo import package\n from quilt.data.foo.package import dataframes\n from quilt.data.foo.package import README\n\n # Contents of the imports\n\n assert isinstance(package, GroupNode)\n assert isinstance(dataframes, GroupNode)\n assert isinstance(dataframes.csv, DataNode)\n assert isinstance(README, DataNode)\n\n assert package.dataframes == dataframes\n assert package.README == README\n\n assert set(dataframes._keys()) == {'xls', 'csv', 'tsv', 'xls_skip', 'nulls'}\n assert set(dataframes._group_keys()) == set()\n assert set(dataframes._data_keys()) == {'xls', 'csv', 'tsv', 'xls_skip', 'nulls'}\n\n assert isinstance(README(), string_types)\n assert isinstance(README._data(), string_types)\n assert isinstance(dataframes.csv(), pd.DataFrame)\n assert isinstance(dataframes.csv._data(), pd.DataFrame)\n\n str(package)\n str(dataframes)\n str(README)\n\n # Bad attributes of imported packages\n\n with self.assertRaises(AttributeError):\n package.foo\n\n with self.assertRaises(AttributeError):\n package.dataframes.foo\n\n with self.assertRaises(AttributeError):\n package.dataframes.csv.foo\n\n # Bad imports\n\n with self.assertRaises(ImportError):\n import quilt.data.foo.bad_package\n\n with self.assertRaises(ImportError):\n import quilt.data.bad_user.bad_package\n\n with self.assertRaises(ImportError):\n from quilt.data.foo.dataframes import blah\n\n with self.assertRaises(ImportError):\n from quilt.data.foo.baz import blah\n\n def test_import_group_as_data(self):\n mydir = os.path.dirname(__file__)\n build_path = os.path.join(mydir, './build_group_data.yml')\n command.build('foo/grppkg', build_path)\n\n # Good imports\n from quilt.data.foo.grppkg import dataframes\n assert isinstance(dataframes, GroupNode)\n assert isinstance(dataframes.csvs.csv, DataNode)\n assert isinstance(dataframes._data(), pd.DataFrame)\n\n # Incompatible Schema\n from quilt.data.foo.grppkg import incompatible\n with self.assertRaises(PackageException):\n incompatible._data()\n\n def test_multiple_package_dirs(self):\n # First level\n mydir = os.path.dirname(__file__)\n build_path = os.path.join(mydir, './build_simple.yml')\n command.build('foo/nested', build_path)\n\n # Second level: different package\n os.mkdir(\"aaa\")\n os.chdir(\"aaa\")\n build_path = os.path.join(mydir, './build.yml')\n command.build('foo/nested', build_path)\n\n # Third level: empty package directory\n os.mkdir(\"bbb\")\n os.chdir(\"bbb\")\n os.mkdir(PACKAGE_DIR_NAME)\n\n # Imports should find the second package\n from quilt.data.foo.nested import dataframes\n\n def test_save(self):\n mydir = os.path.dirname(__file__)\n build_path = os.path.join(mydir, './build.yml')\n command.build('foo/package1', build_path)\n\n from quilt.data.foo import package1\n\n # Build an identical package\n command.build('foo/package2', package1)\n\n from quilt.data.foo import package2\n teststore = PackageStore(self._store_dir)\n contents1 = open(os.path.join(teststore.package_path('foo', 'package1'),\n Package.CONTENTS_DIR,\n package1._package.get_hash())).read()\n contents2 = open(os.path.join(teststore.package_path('foo', 'package2'),\n Package.CONTENTS_DIR,\n package2._package.get_hash())).read()\n assert contents1 == contents2\n\n # Rename an attribute\n package1.dataframes2 = package1.dataframes\n del package1.dataframes\n\n # Modify an existing dataframe\n csv = package1.dataframes2.csv._data()\n csv.at[0, 'Int0'] = 42\n\n # Add a new dataframe\n df = pd.DataFrame(dict(a=[1, 2, 3]))\n package1._set(['new', 'df'], df)\n assert package1.new.df._data() is df\n\n # Add a new file\n file_path = os.path.join(mydir, 'data/foo.csv')\n package1._set(['new', 'file'], file_path)\n assert package1.new.file._data() == file_path\n\n # Add a new group\n package1._add_group('newgroup')\n assert isinstance(package1.newgroup, GroupNode)\n package1.newgroup._add_group('foo')\n assert isinstance(package1.newgroup.foo, GroupNode)\n\n # Overwrite a leaf node\n new_path = os.path.join(mydir, 'data/nuts.csv')\n package1._set(['newgroup', 'foo'], new_path)\n assert package1.newgroup.foo._data() == new_path\n\n # Overwrite the whole group\n package1._set(['newgroup'], new_path)\n assert package1.newgroup._data() == new_path\n\n # Built a new package and verify the new contents\n command.build('foo/package3', package1)\n\n from quilt.data.foo import package3\n\n assert hasattr(package3, 'dataframes2')\n assert not hasattr(package3, 'dataframes')\n\n new_csv = package3.dataframes2.csv._data()\n assert new_csv.xs(0)['Int0'] == 42\n\n new_df = package3.new.df._data()\n assert new_df.xs(2)['a'] == 3\n\n new_file = package3.new.file._data()\n assert isinstance(new_file, string_types)\n\n def test_set_non_node_attr(self):\n mydir = os.path.dirname(__file__)\n build_path = os.path.join(mydir, './build.yml')\n command.build('foo/package1', build_path)\n\n from quilt.data.foo import package1\n\n # Assign a DataFrame as a node\n # (should throw exception)\n df = pd.DataFrame(dict(a=[1, 2, 3]))\n with self.assertRaises(AttributeError):\n package1.newdf = df\n","sub_path":"compiler/quilt/test/test_import.py","file_name":"test_import.py","file_ext":"py","file_size_in_byte":6499,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"73207387","text":"import argparse\nfrom pipeline.interface import TextInterface\n\ndef _launch():\n parser = argparse.ArgumentParser(\n prog = 'python -m pipline',\n description = __doc__\n )\n parser.add_argument(\n '-k',\n '--keep-going',\n action = 'store_true',\n default = False\n )\n parser.add_argument('filename')\n args = parser.parse_args()\n iface = TextInterface(args.keep_going, args.filename)\n iface.cmdloop()\n\nif __name__=='__main__':\n _launch()\n","sub_path":"pipeline/__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":496,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"489128772","text":"from collections import Counter\nfrom wordcloud import WordCloud\nfrom preprocessing.data import (\n Hans,\n HarleyQuinn,\n Jigsaw,\n Joker,\n Fletcher,\n Snowball,\n Plankton\n) \n\n\ndef create_word_cloud(words, name):\n counter = Counter(words)\n cloud = WordCloud(background_color=\"white\")\n cloud.fit_words(counter)\n cloud.to_file('{}.png'.format(name))\n\n\ndef exporter():\n files = {\n \"Hans\": Hans.words,\n \"Fletcher\": Fletcher.words,\n \"Plankton\": Plankton.words,\n \"Snowball\": Snowball.words,\n \"HarleyQuinn\": HarleyQuinn.words,\n \"Jigsaw\": Jigsaw.words,\n \"Joker\": Joker.words\n }\n \n for name, words in files.items():\n create_word_cloud(words, name )\n\nexporter()\n","sub_path":"movie_project/static/img/wordCloudImgs/wordCloud.py","file_name":"wordCloud.py","file_ext":"py","file_size_in_byte":748,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"61950593","text":"''' Basic test of fully rl intersection environment with accelerations as actions. Two lane. Fully rl.\n\nVariables:\n sumo_params {dict} -- [Pass time step, whether safe mode is on or off\n human_lc: strategic -> all advantageous lane changes made\n rl_lc: no_lat_collide -> no lateral collsions\n aggressive -> all lane changes permitted]\n sumo_binary {str} -- [Use either sumo-gui or sumo for visual or non-visual]\n type_params {dict} -- [Types of cars in the system.\n Format {\"name\": (number, (Model, {params}), (Lane Change Model, {params}), initial_speed)}]\n env_params {dict} -- [Params for reward function]\n net_params {dict} -- [Params for network.\n length: road length\n lanes\n speed limit\n resolution: number of edges comprising ring\n net_path: where to store net]\n cfg_params {dict} -- [description]\n initial_config {dict} -- [shuffle: randomly reorder cars to start experiment\n spacing: if gaussian, add noise in start positions\n bunching: how close to place cars at experiment start]\n scenario {[type]} -- [Which road network to use]\n'''\nimport logging\nfrom rllab.envs.normalized_env import normalize\nfrom rllab.misc.instrument import run_experiment_lite, stub\nfrom rllab.algos.trpo import TRPO\nfrom rllab.baselines.linear_feature_baseline import LinearFeatureBaseline\nfrom rllab.policies.gaussian_mlp_policy import GaussianMLPPolicy\nfrom rllab.envs.gym_env import GymEnv\n\nfrom flow.core.params import SumoParams, EnvParams, InitialConfig, NetParams\nfrom flow.core.vehicles import Vehicles\nfrom flow.envs.two_intersection import TwoIntersectionEnvironment\n\nfrom flow.scenarios.intersections.gen import TwoWayIntersectionGenerator\nfrom flow.scenarios.intersections.intersection_scenario import TwoWayIntersectionScenario\nfrom flow.controllers.rlcontroller import RLController\n\nlogging.basicConfig(level=logging.INFO)\n\n\ndef run_task(*_):\n auton_cars = 20\n\n sumo_params = SumoParams(time_step=0.1, human_speed_mode=\"no_collide\", rl_speed_mode= \"no_collide\",\n sumo_binary=\"sumo-gui\")\n\n vehicles = Vehicles()\n vehicles.add_vehicles(\"idm\", (RLController, {}), None, None, 0, 20)\n\n intensity = .2\n v_enter = 10\n env_params = EnvParams(additional_params={\"target_velocity\": v_enter,\n \"control-length\": 150, \"max_speed\": v_enter})\n\n additional_net_params = {\"horizontal_length_in\": 400, \"horizontal_length_out\": 800, \"horizontal_lanes\": 1,\n \"vertical_length_in\": 400, \"vertical_length_out\": 800, \"vertical_lanes\": 1,\n \"speed_limit\": {\"horizontal\": v_enter, \"vertical\": v_enter}}\n net_params = NetParams(no_internal_links=False, additional_params=additional_net_params)\n\n cfg_params = {\"start_time\": 0, \"end_time\": 3000, \"cfg_path\": \"debug/cfg/\"}\n\n initial_config = InitialConfig(spacing=\"custom\", additional_params={\"intensity\": intensity, \"enter_speed\": v_enter})\n\n scenario = TwoWayIntersectionScenario(\"two-way-intersection\", TwoWayIntersectionGenerator,\n vehicles, net_params, initial_config=initial_config)\n\n env = TwoIntersectionEnvironment(env_params, sumo_params, scenario)\n env_name = \"TwoIntersectionEnvironment\"\n pass_params = (env_name, sumo_params, vehicles, env_params, net_params,\n initial_config, scenario)\n\n env = GymEnv(env_name, record_video=False, register_params=pass_params)\n horizon = env.horizon\n env = normalize(env)\n logging.info(\"Experiment Set Up complete\")\n\n print(\"experiment initialized\")\n\n env = normalize(env)\n\n policy = GaussianMLPPolicy(\n env_spec=env.spec,\n hidden_sizes=(64, 64)\n )\n\n baseline = LinearFeatureBaseline(env_spec=env.spec)\n\n algo = TRPO(\n env=env,\n policy=policy,\n baseline=baseline,\n batch_size=30000,\n max_path_length=horizon,\n # whole_paths=True,\n n_itr=200,\n discount=0.999,\n # step_size=0.01,\n )\n algo.train()\n\nfor seed in [1]: # [1, 5, 10, 73, 56]\n run_experiment_lite(\n run_task,\n # Number of parallel workers for sampling\n n_parallel=1,\n # Only keep the snapshot parameters for the last iteration\n snapshot_mode=\"last\",\n # Specifies the seed for the experiment. If this is not provided, a random seed\n # will be used\n seed=seed,\n mode=\"local\",\n exp_prefix=\"intersection-exp\",\n # python_command=flow_config.PYTHON_COMMAND\n # plot=True,\n )\n","sub_path":"examples/two-way-intersection-rl.py","file_name":"two-way-intersection-rl.py","file_ext":"py","file_size_in_byte":4859,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"303556916","text":"import pandas as pd\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn.model_selection import GridSearchCV,ParameterGrid\n\n\ndata = pd.read_csv('./data/BlackFriday.csv')\ndata.fillna(0.0, inplace=True)\nle = LabelEncoder()\ndata['User_ID'] = le.fit_transform(data['User_ID'])\nle = LabelEncoder()\ndata['Product_ID'] = le.fit_transform(data['Product_ID'])\ndata_AGCS = pd.get_dummies(data,columns=['Age','Gender', 'City_Category', 'Stay_In_Current_City_Years'])\ndata_encoded = pd.concat([data.drop('Purchase',axis=1), data_AGCS], axis=1)\ndata_encoded.drop(['Age','Gender', 'City_Category', 'Stay_In_Current_City_Years'], axis=1, inplace=True)\nX = data_encoded.drop('Purchase',axis=1)\ny = data_encoded['Purchase']\n# 标准化\nstd = StandardScaler()\nX = std.fit_transform(X)\nX_train,X_test,y_train,y_test = train_test_split(X,y)\n\n# rfr = RandomForestRegressor(n_estimators=10)\n# rfr.fit(X_train, y_train)\n# y_pre = rfr.predict(X_test)\n# print(y_pre)\n# print('准确率:',rfr.score(X_test, y_test))\n\n\nparameter_space = {\n \"n_estimators\": [2,3,4,5,6,7],\n \"min_samples_leaf\": [2],\n}\nestimator = RandomForestRegressor()\nrfr_cv = GridSearchCV(estimator, parameter_space, cv=4)\nrfr_cv.fit(X_train,y_train)\nprint(rfr_cv.best_params_)\nprint(rfr_cv.best_score_)\n","sub_path":"pre.py","file_name":"pre.py","file_ext":"py","file_size_in_byte":1413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"471743006","text":"#data visualization for GAS usages in Surja Residence\nfrom matplotlib import pyplot as ply\nimport numpy as np\n\nno_of_gas=[1,2,3,4,5]\nno_of_days=[15,14,14,16,12]\n\n#np_data= np_array(data)\nply.plot(no_of_gas,no_of_days)\nply.xlabel(\"No of Gas\")\nply.ylabel(\"No of Days\")\nply.title(\"Usages of Gas\")\nply.show()\n","sub_path":"Matplotlib/GasPlot.py","file_name":"GasPlot.py","file_ext":"py","file_size_in_byte":305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"76050764","text":"# ID: 2018220076\n# Name: Kwon Doyoung\n# -*- coding: utf-8 -*-\nimport numpy as np\n\nfrom sklearn import datasets\nfrom pattern.classifier import classifier, accuracy\n\n\ndef totalResult(train_x, train_y, test_x, test_y, nClass):\n accSet = []\n # model setting\n model = classifier(train_x, train_y, nClass)\n # model 1\n pred = model.predict(test_x, 1)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 1 - accuracy: {:.4f}\".format(acc))\n # model 2\n pred = model.predict(test_x, 2)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 2 - accuracy: {:.4f}\".format(acc))\n # model 3\n pred = model.predict(test_x, 3)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 3 - accuracy: {:.4f}\".format(acc))\n # model 4\n pred = model.predict(test_x, 4)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 4 - accuracy: {:.4f}\".format(acc))\n # model 5\n pred = model.predict(test_x, 5)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 5 - accuracy: {:.4f}\".format(acc))\n # model 6\n pred = model.predict(test_x, 6)\n acc = accuracy(pred, test_y)\n accSet.append(acc)\n print(\"covariance matrix\")\n print(model.covMatrix)\n print(\"mdoel 6 - accuracy: {:.4f}\".format(acc))\n\n return np.array(accSet)\n\n\nif __name__ == \"__main__\":\n iris = datasets.load_iris()\n x = iris.data[:, :]\n y = iris.target\n\n # hw 1 에 대한 결과\n print(\"=======no cross validation part=======\")\n totalResult(x, y, x, y, 3)\n print(\"=======================================\\n\")\n\n # hw2 에 대한 결과\n print(\"===cross validation data processing====\")\n meanAcc = np.zeros([6])\n fold = 5\n for n in range(fold):\n mask = np.array([True for _ in range(150)])\n # R (i+1)th & T (i+1)\n mask[50 - 10 * (n + 1):50 - 10 * n] = False # class 1\n mask[100 - 10 * (n + 1):100 - 10 * n] = False # class 2\n mask[150 - 10 * (n + 1):150 - 10 * n] = False # class 3\n rx = x[mask]\n ry = y[mask]\n tx = x[~mask]\n ty = y[~mask]\n print(\"R{i} & T{i} result\".format(i=n + 1))\n acc = totalResult(rx, ry, tx, ty, 3)\n meanAcc += acc\n meanAcc = meanAcc / fold\n print(\"model mean of accuracy\")\n print(meanAcc)\n","sub_path":"pattern/hw02(class).py","file_name":"hw02(class).py","file_ext":"py","file_size_in_byte":2621,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"420089694","text":"from ex115.lib.interface import *\nfrom ex115.lib.arquivo import *\nfrom time import sleep\n\narq = 'cursoemvideo.txt'\n\nif not arqExiste(arq):\n criarArquivo(arq)\n\n\nwhile True:\n perguta = menu(['Ver pessoas cadatradas', 'Cadastrar nova Pessoa', 'Sair do Sistema'])\n if perguta == 1:\n # Opção de listar o conteúdo de um arquivo.\n lerArquivo(arq)\n elif perguta == 2:\n # Opção de cadastrar nova pessoa.\n cabeçalho('NOVO CADASTRO')\n nome = cadastrarNome()\n idade = leiaInt('Idade: ')\n cadastrar(arq, nome, idade)\n elif perguta == 3:\n # Opção de sair do sistema.\n cabeçalho('Saindo do sistema... Até logo!')\n break\n else:\n # Digitou uma opção errada no menu.\n cor('ERRO: Digite uma opção válida!', 1, False)\n sleep(2)\n","sub_path":"Exercícios/ex115/sistema.py","file_name":"sistema.py","file_ext":"py","file_size_in_byte":828,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"542456934","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('hipchat', '0002_auto_20151213_0657'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='glance',\n name='icon_url',\n field=models.URLField(help_text=b'URL to icon displayed on the left of the glance.', blank=True),\n ),\n migrations.AlterField(\n model_name='glance',\n name='icon_url2',\n field=models.URLField(help_text=b'URL to hi-res icon displayed on the left of the glance.', blank=True),\n ),\n ]\n","sub_path":"hipchat/migrations/0003_auto_20151213_0723.py","file_name":"0003_auto_20151213_0723.py","file_ext":"py","file_size_in_byte":685,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"569855955","text":"import re\nfrom typing import List\nfrom urllib.parse import urljoin\n\nimport httpx\nimport js2py\nfrom faker import Faker\nfrom pydantic import parse_obj_as\n\nfrom kurby.constants import TWIST_URL, ANIME_ENDPOINT\nfrom kurby.decrypt import decrypt\nfrom kurby.schemas import Anime, AnimeDetails, AnimeSource\nfrom kurby.utils import get_chrome_headers\n\nfake = Faker()\n\n\nclass TwistClient(httpx.Client):\n source_key = None\n access_token = None\n\n\ndef get_auth_client() -> TwistClient:\n headers = get_chrome_headers()\n r = httpx.get(url=TWIST_URL, headers=headers)\n r.raise_for_status()\n match = re.search(r\"