diff --git "a/4262.jsonl" "b/4262.jsonl" new file mode 100644--- /dev/null +++ "b/4262.jsonl" @@ -0,0 +1,652 @@ +{"seq_id":"402809948","text":"from diary.modules.resume.serializers import DateSerializer\n\n\nclass DateView(object):\n\n def __init__(self, interactor):\n self.interactor = interactor\n\n def get(self, **kwargs):\n\n user = kwargs.get('user')\n year = kwargs.get('year')\n month = kwargs.get('month')\n day = kwargs.get('day')\n\n date = self.interactor.set_params(\n user,\n year,\n month,\n day\n ).execute()\n\n body = DateSerializer.serialize(date)\n status = 200\n\n return body, status\n","sub_path":"diary/modules/resume/views/date.py","file_name":"date.py","file_ext":"py","file_size_in_byte":719,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"491982388","text":"import numpy as np\n\ndef choosePivot(N):\n return np.random.choice(N)\n\ndef partition(A,pivotId):\n pivot = A[pivotId]\n temp = A[pivotId] \n A[pivotId] = A[0]\n A[0]=temp\n\n N = len(A)\n nextIdToSee = 1\n firstElementLargerThanPivotInPartition = 1\n\n while nextIdToSee pivot:\n pass\n else:\n temp = A[firstElementLargerThanPivotInPartition]\n A[firstElementLargerThanPivotInPartition] = seenNum\n A[nextIdToSee] = temp\n firstElementLargerThanPivotInPartition+=1\n \n nextIdToSee+=1 \n \n A[0] = A[firstElementLargerThanPivotInPartition-1] \n A[firstElementLargerThanPivotInPartition-1] = pivot\n pivotIndexFinal = firstElementLargerThanPivotInPartition-1\n \n return (A,pivotIndexFinal)\n \ndef randSelect(A,i): \n N = len(A)\n pivotId = choosePivot(N) \n (partitioned,finalPivotIndex) = partition(A,pivotId) \n if finalPivotIndex==(i-1): \n return partitioned[finalPivotIndex]\n elif finalPivotIndex < (i-1): \n return randSelect(A[finalPivotIndex:],(i-finalPivotIndex))\n else: \n return randSelect(A[:finalPivotIndex],i)\n\n\nN = 1000\narr = list(np.random.permutation(N))\ni = 512\n\nith_smallest_element = randSelect(arr,i)\n\nprint(f'Array: \\n{arr}')\nprint(f'{i}th smallest element: {ith_smallest_element}')","sub_path":"randomizedSelection.py","file_name":"randomizedSelection.py","file_ext":"py","file_size_in_byte":1442,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"221193935","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\nclass FrequencyAnalysis:\n def __init__(self, audioData, framerate, numSamples, startpoint, endpoint):\n self.data = audioData\n self.frameRate = framerate\n self.numSamples = numSamples\n self.rangeVals = [startpoint, endpoint]\n self.frequencies = self.analyzeFreq()\n\n def analyzeFreq(self):\n data = np.array(self.data)\n data_fft = np.fft.fft(data[self.rangeVals[0]:self.rangeVals[1]],n=self.frameRate)\n frequencies = np.abs(data_fft)\n\n frequencies = frequencies[:int(len(frequencies)/2)]\n\n print(\"The frequency is {} Hz\".format(np.argmax(frequencies)))\n\n plt.subplot(2, 1, 1)\n plt.plot(data[self.rangeVals[0]:self.rangeVals[1]],linewidth = .1)\n plt.title(\"Original audio wave\")\n plt.subplot(2, 1, 2)\n plt.plot(frequencies)\n plt.title(\"Frequencies found\")\n plt.xlim(0, self.frameRate/2)\n plt.show()\n # normalize the values to check for polyphony\n normalizedFrequencies = frequencies/frequencies[np.argmax(frequencies)] # normalize the values\n plt.plot(normalizedFrequencies)\n plt.show()\n for i in len(normalizedFrequencies):\n if normalizedFrequencies[i] > .85:\n finalFrequencies.append(i)\n\n print(finalFrequencies)\n return finalFrequencies\n\n def getFrequency(self):\n return np.argmax(self.frequencies)\n\n","sub_path":"freqanaltesting.py","file_name":"freqanaltesting.py","file_ext":"py","file_size_in_byte":1468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"463840582","text":"# -*- coding:utf-8 -*-\n\n'感知机原始模式'\n\n__author__='litterzhang'\n\nfrom matplotlib import pyplot as plt\nfrom matplotlib import animation\n\n# 初始化感知机参数\ndef init_param(n):\n\tw = [0 for i in range(n)]\n\tb = 0\n\n\tglobal history\n\thistory.append((w[:], b))\n\n\treturn w, b\n\n# 计算点到超平面的距离\ndef calc_dis(point, n, w, b):\n\tdis = 0\n\tfor i in range(n):\n\t\tdis += point[0][i]*w[i]\n\tdis += b\n\tdis *= point[1]\n\treturn dis\n\n# 检查是否存在误分类点\ndef check(training_set, n, w, b):\n\tflag = False\n\n\tfor data in training_set:\n\t\tif calc_dis(data, n, w, b) <= 0:\n\t\t\tflag = True\n\t\t\tw, b = update(data, n, w, b)\n\t\t\tbreak\n\treturn flag, w, b\n\n# 更新感知机参数\ndef update(point, n, w, b):\n\tyt = 1\n\tfor i in range(n):\n\t\tw[i] += yt*point[1]*point[0][i]\n\tb += yt*point[1]\n\n\tglobal history\n\thistory.append((w[:], b))\n\treturn w, b\n\n# 感知机\ndef prece(training_set, n):\n\tw, b = init_param(n)\n\n\twhile True:\n\t\tflag, w, b = check(training_set, n, w, b)\n\t\tif not flag: break\n\treturn w, b\n\n# initialization function: plot the background of each frame\ndef init():\n\tline.set_data([], [])\n\tx, y, x_, y_ = [], [], [], []\n\tfor p in training_set:\n\t\tif p[1] > 0:\n\t\t\tx.append(p[0][0])\n\t\t\ty.append(p[0][1])\n\t\telse:\n\t\t\tx_.append(p[0][0])\n\t\t\ty_.append(p[0][1])\n\n\tplt.plot(x, y, 'bo', x_, y_, 'rx')\n\tplt.axis([-6, 6, -6, 6])\n\tplt.grid(True)\n\tplt.xlabel('x')\n\tplt.ylabel('y')\n\tplt.title('Perceptron Algorithm')\n\treturn line, label\n\n# animation function. this is called sequentially\ndef animate(i):\n\tglobal history, ax, line, label\n\n\tw = history[i][0]\n\tb = history[i][1]\n\n\tif w[0] == 0 and w[1] == 0:\n\t\treturn line, label\n\tif w[1] == 0:\n\t\tx = -b / w[0]\n\t\ty1 = 7\n\t\ty2 = -7\n\t\tline.set_data([x, x], [y1, y2])\n\t\ty1 = 0\n\t\tlabel.set_text(history[i])\n\t\tlabel.set_position([x, y1])\n\telse:\n\t\tx1 = -7\n\t\ty1 = -(b + w[0] * x1) / w[1]\n\t\tx2 = 7\n\t\ty2 = -(b + w[0] * x2) / w[1]\n\t\tline.set_data([x1, x2], [y1, y2])\n\t\tx1 = 0\n\t\ty1 = -(b + w[0] * x1) / w[1]\n\t\tlabel.set_text(history[i])\n\t\tlabel.set_position([x1, y1])\n\treturn line, label\n\ndef draw():\n\tglobal history, ax, line, label\n\n\t# first set up the figure, the axis, and the plot element we want to animate\n\tfig = plt.figure('Perceptron')\n\tax = plt.axes(xlim=(0, 2), ylim=(-2, 2))\n\tline, = ax.plot([], [], 'g', lw=2)\n\tlabel = ax.text([], [], '')\n \n\t# call the animator. blit=true means only re-draw the parts that have changed.\n\tanim = animation.FuncAnimation(fig, animate, init_func=init, frames=len(history), interval=1000, repeat=True, blit=True)\n\tplt.show()\n\t\n\tanim.save('perceptron_origin.gif', fps=2, writer='imagemagick')\n\n\nif __name__=='__main__':\n\ttraining_set = [((4, 3), 1), ((1, 1), -1), ((3, 3), 1)]\n\tn = 2\n\t\n\thistory = list()\n\n\tprece(training_set, n)\n\t\n\tdraw()\n\n\n","sub_path":"Perceptron/preceptron_origin.py","file_name":"preceptron_origin.py","file_ext":"py","file_size_in_byte":2719,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"590045226","text":"from rest_framework import serializers\n\nfrom protein.models import Protein, ProteinConformation, ProteinFamily, Species, ProteinSource, ProteinSegment\nfrom residue.models import Residue, ResidueNumberingScheme, ResidueGenericNumber\nfrom structure.models import Structure\n\n\nclass ProteinSerializer(serializers.ModelSerializer):\n family = serializers.SlugRelatedField(read_only=True, slug_field='slug')\n species = serializers.StringRelatedField(read_only=True)\n source = serializers.StringRelatedField(read_only=True)\n residue_numbering_scheme = serializers.SlugRelatedField(read_only=True, slug_field='short_name')\n class Meta:\n model = Protein\n fields = ('entry_name', 'name', 'accession', 'family', 'species', 'source', 'residue_numbering_scheme',\n 'sequence')\n\n\nclass ProteinFromConformationSerializer(serializers.ModelSerializer):\n protein = serializers.SlugRelatedField(read_only=True, slug_field='entry_name')\n class Meta:\n model = ProteinConformation\n fields = ('protein', )\n\n\nclass ParentProteinFamilySerializer(serializers.ModelSerializer):\n class Meta:\n model = ProteinFamily\n fields = ('slug', 'name')\n\n\nclass ProteinFamilySerializer(serializers.ModelSerializer):\n parent = ParentProteinFamilySerializer(read_only=True)\n class Meta:\n model = ProteinFamily\n fields = ('slug', 'name', 'parent')\n\n\nclass SpeciesSerializer(serializers.ModelSerializer):\n class Meta:\n model = Species\n fields = ('latin_name', 'common_name')\n\n\nclass ResidueNumberingSchemeSerializer(serializers.ModelSerializer):\n class Meta:\n model = ResidueNumberingScheme\n fields = ('slug', 'short_name')\n\n\nclass ResidueGenericNumberSerializer(serializers.ModelSerializer):\n scheme = serializers.SlugRelatedField(read_only=True, slug_field='short_name')\n class Meta:\n model = ResidueGenericNumber\n fields = ('scheme', 'label')\n\n\nclass ResidueSerializer(serializers.ModelSerializer):\n protein_segment = serializers.StringRelatedField(read_only=True)\n display_generic_number = serializers.StringRelatedField(read_only=True)\n class Meta:\n model = Residue\n fields = ('sequence_number', 'amino_acid', 'protein_segment', 'display_generic_number')\n\n\nclass ResidueExtendedSerializer(serializers.ModelSerializer):\n protein_segment = serializers.StringRelatedField(read_only=True)\n display_generic_number = serializers.StringRelatedField(read_only=True)\n alternative_generic_numbers = ResidueGenericNumberSerializer(read_only=True, many=True)\n class Meta:\n model = Residue\n fields = ('sequence_number', 'amino_acid', 'protein_segment', 'display_generic_number',\n 'alternative_generic_numbers')\n\n\nclass StructureSerializer(serializers.ModelSerializer):\n pdb_code = serializers.SlugRelatedField(read_only=True, slug_field='index')\n protein_conformation = ProteinFromConformationSerializer()\n class Meta:\n model = Structure\n fields = ('pdb_code', 'resolution', 'protein_conformation')","sub_path":"api/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":3078,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"304806165","text":"from django.contrib.auth.models import User, Group\nfrom django.db.models import fields\nfrom django.db.models.base import Model\nfrom rest_framework import serializers\nfrom .models import *\n\n\nclass tarjetaSerializer(serializers.ModelSerializer):\n ciudadnombre = serializers.CharField(\n source='ciudad.nombre',\n read_only=True\n )\n\n class Meta:\n model = tarjeta\n fields = ['ciudadnombre', 'ciudad_id', 'id', 'titular',\n 'numeroTarjeta', 'fechaExpiracion', 'cvv']\n\n\nclass personaSerializer(serializers.ModelSerializer):\n ciudadnombre = serializers.CharField(\n source='ciudad.nombre',\n read_only=True\n )\n\n class Meta:\n model = persona\n fields = ['nombre', 'direccion', 'telefono',\n 'usuario', 'contrasena', 'ciudadnombre']\n extra_kwargs = {'contrasena': {'write_only': True, 'required': True}}\n\n\nclass ciudadSerializer(serializers.ModelSerializer):\n class Meta:\n model = ciudad\n fields = ['id', 'nombre']\n \nclass categoriaSerializer(serializers.ModelSerializer):\n class Meta:\n model = categoria\n fields = ['id', 'nombre']\n\n\n\nclass productoSerializer (serializers.ModelSerializer): \n categoria=serializers.CharField(\n source='categoria.nombre',\n read_only=True\n )\n \n class Meta:\n model = producto\n fields = ['nombre', 'categoria','activo','precio']\n \n \nclass GrupoyModificacionSerializer(serializers.ModelSerializer):\n \n nombreModificacion=serializers.CharField(\n source='modificador.nombre',\n read_only=True\n ) \n class Meta:\n model = unionModificacion\n fields = ['id','grupoModificador','nombreModificacion']\n \n \nclass grupo_modificaSerializer (serializers.ModelSerializer):\n \n class Meta:\n model = grupo_modifica\n fields = ['id','nombre']\n\nclass conexionProductoSerializer (serializers.ModelSerializer):\n productonombre=serializers.CharField(\n source='producto.nombre',\n read_only=True\n ) \n productoprecio=serializers.CharField(\n source='producto.precio',\n read_only=True\n )\n grupoModificador=serializers.CharField(\n source='grupoModificador.nombre',\n read_only=True\n )\n \n class Meta:\n model = conexionProducto\n fields = ['producto','productonombre','productoprecio','unionModificacion','grupoModificador']\n\n \n\n \n","sub_path":"BackEnd/DjangoApi/api/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":2487,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"467914472","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Aug 21 21:26:31 2017\n\n@author: EthanMorse\n\"\"\"\n\nimport os \n\n# run once for each path to create a .txt file, never use again\n# account_type is school, personal, business, etc in string form\ndef create_old_file_list(account_type):\n \n file = open(\"old_files_dict_\" + account_type.lower() + \".txt\",\"w\")\n file.close()\n \n \n# update old file list so it holds current files in dictionary form\ndef update_file_list(path, account_type):\n \n old_files = open(\"old_files_dict_\" + account_type.lower() + \".txt\",\"w\")\n old_files_dict = {}\n \n # iterate through files in path and enter filename into \n # dictionary, with last mod. times as its value\n for file in os.listdir(path):\n old_files_dict[str(file)] = os.path.getmtime(path + \"/\" + str(file))\n \n old_files.write(str(old_files_dict)) # write entire dict into file\n old_files.close()\n \n \ndef check_if_modified(path, account_type, mod_file):\n \n with open(\"old_files_dict_\" + account_type + \".txt\",\"r\") as file:\n old_files = file.readlines()[0] # read, subscript string (dict)\n \n old_files_dict = eval(old_files) # convert back to dict from string\n \n try:\n # check if last mod. time is same as most recent mod. time\n if os.path.getmtime(path + \"/\" + mod_file) == old_files_dict[mod_file]:\n return False # if NOT modified\n else:\n return True # if modified\n \n except KeyError:\n return None","sub_path":"file_functions.py","file_name":"file_functions.py","file_ext":"py","file_size_in_byte":1539,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"94934567","text":"\r\nlist1 = [1,2,3]\r\nlist2=[4,5,6]\r\n\r\nlist3=list1+list2\r\n\r\nnumbers=[1,2,3,4,5]\r\nn=5\r\nif n in numbers:\r\n print('{}가 있다'.format(n))\r\n\r\ndel list1[1]\r\n#or\r\n#remove list(2)\r\n\r\nprint(list1)\r\n","sub_path":"Try! helloworld/week 5/practice 2 (week 5).py","file_name":"practice 2 (week 5).py","file_ext":"py","file_size_in_byte":192,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"214113876","text":"import cv2\nimport numpy as np\nimport glob\nimport time\n\n\ndef load_all_image_from_path(path):\n image_list = []\n for filename in glob.glob(path):\n # load image in gray scale\n im = cv2.imread(filename, cv2.IMREAD_GRAYSCALE)\n image_list.append(im)\n return image_list\n\n\ndef display_image(image, title=\"image\"):\n cv2.imshow(title, image)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n\n\ndef combine_image(image1, image2):\n h1, w1 = image1.shape[:2]\n h2, w2 = image2.shape[:2]\n print(\"------------------------------------\\n\\n\")\n\n # create empty matrix\n vis = np.zeros((max(h1, h2), w1 + w2), np.uint8)\n\n # combine 2 images\n vis[:h1, :w1] = image1\n vis[:h2, w1:w1 + w2] = image2\n return vis\n\n\ndef orb_with_flann(image_query, image_train):\n # init feature detector\n orb = cv2.ORB_create(nfeatures=2000) # default features is 500\n\n # find key point and descriptor\n kp_logo, des_logo = orb.detectAndCompute(image_train, None)\n kp_img, des_img = orb.detectAndCompute(image_query, None)\n\n # view key point\n # result_image_train = cv2.drawKeypoints(image_train, kp_logo, None, flags=0)\n # result_image_query = cv2.drawKeypoints(image_query, kp_img, None, flags=0)\n # display_image(result_image_train,\"train\")\n # display_image(result_image_query,\"query\")\n\n # FLANN parameters\n flann_index_lsh = 6\n index_params = dict(algorithm=flann_index_lsh,\n table_number=12,\n key_size=20,\n multi_probe_level=2)\n search_params = dict(checks=100) # or pass empty dictionary\n\n # create FLANN\n flann = cv2.FlannBasedMatcher(index_params, search_params)\n\n # perform matching\n flann_matches = flann.knnMatch(des_logo, des_img, k=2)\n\n # View match without fillter.\n # img3 = cv2.drawMatchesKnn(image_train, kp_logo, image_query, kp_img, flann_matches, None)\n # display_image(img3)\n\n # Need to draw only good matches, so create a mask\n matches_mask = [[0, 0] for i in range(len(flann_matches))]\n\n # ratio test as per Lowe's paper\n good = []\n for index in range(len(flann_matches)):\n if len(flann_matches[index]) == 2:\n m, n = flann_matches[index]\n if m.distance < 0.8 * n.distance: # threshold of ratio testing\n matches_mask[index] = [1, 0]\n good.append(flann_matches[index])\n\n # draw match after filter\n draw_params = dict(\n singlePointColor=(255, 0, 0),\n matchesMask=matches_mask,\n flags=2)\n\n img3 = cv2.drawMatchesKnn(image_train, kp_logo, image_query, kp_img, flann_matches, None, **draw_params)\n\n # font = cv2.FONT_HERSHEY_SIMPLEX\n # cv2.putText(img3, str(len(good)), (0, 50), font, 2, (0, 0, 255), 2, cv2.LINE_AA)\n\n display_image(img3) # view matching image\n\n return image_train, len(good)\n\n\ndef find_best_match_index(match):\n index_best = 0\n best_length = 0\n for index, (image, good_match_length) in enumerate(match):\n if good_match_length >= best_length:\n best_length = good_match_length\n index_best = index\n return index_best\n\n\nif __name__ == '__main__':\n # load image\n train_image_list = load_all_image_from_path(\"train_image/*\")\n query_image_list = load_all_image_from_path(\"sample_image/*\")\n\n for index, query_image in enumerate(query_image_list):\n matches = []\n\n # get process time\n start = int(round(time.time() * 1000))\n for train_image in train_image_list:\n matches.append(orb_with_flann(query_image, train_image))\n end = int(round(time.time() * 1000)) - start\n print(\"process time: \" + end.__str__())\n\n # find the best match\n best_match_index = find_best_match_index(matches)\n\n # combine query image with the best match image so easy view\n result_image = combine_image(query_image, matches[best_match_index][0])\n\n font = cv2.FONT_HERSHEY_SIMPLEX\n result_image = cv2.cvtColor(result_image, cv2.COLOR_GRAY2BGR)\n result_image = cv2.putText(result_image, str(matches[best_match_index][1]) + \" - \" + \"time: \" + str(end),\n (0, 50), font, 2, (0, 0, 255),\n 2, cv2.LINE_AA)\n display_image(result_image)\n\n # save result image\n # cv2.imwrite(\"result/result\" + str(index) + \".jpg\", result_image)\n\n","sub_path":"wine_matcher.py","file_name":"wine_matcher.py","file_ext":"py","file_size_in_byte":4422,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"132599123","text":"###JAX implementation of ADAM\nimport jax.numpy as jnp\nfrom jax.experimental.optimizers import adam\nfrom jax.lax import while_loop,cond\nfrom jax import grad,value_and_grad, jit\nfrom functools import partial\n\n### Optimizer loop and termination\n@partial(jit,static_argnums = (0))\ndef minimize_ADAM(fun,x0,data,step_size = 0.01,f_tol = 1e-5,n_iter_max = 1000): \n \"\"\"Minimize function with ADAM.\n Args:\n fun: function to minimize, which takes (x,data) as input\n x0: ndarray: initial solution \n data: ndarray: full data to subsample\n \n #Optimizer hyperparams\n step_size: positive scalar for step-size to pass into minimize_adam\n f_tol: positive scalar for value of norm of f change before termination\n n_iter_max: maximum number of sgd steps before termination\n\n Returns:\n x_opt: Optimized x\n loss: Minimized loss value\n n_iter: Number of iterations at termination\n delta_f: Value of change of f at termination\n \"\"\"\n\n #function wrappers\n value_and_grad_fun = jit(value_and_grad(fun))\n\n @jit #wrapper around step to allow for termination checks\n def step(carry): \n i,loss,g,opt_state,key,n = carry\n key,*subkey = random.split(key)\n ind = random.shuffle(key,jnp.arange(n))\n params = get_params(opt_state)\n loss,g = value_and_grad_fun(params,data[ind])\n i=i+1\n carry = i,loss,g,opt_update(i, g, opt_state)\n return carry\n\n @jit #termination condition on gradient norm or number of iterations\n def converged(carry): \n i,loss,g,opt_state,key,n = carry\n delta_f = 1\n return jnp.logical_and(norm_g>g_tol,i10,}円\".format(num0, num1))\n","sub_path":"Python/09/Sample3.py","file_name":"Sample3.py","file_ext":"py","file_size_in_byte":425,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"54"} +{"seq_id":"490338808","text":"#!/bin/python\n# -*- coding: utf-8 -*-\n\nimport sys\n\nfrom PyQt4 import QtCore, QtGui\n\nfrom model.model import Encuesta\nfrom model.model import Pregunta\nfrom model.model import Opcion\nfrom model.model import TipoRespuesta\nfrom model.model import Secuencia\nfrom model.model import Ocurrencia\nimport model.model as modelo\n\nfrom utils.verifier import Verifier\n\nclass QuestionDisplay(object):\n # Array de elementos qt a mostrar\n\n def __init__(self, pregunta):\n self.elementos = []\n self.pregunta = pregunta\n\n def elegirEstrategia(self):\n # Vemos de qué se trata y elegimos lo mejor para hacer\n\n if (self.pregunta.idTipoRespuesta == modelo.OPC_S_TXT):\n # Texto corto\n self.showOpciones = self.showOpcionesTextoCorto\n\n elif self.pregunta.idTipoRespuesta == modelo.OPC_COMBO:\n # Combo\n self.showOpciones = self.showOpcionesCombo\n\n elif self.pregunta.idTipoRespuesta == modelo.OPC_RADIO:\n # Radio\n self.showOpciones = self.showOpcionesRadio\n\n elif self.pregunta.idTipoRespuesta == modelo.OPC_CHECK:\n # Checkboxes\n self.showOpciones = self.showOpcionesCheck\n\n def removerPregunta(self, ui):\n # Limpia el display de preguntas\n for i in reversed(range(ui.opcionPregunta.count())):\n ui.opcionPregunta.itemAt(i).widget().setParent(None)\n\n def displayPregunta(self, ui):\n ui.textoPregunta.setText(QtCore.QString(self.pregunta.texto_pregunta))\n opciones = modelo.session.query(Opcion).\\\n filter(Opcion.idPregunta.in_([self.pregunta.idPregunta])).all()\n\n # Elemento qt usado en las opciones\n tipoRespuestaOpcion = modelo.session.query(TipoRespuesta).\\\n filter(TipoRespuesta.idTipoRespuesta.in_([opciones[0].idTipoRespuesta])).all()\n\n self.removerPregunta(ui)\n\n self.showOpciones(tipoRespuestaOpcion, ui, opciones)\n\n #print('opciones: {}\\nelementos {}'.format(opciones, self.elementos))\n return opciones, self.elementos\n\n def showOpcionesCombo(self,tipoRespuestaOpcion, ui, opciones):\n lista = []\n row = 0\n nombre = 'QtGui.' + tipoRespuestaOpcion[0].elemento_qt_usado\n Elemento = eval(nombre)\n self.elementos.append(Elemento())\n\n for o in opciones:\n lista.append(o.texto_opcion)\n\n # Acá ya está el elemento ComboBox\n self.elementos[0].addItems(lista)\n # Mostramos el coso en la ventana\n ui.opcionPregunta.addWidget(self.elementos[0],row,0)\n\n def showOpcionesCheck(self, tipoRespuestaOpcion, ui, opciones):\n lista = []\n row = 0\n btnId = 1\n nombre = 'QtGui.' + tipoRespuestaOpcion[0].elemento_qt_usado\n Elemento = eval(nombre)\n #print(Elemento)\n\n for o in opciones:\n choice = Elemento(o.texto_opcion)\n self.elementos.append(choice)\n ui.opcionPregunta.addWidget(self.elementos[row],row,0)\n row +=1\n btnId += 1\n\n def showOpcionesRadio(self, tipoRespuestaOpcion, ui, opciones):\n lista = []\n row = 0\n btnId = 1\n nombre = 'QtGui.' + tipoRespuestaOpcion[0].elemento_qt_usado\n Elemento = eval(nombre)\n #print(Elemento)\n self.elementos.append(QtGui.QButtonGroup())\n\n for o in opciones:\n choice = Elemento(o.texto_opcion)\n self.elementos[0].addButton(choice,btnId)\n ui.opcionPregunta.addWidget(self.elementos[0].button(row+1),row,0)\n row +=1\n btnId += 1\n\n\n def showOpcionesTextoCorto(self, tipoRespuestaOpcion, ui, opciones):\n row = 0\n for o in opciones:\n # Se muestra todo en formato